Name
Papers
Collaborators
RONALD PARR
50
39
Citations 
PageRank 
Referers 
2428
186.85
3432
Referees 
References 
528
501
Search Limit
1001000
Title
Citations
PageRank
Year
Revisiting the Softmax Bellman Operator: New Benefits and New Perspective00.342019
Revisiting the Softmax Bellman Operator: Theoretical Properties and Practical Benefits.00.342018
Efficient PAC-Optimal Exploration in Concurrent, Continuous State MDPs with Delayed Updates.60.512016
Distance Minimization for Reward Learning from Scored Trajectories.50.442016
Improving PAC Exploration Using the Median Of Means.00.342016
Linear Feature Encoding for Reinforcement Learning.00.342016
PAC Optimal Exploration in Continuous Space Markov Decision Processes.190.832013
Sample Complexity and Performance Bounds for Non-Parametric Approximate Linear Programming.10.362013
Policy Iteration for Factored MDPs507.842013
Value function approximation in zero-sum markov games160.942013
Computing Optimal Strategies to Commit to in Stochastic Games.90.612012
Greedy Algorithms for Sparse Reinforcement Learning.210.892012
Value Function Approximation in Noisy Environments Using Locally Smoothed Regularized Approximate Linear Programs40.522012
Computing Stackelberg strategies in stochastic games30.422012
Security games with multiple attacker resources261.182011
Generalized Value Functions for Large Action Sets.110.662011
Solving Stackelberg games with uncertain observability302.032011
Efficient solution algorithms for factored MDPs1517.212011
Non-Parametric Approximate Linear Programming for MDPs.100.692011
Counting Objects with a Combination of Horizontal and Overhead Sensors00.342010
Feature Selection Using Regularization in Approximate Linear Programs for Markov Decision Processes321.582010
Linear Complementarity for Regularized Policy Evaluation and Improvement.251.382010
Multi-step multi-sensor hider-seeker games302.402009
Kernelized value function approximation for reinforcement learning471.662009
An analysis of linear models, linear value-function approximation, and feature selection for reinforcement learning652.482008
Planning Aims for a Network of Horizontal and Overhead Sensors10.412008
Nonmyopic Multiaspect Sensing With Partially Observable Markov Decision Processes180.952007
Point-Based Policy Iteration100.612007
Analyzing feature generation for value-function approximation593.222007
Efficient Selection of Disambiguating Actions for Stereo Vision00.342006
Hierarchical Linear/Constant Time SLAM Using Particle Filters for Dense Maps231.232005
Learning probabilistic motion models for mobile robots302.062004
DP-SLAM 2.0563.102004
Reinforcement Learning as Classification: Leveraging Modern Classifiers704.982003
Least-squares policy iteration51925.742003
DP-SLAM: fast, robust simultaneous localization and mapping without predetermined landmarks1238.622003
Approximate policy iteration using large-margin classifiers00.342003
Coordinated Reinforcement Learning804.772002
Learning in Zero-Sum Team Markov Games Using Factored Value Functions50.492002
XPathLearner: an on-line self-tuning Markov histogram for XML path selectivity estimation491.892002
Least-Squares Methods in Reinforcement Learning for Control261.992002
Model-Free Least-Squares Policy Iteration403.682001
Max-norm projections for factored MDPs576.762001
Multiagent Planning with Factored MDPs1158.592001
Making Rational Decisions Using Adaptive Utility Elicitation14612.052000
Computing Factored Value Functions for Policies in Structured MDPs709.501999
Reinforcement learning with hierarchies of machines22118.141997
Generalized Prioritized Sweeping.00.341997
Approximating optimal policies for partially observable stochastic domains6621.651995
Provably bounded-optimal agents839.081995