Title
A survey of inverse reinforcement learning: Challenges, methods and progress
Abstract
Inverse reinforcement learning (IRL) is the problem of inferring the reward function of an agent, given its policy or observed behavior. Analogous to RL, IRL is perceived both as a problem and as a class of methods. By categorically surveying the extant literature in IRL, this article serves as a comprehensive reference for researchers and practitioners of machine learning as well as those new to it to understand the challenges of IRL and select the approaches best suited for the problem on hand. The survey formally introduces the IRL problem along with its central challenges such as the difficulty in performing accurate inference and its generalizability, its sensitivity to prior knowledge, and the disproportionate growth in solution complexity with problem size. The article surveys a vast collection of foundational methods grouped together by the commonality of their objectives, and elaborates how these methods mitigate the challenges. We further discuss extensions to the traditional IRL methods for handling imperfect perception, an incomplete model, learning multiple reward functions and nonlinear reward functions. The article concludes the survey with a discussion of some broad advances in the research area and currently open research questions.
Year
DOI
Venue
2018
10.1016/j.artint.2021.103500
Artificial Intelligence
Keywords
Field
DocType
Reinforcement learning,Reward function,Learning from demonstration,Generalization,Learning accuracy,Survey
Open research,Generalizability theory,Inference,Correctness,Inverse reinforcement learning,Artificial intelligence,Perception,Machine learning,Mathematics
Journal
Volume
Issue
ISSN
297
1
0004-3702
Citations 
PageRank 
References 
5
0.47
50
Authors
2
Name
Order
Citations
PageRank
Saurabh Arora151.48
Prashant Doshi292690.23