Title
A Framework For Learning From Demonstration With Minimal Human Effort
Abstract
We consider robot learning in the context of shared autonomy, where control of the system can switch between a human teleoperator and autonomous control. In this setting we address reinforcement learning, and learning from demonstration, where there is a cost associated with human time. This cost represents the human time required to teleoperate the robot, or recover the robot from failures. For each episode, the agent must choose between requesting human teleoperation, or using one of its autonomous controllers. In our approach, we learn to predict the success probability for each controller, given the initial state of an episode. This is used in a contextual multi-armed bandit algorithm to choose the controller for the episode. A controller is learnt online from demonstrations and reinforcement learning so that autonomous performance improves, and the system becomes less reliant on the teleoperator with more experience. We show that our approach to controller selection reduces the human cost to perform two simulated tasks and a single real-world task.
Year
DOI
Venue
2020
10.1109/LRA.2020.2970619
IEEE ROBOTICS AND AUTOMATION LETTERS
Keywords
DocType
Volume
Learning from demonstration, human-centered robotics, telerobotics and teleoperation
Journal
5
Issue
ISSN
Citations 
2
2377-3766
3
PageRank 
References 
Authors
0.40
0
3
Name
Order
Citations
PageRank
Marc Rigter141.78
Bruno Lacerda28512.96
Nick Hawes332134.18