Title
Including Uncertainty when Learning from Human Corrections.
Abstract
It is difficult for humans to efficiently teach robots how to correctly perform a task. One intuitive solution is for the robot to iteratively learn the humanu0027s preferences from corrections, where the human improves the robotu0027s current behavior at each iteration. When learning from corrections, we argue that while the robot should estimate the most likely human preferences, it should also know what it does not know, and integrate this uncertainty when making decisions. We advance the state-of-the-art by introducing a Kalman filter for learning from corrections: this approach also maintains the uncertainty of the estimated human preferences. Next, we demonstrate how uncertainty can be leveraged for active learning and risk-sensitive deployment. Our results indicate that maintaining and leveraging uncertainty leads to faster learning from human corrections.
Year
Venue
DocType
2018
CoRL
Conference
Volume
Citations 
PageRank 
abs/1806.02454
2
0.38
References 
Authors
13
2
Name
Order
Citations
PageRank
dylan p losey15210.77
Marcia K. O'Malley245674.32