Title
When a Robot Reaches Out for Human Help.
Abstract
In many realistic planning situations, any policy has a nonzero probability of reaching a dead-end. In such cases, a popular approach is to plan to maximize the probability of reaching the goal. While this strategy increases the robustness and expected autonomy of the robot, it considers that the robot gives up on the task whenever a dead-end is encountered. In this work, we consider planning for agents that proactively and autonomously resort to human help when an unavoidable dead-end is encountered (the so-called symbiotic agents). To this end, we develop a new class of Goal-Oriented Markov Decision Process that includes a set of human actions that ensures the existence of a proper policy, one that possibly resorts to human help. We discuss two different optimization criteria: minimizing the probability to use human help and minimizing the expected cumulative cost with a finite penalty for using human help for the first time. We show that for a large enough penalty both criteria are equivalent. We report on experiments with standard probabilistic planning domains for reasonably large problems.
Year
DOI
Venue
2018
10.1007/978-3-030-03928-8_23
ADVANCES IN ARTIFICIAL INTELLIGENCE - IBERAMIA 2018
Keywords
Field
DocType
Probabilistic planning,Shortest stochastic path,Human-robot collaboration
Mathematical optimization,Computer science,Autonomy,Markov decision process,Robustness (computer science),Probabilistic logic,Robot
Conference
Volume
ISSN
Citations 
11238
0302-9743
0
PageRank 
References 
Authors
0.34
15
4
Name
Order
Citations
PageRank
Ignasi Andrés100.34
Leliane Nunes de Barros232.74
Denis Deratani Mauá316524.64
Thiago D. Simão402.70