Abstract | ||
---|---|---|
This paper introduces a novel idea in human-aided reinforcement learning - the concept of criticality. The criticality of a state indicates how much the choice of action in that particular state influences the expected return. In order to develop an intuition for the concept, we present examples of plausible criticality functions in multiple environments. Furthermore, we formulate a practical application of criticality in reinforcement learning: the criticality-based varying stepnumber algorithm (CVS) - a flexible stepnumber algorithm that utilizes the criticality function, provided by a human, in order to avoid the problem of choosing an appropriate stepnumber in n-step algorithms such as n-step SARSA and n-step Tree Backup. We present experiments in the Atari Pong environment demonstrating that CVS is able to outperform popular learning algorithms such as Deep Q-Learning and Monte Carlo. |
Year | DOI | Venue |
---|---|---|
2019 | 10.1109/ICTAI.2019.00043 | 2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI) |
Keywords | Field | DocType |
Human aided reinforcement learning Human agent interaction | Monte Carlo method,Computer science,Intuition,Artificial intelligence,Criticality,Machine learning,Expected return,Backup,Reinforcement learning | Conference |
Volume | ISSN | ISBN |
abs/1810.07254 | 1082-3409 | 978-1-7281-3799-5 |
Citations | PageRank | References |
0 | 0.34 | 4 |
Authors | ||
2 |
Name | Order | Citations | PageRank |
---|---|---|---|
Yitzhak Spielberg | 1 | 0 | 0.34 |
Amos Azaria | 2 | 272 | 32.02 |