Title
The Concept of Criticality in Reinforcement Learning
Abstract
This paper introduces a novel idea in human-aided reinforcement learning - the concept of criticality. The criticality of a state indicates how much the choice of action in that particular state influences the expected return. In order to develop an intuition for the concept, we present examples of plausible criticality functions in multiple environments. Furthermore, we formulate a practical application of criticality in reinforcement learning: the criticality-based varying stepnumber algorithm (CVS) - a flexible stepnumber algorithm that utilizes the criticality function, provided by a human, in order to avoid the problem of choosing an appropriate stepnumber in n-step algorithms such as n-step SARSA and n-step Tree Backup. We present experiments in the Atari Pong environment demonstrating that CVS is able to outperform popular learning algorithms such as Deep Q-Learning and Monte Carlo.
Year
DOI
Venue
2019
10.1109/ICTAI.2019.00043
2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)
Keywords
Field
DocType
Human aided reinforcement learning Human agent interaction
Monte Carlo method,Computer science,Intuition,Artificial intelligence,Criticality,Machine learning,Expected return,Backup,Reinforcement learning
Conference
Volume
ISSN
ISBN
abs/1810.07254
1082-3409
978-1-7281-3799-5
Citations 
PageRank 
References 
0
0.34
4
Authors
2
Name
Order
Citations
PageRank
Yitzhak Spielberg100.34
Amos Azaria227232.02