Title
Interpretable Approximation of a Deep Reinforcement Learning Agent as a Set of If-Then Rules
Abstract
In many industrial applications, one of the major bottlenecks in using advanced learning-based methods (such as reinforcement learning) for controls is the lack of interpretability of the trained agent. In this paper, we present a methodology for translating a trained reinforcement learning agent into a set of simple and easy to interpret if-then rules by using the proven universal approximation property of the rules with fuzzy predicates. Proposed methodology combines the optimality of reinforcement learning with interpretability of the theory of approximate reasoning, thus making reinforcement learning-based solutions more accessible to industrial practitioners. The framework presented in this paper has the potential to help address the fundamental problem in widespread adoption of reinforcement learning in industrial applications.
Year
DOI
Venue
2019
10.1109/ICMLA.2019.00041
2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA)
Keywords
Field
DocType
Reinforcement learning,car following,explainable AI,Fuzzy rules
Car following,Interpretability,Computer science,Fuzzy predicates,Approximate reasoning,Artificial intelligence,Approximation property,Machine learning,Reinforcement learning
Conference
ISBN
Citations 
PageRank 
978-1-7281-4551-8
0
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Subramanya P. Nageshrao1454.95
Bruno Costa2367.51
Dimitar P. Filev300.34