Title | ||
---|---|---|
Interpretable Approximation of a Deep Reinforcement Learning Agent as a Set of If-Then Rules |
Abstract | ||
---|---|---|
In many industrial applications, one of the major bottlenecks in using advanced learning-based methods (such as reinforcement learning) for controls is the lack of interpretability of the trained agent. In this paper, we present a methodology for translating a trained reinforcement learning agent into a set of simple and easy to interpret if-then rules by using the proven universal approximation property of the rules with fuzzy predicates. Proposed methodology combines the optimality of reinforcement learning with interpretability of the theory of approximate reasoning, thus making reinforcement learning-based solutions more accessible to industrial practitioners. The framework presented in this paper has the potential to help address the fundamental problem in widespread adoption of reinforcement learning in industrial applications. |
Year | DOI | Venue |
---|---|---|
2019 | 10.1109/ICMLA.2019.00041 | 2019 18th IEEE International Conference On Machine Learning And Applications (ICMLA) |
Keywords | Field | DocType |
Reinforcement learning,car following,explainable AI,Fuzzy rules | Car following,Interpretability,Computer science,Fuzzy predicates,Approximate reasoning,Artificial intelligence,Approximation property,Machine learning,Reinforcement learning | Conference |
ISBN | Citations | PageRank |
978-1-7281-4551-8 | 0 | 0.34 |
References | Authors | |
0 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Subramanya P. Nageshrao | 1 | 45 | 4.95 |
Bruno Costa | 2 | 36 | 7.51 |
Dimitar P. Filev | 3 | 0 | 0.34 |