Title
Modelling Agent Policies with Interpretable Imitation Learning.
Abstract
As we deploy autonomous agents in safety-critical domains, it becomes important to develop an understanding of their internal mechanisms and representations. We outline an approach to imitation learning for reverse-engineering black box agent policies in MDP environments, yielding simplified, interpretable models in the form of decision trees. As part of this process, we explicitly model and learn agents' latent state representations by selecting from a large space of candidate features constructed from the Markov state. We present initial promising results from an implementation in a multi-agent traffic environment.
Year
DOI
Venue
2020
10.1007/978-3-030-73959-1_16
TAILOR
DocType
Citations 
PageRank 
Conference
0
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Tom Bewley100.68
Jonathan Lawry217219.06
Arthur Richards326826.94