Title
Extracting Interpretable Concept-Based Decision Trees from CNNs.
Abstract
In an attempt to gather a deeper understanding of how convolutional neural networks (CNNs) reason about human-understandable concepts, we present a method to infer labeled concept data from hidden layer activations and interpret the concepts through a shallow decision tree. The decision tree can provide information about which concepts a model deems important, as well as provide an understanding how the concepts interact with each other. Experiments demonstrate that the extracted decision tree is capable of accurately representing the original CNN's classifications at low tree depths, thus encouraging human-in-the-loop understanding of discriminative concepts.
Year
Venue
DocType
2019
CoRR
Journal
Volume
Citations 
PageRank 
abs/1906.04664
0
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Conner Chyung100.34
michael tsang2142.58
Yan Liu316844.76