Title
Interpreting Cnns Via Decision Trees
Abstract
This paper(1) aims to quantitatively explain the rationales of each prediction that is made by a pre-trained convolutional neural network (CNN). We propose to learn a decision tree, which clarifies the specific reason for each prediction made by the CNN at the semantic level. I.e. the decision tree decomposes feature representations in high cony-layers of the CNN into elementary concepts of object parts. In this way, the decision tree tells people which object parts activate which filters for the prediction and how much each object part contributes to the prediction score. Such semantic and quantitative explanations for CNN predictions have specific values beyond the traditional pixel-level analysis of CNNs. More specifically, our method mines all potential decision modes of the CNN, where each mode represents a typical case of how the CNN uses object parts for prediction. The decision tree organizes all potential decision modes in a coarse-to-fine manner to explain CNN predictions at different fine-grained levels. Experiments have demonstrated the effectiveness of the proposed method.
Year
DOI
Venue
2019
10.1109/CVPR.2019.00642
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019)
Field
DocType
ISSN
Decision tree,Pattern recognition,Computer science,Artificial intelligence
Conference
1063-6919
Citations 
PageRank 
References 
5
0.38
0
Authors
4
Name
Order
Citations
PageRank
Quanshi Zhang128826.67
Yang Yu248455.96
Haotian Ma352.75
Ying Nian Wu41652267.72