Title
Generating Contrastive Explanations with Monotonic Attribute Functions.
Abstract
Explaining decisions of deep neural networks is a hot research topic with applications in medical imaging, video surveillance, and self driving cars. Many methods have been proposed in literature to explain these decisions by identifying relevance of different pixels. In this paper, we propose a method that can generate contrastive explanations for such data where we not only highlight aspects that are in themselves sufficient to justify the classification by the deep model, but also new aspects which if added will change the classification. One of our key contributions is how we define "addition" for such rich data in a formal yet humanly interpretable way that leads to meaningful results. This was one of the open questions laid out in Dhurandhar et.al. (2018) [5], which proposed a general framework for creating (local) contrastive explanations for deep models. We showcase the efficacy of our approach on CelebA and Fashion-MNIST in creating intuitive explanations that are also quantitatively superior compared with other state-of-the-art interpretability methods.
Year
Venue
DocType
2019
arXiv: Learning
Journal
Volume
Citations 
PageRank 
abs/1905.12698
1
0.35
References 
Authors
0
6
Name
Order
Citations
PageRank
Ronny Luss110210.30
Pin-Yu Chen264674.59
Amit Dhurandhar37019.33
Prasanna Sattigeri48517.23
Karthikeyan Shanmugam545637.49
Chun Chen Tu6383.20