Title
Collaborative Explanation of Deep Models with Limited Interaction for Trade Secret and Privacy Preservation
Abstract
An ever increasing number of decisions affecting our lives are made by algorithms. For this reason, algorithmic transparency is becoming a pressing need: automated decisions should be explainable and unbiased. A straightforward solution is to make the decision algorithms open-source, so that everyone can verify them and reproduce their outcome. However, in many situations, the source code or the training data of algorithms cannot be published for industrial or intellectual property reasons, as they are the result of long and costly experience (e.g. this is typically the case in banking or insurance). We present an approach whereby individual subjects on whom automated decisions are made can elicit in a collaborative and privacy-preserving manner a rule-based approximation of the model underlying the decision algorithm, based on limited interaction with the algorithm or even only on how they have been classified. Furthermore, being rule-based, the approximation thus obtained can be used to detect potential discrimination. We present empirical work to demonstrate the practicality of our ideas.
Year
DOI
Venue
2019
10.1145/3308560.3317586
Companion Proceedings of The 2019 World Wide Web Conference
Keywords
Field
DocType
Auditing, Explainability, Machine Learning, Privacy, Transparency
Training set,Data science,Transparency (graphic),World Wide Web,Audit,Computer science,Source code,Intellectual property,Trade secret
Conference
ISBN
Citations 
PageRank 
978-1-4503-6675-5
0
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Josep Domingo-Ferrer13231404.42
Cristina Pérez-Solà2298.86
Alberto Blanco-Justicia3146.77