Title
Friends Don't Let Friends Deploy Black-Box Models: The Importance of Intelligibility in Machine Learning
Abstract
Every data set is flawed, often in ways that are unanticipated and difficult to detect. If you can't understand what your model has learned, then you almost certainly are shipping models that are less accurate than they could be and which might even be risky. Historically there has been a tradeoff between accuracy and intelligibility: accurate models such as neural nets, boosted tress and random forests are not very intelligible, and intelligible models such as logistic regression and small trees or decision lists usually are less accurate. In mission-critical domains such as healthcare, where being able to understand, validate, edit and ultimately trust a model is important, one often had to choose less accurate models. But this is changing. We have developed a learning method based on generalized additive models with pairwise interactions (GA2Ms) that is as accurate as full complexity models yet even more interpretable than logistic regression. In this talk I'll highlight the kinds of problems that are lurking in all of our datasets, and how these interpretable, high-performance GAMs are making what was previously hidden, visible. I'll also show how we're using these models to uncover bias in models where fairness and transparency are important. (Code for the models has recently been released open-source.)
Year
DOI
Field
2019
10.1145/3292500.3340414
Black box (phreaking),Pairwise comparison,Computer science,Decision list,Artificial intelligence,Artificial neural network,Random forest,Generalized additive model,Logistic regression,Machine learning,Intelligibility (communication)
DocType
ISSN
ISBN
Conference
978-1-4503-6201-6
978-1-4503-6201-6
Citations 
PageRank 
References 
0
0.34
0
Authors
1
Name
Order
Citations
PageRank
Rich Caruana14503655.71