Title
Interpretation of linear classifiers by means of feature relevance bounds.
Abstract
Research on feature relevance and feature selection problems goes back several decades, but the importance of these areas continues to grow as more and more data becomes available, and machine learning methods are used to gain insight and interpret, rather than solely to solve classification or regression problems. Despite the fact that feature relevance is often discussed, it is frequently poorly defined, and the feature selection problems studied are subtly different. Furthermore, the problem of finding all features relevant for a classification problem has only recently started to gain traction, despite its importance for interpretability and integrating expert knowledge. In this paper, we attempt to unify commonly used concepts and to give an overview of the main questions and results. We formalize two interpretations of the all-relevant problem and propose a polynomial method to approximate one of them for the important hypothesis class of linear classifiers, which also enables a distinction between strongly and weakly relevant features.
Year
DOI
Venue
2018
10.1016/j.neucom.2017.11.074
Neurocomputing
Keywords
Field
DocType
Feature relevance,Feature selection,Interpretability,All-relevant,Linear classification
Interpretability,Pattern recognition,Feature selection,Random subspace method,Feature relevance,Artificial intelligence,Regression problems,Mathematics,Machine learning,Polynomial method
Journal
Volume
ISSN
Citations 
298
0925-2312
2
PageRank 
References 
Authors
0.39
14
4
Name
Order
Citations
PageRank
Christina Göpfert121.75
Lukas Pfannschmidt220.39
Jan Philip Göpfert341.78
Barbara Hammer474.52