Title
The Coherent Loss Function for Classification.
Abstract
A prediction rule in binary classification that aims to achieve the lowest probability of misclassification involves minimizing over a non-convex, 0-1 loss function, which is typically a computationally intractable optimization problem. To address the intractability, previous methods consider minimizing the cumulative loss – the sum of convex surrogates of the 0-1 loss of each sample. In this paper, we revisit this paradigm and develop instead an axiomatic framework by proposing a set of salient properties on functions for binary classification and then propose the coherent loss approach, which is a tractable upper-bound of the empirical classification error over the entire sample set. We show that the proposed approach yields a strictly tighter approximation to the empirical classification error than any convex cumulative loss approach while preserving the convexity of the underlying optimization problem, and this approach for binary classification also has a robustness interpretation which builds a connection to robust SVMs. The experimental results show that our approach outperforms the standard SVM when additional constraints are imposed.
Year
Venue
Field
2014
ICML
Mathematical optimization,Convexity,Hinge loss,Binary classification,Axiom,Computer science,Support vector machine,Regular polygon,Robustness (computer science),Artificial intelligence,Optimization problem,Machine learning
DocType
Citations 
PageRank 
Conference
2
0.40
References 
Authors
9
3
Name
Order
Citations
PageRank
Wenzhuo Yang1143.02
Melvyn Sim21909117.68
Xu, Huan3111671.73