Title
Information Measures, Experiments, Multi-category Hypothesis Tests, and Surrogate Losses.
Abstract
We provide a unifying view of statistical information measures, multi-class classification problems, multi-way Bayesian hypothesis testing, and loss functions, elaborating equivalence results between all of these objects. In particular, we consider a particular generalization of $f$-divergences to multiple distributions, and we show that there is a constructive equivalence between $f$-divergences, statistical information (in the sense of uncertainty as elaborated by DeGroot), and loss functions for multi-category classification. We also study an extension of our results to multi-class classification problems in which we must both infer a discriminant function $gamma$ and a data representation (or, in the setting of a hypothesis testing problem, an experimental design), represented by a quantizer $mathsf{q}$ from a family of possible quantizers $mathsf{Q}$. There, we give a complete characterization of the equivalence between loss functions, meaning that optimizing either of two losses yields the same optimal discriminant and quantizer $mathsf{q}$. A main consequence of our results is to describe those convex loss functions that are Fisher consistent for jointly choosing a data representation and minimizing the (weighted) probability of error in multi-category classification and hypothesis testing problems.
Year
Venue
Field
2016
arXiv: Statistics Theory
Discrete mathematics,External Data Representation,Constructive,Discriminant,Regular polygon,Equivalence (measure theory),Statistics,Quantization (signal processing),Statistical hypothesis testing,Mathematics,Discriminant function analysis
DocType
Volume
Citations 
Journal
abs/1603.00126
0
PageRank 
References 
Authors
0.34
5
3
Name
Order
Citations
PageRank
John C. Duchi1164898.77
Khashayar Khosravi262.45
Feng Ruan331.75