Title | ||
---|---|---|
Information Measures, Experiments, Multi-category Hypothesis Tests, and Surrogate Losses. |
Abstract | ||
---|---|---|
We provide a unifying view of statistical information measures, multi-class classification problems, multi-way Bayesian hypothesis testing, and loss functions, elaborating equivalence results between all of these objects. In particular, we consider a particular generalization of $f$-divergences to multiple distributions, and we show that there is a constructive equivalence between $f$-divergences, statistical information (in the sense of uncertainty as elaborated by DeGroot), and loss functions for multi-category classification. We also study an extension of our results to multi-class classification problems in which we must both infer a discriminant function $gamma$ and a data representation (or, in the setting of a hypothesis testing problem, an experimental design), represented by a quantizer $mathsf{q}$ from a family of possible quantizers $mathsf{Q}$. There, we give a complete characterization of the equivalence between loss functions, meaning that optimizing either of two losses yields the same optimal discriminant and quantizer $mathsf{q}$. A main consequence of our results is to describe those convex loss functions that are Fisher consistent for jointly choosing a data representation and minimizing the (weighted) probability of error in multi-category classification and hypothesis testing problems. |
Year | Venue | Field |
---|---|---|
2016 | arXiv: Statistics Theory | Discrete mathematics,External Data Representation,Constructive,Discriminant,Regular polygon,Equivalence (measure theory),Statistics,Quantization (signal processing),Statistical hypothesis testing,Mathematics,Discriminant function analysis |
DocType | Volume | Citations |
Journal | abs/1603.00126 | 0 |
PageRank | References | Authors |
0.34 | 5 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
John C. Duchi | 1 | 1648 | 98.77 |
Khashayar Khosravi | 2 | 6 | 2.45 |
Feng Ruan | 3 | 3 | 1.75 |