Abstract | ||
---|---|---|
Metrics specifying distances between data points can be learned in a discriminative manner or from generative models. In this paper, we show how to unify generative and discriminative learning of metrics via a kernel learning framework. Specifically, we learn local metrics optimized from parametric generative models. These are then used as base kernels to construct a global kernel that minimizes a discriminative training criterion. We consider both linear and nonlinear combinations of local metric kernels. Our empirical results show that these combinations significantly improve performance on classification tasks. The proposed learning algorithm is also very efficient, achieving order of magnitude speedup in training time compared to previous discriminative baseline methods. |
Year | Venue | Keywords |
---|---|---|
2011 | CoRR | discrimination learning,artificial intelligent |
Field | DocType | Volume |
Online machine learning,Instance-based learning,Stability (learning theory),Semi-supervised learning,Multi-task learning,Active learning (machine learning),Unsupervised learning,Artificial intelligence,Discriminative model,Machine learning,Mathematics | Journal | abs/1109.3940 |
Citations | PageRank | References |
2 | 0.39 | 14 |
Authors | ||
4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Yuan Shi | 1 | 6 | 2.24 |
Yung-Kyun Noh | 2 | 39 | 8.62 |
Fei Sha | 3 | 3429 | 240.64 |
Daniel D. Lee | 4 | 1136 | 109.20 |