Title
Large Margin Discriminant Dimensionality Reduction in Prediction Space.
Abstract
In this paper we establish a duality between boosting and SVM, and use this to derive a novel discriminant dimensionality reduction algorithm. In particular, using the multiclass formulation of boosting and SVM we note that both use a combination of mapping and linear classification to maximize the multiclass margin. In SVM this is implemented using a pre-defined mapping (induced by the kernel) and optimizing the linear classifiers. In boosting the linear classifiers are pre-defined and the mapping (predictor) is learned through combination of weak learners. We argue that the intermediate mapping, e.g. boosting predictor, is preserving the discriminant aspects of the data and by controlling the dimension of this mapping it is possible to achieve discriminant low dimensional representations for the data. We use the aforementioned duality and propose a new method, Large Margin Discriminant Dimensionality Reduction (LADDER) that jointly learns the mapping and the linear classifiers in an efficient manner. This leads to a data-driven mapping which can embed data into any number of dimensions. Experimental results show that this embedding can significantly improve performance on tasks such as hashing and image/scene classification.
Year
Venue
Field
2016
NIPS
Dimensionality reduction,Computer science,Multiple discriminant analysis,Artificial intelligence,Kernel (linear algebra),Optimal discriminant analysis,Mathematical optimization,Pattern recognition,Support vector machine,Boosting (machine learning),Linear discriminant analysis,Linear classifier,Machine learning
DocType
Citations 
PageRank 
Conference
0
0.34
References 
Authors
0
4
Name
Order
Citations
PageRank
Mohammad J. Saberian126612.11
Jose Costa Pereira268717.58
Nuno Vasconcelos35410273.99
Can Xu442.14