Title
Deep Metric Learning with Data Summarization.
Abstract
We present Deep Stochastic Neighbor Compression DSNC, a framework to compress training data for instance-based methods such as k-nearest neighbors. We accomplish this by inferring a smaller set of pseudo-inputs in a new feature space learned by a deep neural network. Our framework can equivalently be seen as jointly learning a nonlinear distance metric induced by the deep feature space and learning a compressed version of the training data. In particular, compressing the data in a deep feature space makes DSNC robust against label noise and issues such as within-class multi-modal distributions. This leads to DSNC yielding better accuracies and faster predictions at test time, as compared to other competing methods. We conduct comprehensive empirical evaluations, on both quantitative and qualitative tasks, and on several benchmark datasets, to show its effectiveness as compared to several baselines.
Year
DOI
Venue
2016
10.1007/978-3-319-46128-1_49
ECML/PKDD
Field
DocType
Citations 
Training set,Data mining,Automatic summarization,Feature vector,Stochastic gradient descent,Nonlinear system,Convolutional neural network,Computer science,Metric (mathematics),Artificial intelligence,Artificial neural network,Machine learning
Conference
2
PageRank 
References 
Authors
0.37
22
5
Name
Order
Citations
PageRank
Wenlin Wang1517.06
Changyou Chen236536.95
Wenlin Chen335823.45
Piyush Rai460436.79
L. Carin54603339.36