Title
Large-Scale Bayesian Multi-Label Learning via Topic-Based Label Embeddings
Abstract
We present a scalable Bayesian multi-label learning model based on learning low-dimensional label embeddings. Our model assumes that each label vector is generated as a weighted combination of a set of topics (each topic being a distribution over labels), where the combination weights (i.e., the embeddings) for each label vector are conditioned on the observed feature vector. This construction, coupled with a Bernoulli-Poisson link function for each label of the binary label vector, leads to a model with a computational cost that scales in the number of positive labels in the label matrix. This makes the model particularly appealing for real-world multi-label learning problems where the label matrix is usually very massive but highly sparse. Using a data-augmentation strategy leads to full local conjugacy in our model, facilitating simple and very efficient Gibbs sampling, as well as an Expectation Maximization algorithm for inference. Also, predicting the label vector at test time does not require doing an inference for the label embeddings and can be done in closed form. We report results on several benchmark data sets, comparing our model with various state-of-the art methods.
Year
Venue
Field
2015
Annual Conference on Neural Information Processing Systems
Feature vector,Pattern recognition,Matrix (mathematics),Computer science,Inference,Expectation–maximization algorithm,Artificial intelligence,Gibbs sampling,Machine learning,Binary number,Scalability,Bayesian probability
DocType
Volume
ISSN
Conference
28
1049-5258
Citations 
PageRank 
References 
10
0.49
20
Authors
4
Name
Order
Citations
PageRank
Piyush Rai160436.79
Changwei Hu2575.95
Ricardo Henao328623.85
L. Carin44603339.36