Title
Multiple-Instance Learning from Distributions.
Abstract
We propose a new theoretical framework for analyzing the multiple-instance learning (MIL) setting. In MIL, training examples are provided to a learning algorithm in the form of labeled sets, or "bags," of instances. Applications of MIL include 3-D quantitative structure activity relationship prediction for drug discovery and content-based image retrieval for web search. The goal of an algorithm is to learn a function that correctly labels new bags or a function that correctly labels new instances. We propose that bags should be treated as latent distributions from which samples are observed. We show that it is possible to learn accurate instance- and bag-labeling functions in this setting as well as functions that correctly rank bags or instances under weak assumptions. Additionally, our theoretical results suggest that it is possible to learn to rank efficiently using traditional, well-studied "supervised" learning approaches. We perform an extensive empirical evaluation that supports the theoretical predictions entailed by the new framework. The proposed theoretical framework leads to a better understanding of the relationship between the MI and standard supervised learning settings, and it provides new methods for learning from MI data that are more accurate, more efficient, and have better understood theoretical properties than existing MI-specific algorithms.
Year
Venue
Keywords
2016
JOURNAL OF MACHINE LEARNING RESEARCH
multiple-instance learning,learning theory,ranking,classification
Field
DocType
Volume
Stability (learning theory),Instance-based learning,Semi-supervised learning,Pattern recognition,Ranking,Computer science,Learning theory,Image retrieval,Supervised learning,Artificial intelligence,Machine learning
Journal
17
ISSN
Citations 
PageRank 
1532-4435
0
0.34
References 
Authors
0
2
Name
Order
Citations
PageRank
Gary Doran1313.75
Soumya Ray2948.89