Title
Listwise Learning to Rank from Crowds.
Abstract
Learning to rank has received great attention in recent years as it plays a crucial role in many applications such as information retrieval and data mining. The existing concept of learning to rank assumes that each training instance is associated with a reliable label. However, in practice, this assumption does not necessarily hold true as it may be infeasible or remarkably expensive to obtain reliable labels for many learning to rank applications. Therefore, a feasible approach is to collect labels from crowds and then learn a ranking function from crowdsourcing labels. This study explores the listwise learning to rank with crowdsourcing labels obtained from multiple annotators, who may be unreliable. A new probabilistic ranking model is first proposed by combining two existing models. Subsequently, a ranking function is trained by proposing a maximum likelihood learning approach, which estimates ground-truth labels and annotator expertise, and trains the ranking function iteratively. In practical crowdsourcing machine learning, valuable side information (e.g., professional grades) about involved annotators is normally attainable. Therefore, this study also investigates learning to rank from crowd labels when side information on the expertise of involved annotators is available. In particular, three basic types of side information are investigated, and corresponding learning algorithms are consequently introduced. Further, the top-k learning to rank from crowdsourcing labels are explored to deal with long training ranking lists. The proposed algorithms are tested on both synthetic and real-world data. Results reveal that the maximum likelihood estimation approach significantly outperforms the average approach and existing crowdsourcing regression methods. The performances of the proposed algorithms are comparable to those of the learning model in consideration reliable labels. The results of the investigation further indicate that side information is helpful in inferring both ranking functions and expertise degrees of annotators.
Year
DOI
Venue
2016
10.1145/2910586
TKDD
Keywords
Field
DocType
Listwise learning to rank,crowdsourcing,multiple annotators,probabilistic ranking model,side information
Data mining,Crowds,Learning to rank,Ranking,Ranking SVM,Regression,Crowdsourcing,Computer science,Concept learning,Artificial intelligence,Probabilistic logic,Machine learning
Journal
Volume
Issue
ISSN
11
1
1556-4681
Citations 
PageRank 
References 
1
0.35
33
Authors
5
Name
Order
Citations
PageRank
Ou Wu129722.76
Qiang You272.17
Fen Xia3422.26
Lei Ma410.35
Weiming Hu55300261.38