Title
Crowdsourcing for search and data mining
Abstract
The advent of crowdsourcing is revolutionizing data annotation, evaluation, and other traditionally manual-labor intensive processes by dramatically reducing the time, cost, and effort involved. This in turn is driving a disruptive shift in search and data mining methodology in areas such as: Evaluation: the Cranfield paradigm for search evaluation requires manually assessing document relevance to search queries. Recent work on stochastic evaluation has reduced but not removed this need for manual assessment. Supervised Learning: while traditional costs associated with data annotation have driven recent machine learning work (e.g. Learning to Rank) toward greater use of unsupervised and semi-supervised methods, the emergence of crowdsourcing has made labeled data far easier to acquire, thereby driving a potential resurgence in fully-supervised methods. Applications: Crowdsourcing has introduced exciting new opportunities to integrate human labor into automated systems: handling difficult cases where automation fails, exploiting the breadth of backgrounds, geographic dispersion, real-time response, etc.
Year
DOI
Venue
2011
10.1145/1935826.1935828
SIGIR Forum
Keywords
DocType
Volume
data mining,automated system,hong kong,crowdsourcing,difficult case,data annotation,empirical method,refereed paper,fourth acm international conference,workshop proceeding,recent machine,presentation slide,data mining methodology,recent advance,onweb search,search evaluation,search,cranfield paradigm,recent work,novel application,stochastic evaluation,supervised learning,real time,machine learning,learning to rank
Conference
45
Issue
Citations 
PageRank 
1
6
0.59
References 
Authors
11
3
Name
Order
Citations
PageRank
Vitor R. Carvalho167236.38
Matthew Lease2132684.06
Emine Yilmaz3145996.39