Abstract | ||
---|---|---|
Identifying the labels of points of interest (POIs), aka POI labelling, provides significant benefits in location-based services. However, the quality of raw labels manually added by users or generated by artificial algorithms cannot be guaranteed. Such low-quality labels decrease the usability and result in bad user experiences. In this paper, by observing that crowdsourcing is a best-fit for computer-hard tasks, we leverage crowdsourcing to improve the quality of POI labelling. To our best knowledge, this is the first work on crowdsourced POI labelling tasks. In particular, there are two sub-problems: (1) how to infer the correct labels for each POI based on workers' answers, and (2) how to effectively assign proper tasks to workers in order to make more accurate inference for next available workers. To address these two problems, we propose a framework consisting of an inference model and an online task assigner. The inference model measures the quality of a worker on a POI by elaborately exploiting (i) worker's inherent quality, (ii) the spatial distance between the worker and the POI, and (iii) the POI influence, which can provide reliable inference results once a worker submits an answer. As workers are dynamically coming, the online task assigner judiciously assigns proper tasks to them so as to benefit the inference. The inference model and task assigner work alternately to continuously improve the overall quality. We conduct extensive experiments on a real crowdsourcing platform, and the results on two real datasets show that our method significantly outperforms state-of-the-art approaches. |
Year | DOI | Venue |
---|---|---|
2016 | 10.1109/ICDE.2016.7498229 | 2016 IEEE 32nd International Conference on Data Engineering (ICDE) |
Keywords | Field | DocType |
location-aware result inference,task assignment,points of interest,location-based services,raw labels,artificial algorithms,low-quality labels,bad user experiences,best-fit,computer-hard tasks,crowdsourced POI labelling tasks,correct labels,online task assigner,inference model measures,spatial distance,POI influence,crowdsourcing platform | Data mining,Computer science,Inference,Crowdsourcing,Usability,Labelling,Point of interest,Location aware,Spatial database,Database,AKA | Conference |
ISSN | Citations | PageRank |
1084-4627 | 29 | 0.74 |
References | Authors | |
23 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Huiqi Hu | 1 | 79 | 6.22 |
Yudian Zheng | 2 | 418 | 16.91 |
Zhifeng Bao | 3 | 683 | 62.90 |
Guoliang Li | 4 | 3077 | 154.70 |
Jianhua Feng | 5 | 2713 | 121.30 |
Reynold Cheng | 6 | 3069 | 154.13 |