Title
An Open Source Tool For Crowd-Sourcing The Manual Annotation Of Texts
Abstract
Manually annotated data is the basis for a large number of tasks in natural language processing as either: evaluation or training data. The annotation of large amounts of data by dedicated full-time annotators can be an expensive task, which may be beyond the budgets of many research projects. An alternative is crowd-sourcing where annotations are split among many part time annotators. This paper presents a freely available open-source platform for crowd-sourcing manual annotation tasks, and describes its application to annotating causative relations.
Year
DOI
Venue
2014
10.1007/978-3-319-09761-9_31
COMPUTATIONAL PROCESSING OF THE PORTUGUESE LANGUAGE
Keywords
Field
DocType
crowd-sourcing, annotations, causative relations
Training set,Annotation,Information retrieval,Computer science,Manual annotation,Natural language processing,Artificial intelligence
Conference
Volume
ISSN
Citations 
8775
0302-9743
0
PageRank 
References 
Authors
0.34
11
6