Title
Tag Dictionaries Accelerate Manual Annotation
Abstract
Expert human input can contribute in various ways to facilitate automatic annotation of natural language text. For example, a part-of-speech tagger can be trained on labeled input provided offline by experts. In addition, expert input can be solicited by way of active learning to make the most of annotator expertise. However, hiring individuals to perform manual annotation is costly both in terms of money and time. This paper reports on a user study that was performed to determine the degree of effect that a part-of-speech dictionary has on a group of subjects performing the annotation task. The user study was conducted using a modular, web-based interface created specifically for text annotation tasks. The user study found that for both native and non-native English speakers a dictionary with greater than 60% coverage was effective at reducing annotation time and increasing annotator accuracy. On the basis of this study, we predict that using a part-of-speech tag dictionary with coverage greater than 60% can reduce the cost of annotation in terms of both time and money.
Year
Venue
Keywords
2010
LREC 2010 - SEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION
part of speech,active learning
Field
DocType
Citations 
Annotation,Active learning,Information retrieval,Computer science,Manual annotation,Image retrieval,Natural language,Natural language processing,Artificial intelligence,Modular design,Text annotation
Conference
4
PageRank 
References 
Authors
0.44
8
8
Name
Order
Citations
PageRank
Marc Carmen1523.51
Paul Felt2235.35
Robbie Haertel3817.19
Deryle Lonsdale410613.95
Peter McClanahan5664.81
Owen Merkling680.93
Eric K. Ringger727239.24
Kevin D. Seppi833541.46