Title
A general and multi-lingual phrase chunking model based on masking method
Abstract
Several phrase chunkers have been proposed over the past few years. Some state-of-the-art chunkers achieved better performance via integrating external resources, e.g., parsers and additional training data, or combining multiple learners. However, in many languages and domains, such external materials are not easily available and the combination of multiple learners will increase the cost of training and testing. In this paper, we propose a mask method to improve the chunking accuracy. The experimental results show that our chunker achieves better performance in comparison with other deep parsers and chunkers. For CoNLL-2000 data set, our system achieves 94.12 in F rate. For the base-chunking task, our system reaches 92.95 in F rate. When porting to Chinese, the performance of the base-chunking task is 92.36 in F rate. Also, our chunker is quite efficient. The complete chunking time of a 50K words document is about 50 seconds.
Year
DOI
Venue
2006
10.1007/11671299_17
CICLing
Keywords
Field
DocType
chunking accuracy,additional training data,base-chunking task,masking method,phrase chunkers,multi-lingual phrase,conll-2000 data,multiple learner,f rate,better performance,complete chunking time,state-of-the-art chunkers
Noun phrase,Masking (art),Computer science,Phrase chunking,Phrase,Speech recognition,Polynomial kernel,Artificial intelligence,Chunking (psychology),Natural language processing,Porting,Parsing
Conference
Volume
ISSN
ISBN
3878
0302-9743
3-540-32205-1
Citations 
PageRank 
References 
17
0.85
17
Authors
3
Name
Order
Citations
PageRank
Yu-Chieh Wu124723.16
Chia-Hui Chang2106264.41
Yue-Shi Lee354341.14