Title
Learning representations for weakly supervised natural language processing tasks
Abstract
Finding the right representations for words is critical for building accurate NLP systems when domain-specific labeled data for the task is scarce. This article investigates novel techniques for extracting features from n-gram models, Hidden Markov Models, and other statistical language models, including a novel Partial Lattice Markov Random Field model. Experiments on part-of-speech tagging and information extraction, among other tasks, indicate that features taken from statistical language models, in combination with more traditional features, outperform traditional representations alone, and that graphical model representations outperform n-gram models, especially on sparse and polysemous words.
Year
DOI
Venue
2014
10.1162/COLI_a_00167
Computational Linguistics
Field
DocType
Volume
Markov random field,Computer science,Information extraction,Artificial intelligence,Natural language processing,Graphical model,Labeled data,Hidden Markov model,Language model,Machine learning
Journal
40
Issue
ISSN
Citations 
1
0891-2017
23
PageRank 
References 
Authors
0.95
75
6
Name
Order
Citations
PageRank
Fei Huang150656.44
Arun Ahuja2727.45
Doug Downey31908119.79
Yi Yang4271.68
Yuhong Guo577449.28
Alexander Yates689851.53