Abstract | ||
---|---|---|
This work shows how to improve state-of-the-art monolingual natural language processing models using unannotated bilingual text. We build a multiview learning objective that enforces agreement between monolingual and bilingual models. In our method the first, monolingual view consists of supervised predictors learned separately for each language. The second, bilingual view consists of log-linear predictors learned over both languages on bilingual text. Our training procedure estimates the parameters of the bilingual model using the output of the monolingual model, and we show how to combine the two models to account for dependence between views. For the task of named entity recognition, using bilingual predictors increases F1 by 16.1% absolute over a supervised monolingual model, and retraining on bilingual predictions increases monolingual model F1 by 14.6%. For syntactic parsing, our bilingual predictor increases F1 by 2.1% absolute, and retraining a monolingual model on its output gives an improvement of 2.0%. |
Year | Venue | Keywords |
---|---|---|
2010 | CoNLL | monolingual view,supervised monolingual model,bilingual text,unannotated bilingual text,bilingual predictor,monolingual model,bilingual model,bilingual prediction,state-of-the-art monolingual natural language,bilingual view,natural language processing |
Field | DocType | Citations |
Syntactic parsing,Computer science,Multiview learning,Speech recognition,Natural language processing,Artificial intelligence,Named-entity recognition,Retraining | Conference | 23 |
PageRank | References | Authors |
0.90 | 23 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
David Burkett | 1 | 190 | 8.73 |
Slav Petrov | 2 | 2405 | 107.56 |
John Blitzer | 3 | 3330 | 193.45 |
Dan Klein | 4 | 8083 | 495.21 |