Abstract | ||
---|---|---|
Neural Machine Translation (NMT) systems are known to degrade when confronted with noisy data, especially when the system is trained only on clean data. In this paper, we show that augmenting training data with sentences containing artificially-introduced grammatical errors can make the system more robust to such errors. In combination with an automatic grammar error correction system, we can recover 1.5 BLEU out of 2.4 BLEU lost due to grammatical errors. We also present a set of Spanish translations of the JFLEG grammar error correction corpus, which allows for testing NMT robustness to real grammatical errors. |
Year | Venue | Field |
---|---|---|
2018 | arXiv: Computation and Language | Training set,BLEU,Noisy data,Computer science,Machine translation,Error detection and correction,Grammar,Robustness (computer science),Natural language processing,Artificial intelligence |
DocType | Volume | Citations |
Journal | abs/1808.06267 | 0 |
PageRank | References | Authors |
0.34 | 23 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Alison Lui | 1 | 0 | 0.68 |
Antonios Anastasopoulos | 2 | 122 | 17.13 |
David Chiang | 3 | 2843 | 144.76 |