Title | ||
---|---|---|
Morpho-syntactic information for automatic error analysis of statistical machine translation output |
Abstract | ||
---|---|---|
Evaluation of machine translation output is an important but difficult task. Over the last years, a variety of automatic evaluation measures have been studied, some of them like Word Error Rate (WER), Position Independent Word Error Rate (PER) and BLEU and NIST scores have become widely used tools for comparing different systems as well as for evaluating improvements within one system. However, these measures do not give any details about the nature of translation errors. Therefore some analysis of the generated output is needed in order to identify the main problems and to focus the research efforts. On the other hand, human evaluation is a time consuming and expensive task. In this paper, we investigate methods for using of morpho-syntactic information for automatic evaluation: standard error measures WER and PER are calculated on distinct word classes and forms in order to get a better idea about the nature of translation errors and possibilities for improvements. |
Year | Venue | Field |
---|---|---|
2006 | WMT@HLT-NAACL | BLEU,Computer science,Evaluation of machine translation,Word error rate,Machine translation,Speech recognition,Morpho,NIST,Syntax,Standard error |
DocType | Citations | PageRank |
Conference | 19 | 1.02 |
References | Authors | |
13 | 8 |
Name | Order | Citations | PageRank |
---|---|---|---|
Maja Popović | 1 | 169 | 13.09 |
Hermann Ney | 2 | 14178 | 1506.93 |
Adrià de Gispert | 3 | 472 | 35.22 |
José B. Mariño | 4 | 510 | 64.66 |
Deepa Gupta | 5 | 75 | 11.91 |
marcello federico | 6 | 2420 | 179.56 |
Patrik Lambert | 7 | 277 | 23.36 |
Rafael Banchs | 8 | 120 | 8.91 |