Abstract | ||
---|---|---|
Machine translation is one of the most classic application technologies in artificial intelligence and natural language processing. Neural machine translation models generally adopt an encoder–decoder architecture for modeling the entire translation process. However, without considering target context (e.g., decoding state) to guide the encoding, encoded source representations struggle to put great emphasis on important information for predicting some target word, yielding the weakness in generating more discriminative attentive representations across different decoding steps. Towards tackling this issue, we propose a novel encoder–refiner–decoder framework, which dynamically refines the source representations based on the generated target-side information at each decoding step. Since the refining operations are time-consuming, we propose a policy network to decide when to refine at specific decoding steps. We solve such a problem using the Gumbel-Softmax reparameterization, which makes our network differentiable and trainable through standard stochastic gradient methods. Experimental results on both Chinese–English and English–German translation tasks show that the proposed approach significantly and consistently improves translation performance over the standard encoder–decoder framework. Furthermore, when refining strategy is applied, experimental results still show a reasonable improvement over the baseline with much decrease in decoding speed. |
Year | DOI | Venue |
---|---|---|
2022 | 10.1007/s13042-022-01515-9 | International Journal of Machine Learning and Cybernetics |
Keywords | DocType | Volume |
Natural language processing, Neural machine translation, Stochastic gradient estimation, Gumbel-Softmax reparameterization | Journal | 13 |
Issue | ISSN | Citations |
8 | 1868-8071 | 0 |
PageRank | References | Authors |
0.34 | 8 | 7 |
Name | Order | Citations | PageRank |
---|---|---|---|
Xinwei Geng | 1 | 0 | 1.35 |
Longyue Wang | 2 | 72 | 18.24 |
Xing Wang | 3 | 58 | 10.07 |
Yang, Mingtao | 4 | 0 | 0.34 |
xiaocheng feng | 5 | 62 | 12.05 |
Bing Qin | 6 | 1076 | 72.82 |
Zhaopeng Tu | 7 | 518 | 39.95 |