Title
Marian: Cost-effective High-Quality Neural Machine Translation in C plus
Abstract
This paper describes the submissions of the "Marian" team to the WNMT 2018 shared task. We investigate combinations of teacher-student training, low-precision matrix products, auto-tuning and other methods to optimize the Transformer model on GPU and CPU. By further integrating these methods with the new averaging attention networks, a recently introduced faster Transformer variant, we create a number of high-quality, high-performance models on the GPU and CPU, dominating the Pareto frontier for this shared task.
Year
Venue
DocType
2018
NEURAL MACHINE TRANSLATION AND GENERATION
Conference
Volume
Citations 
PageRank 
abs/1805.12096
0
0.34
References 
Authors
4
5
Name
Order
Citations
PageRank
Marcin Junczys-Dowmunt131224.24
Kenneth Heafield257939.46
Hieu Hoang3151868.35
Roman Grundkiewicz410911.75
Anthony Aue529016.87