Title
A Convolutional Encoder Model For Neural Machine Translation
Abstract
The prevalent approach to neural machine translation relies on bi-directional LSTMs to encode the source sentence. We present a faster and simpler architecture based on a succession of convolutional layers. This allows to encode the source sentence simultaneously compared to recurrent networks for which computation is constrained by temporal dependencies. On WMT'16 EnglishRomanian translation we achieve competitive accuracy to the state-of-the-art and on WMT'15 English-German we outperform several recently published results. Our models obtain almost the same accuracy as a very deep LSTM setup on WMT'14 English-French translation. We speed up CPU decoding by more than two times at the same or higher accuracy as a strong bidirectional LSTM.
Year
DOI
Venue
2017
10.18653/v1/P17-1012
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1
DocType
Volume
Citations 
Conference
abs/1611.02344
10
PageRank 
References 
Authors
0.56
0
4
Name
Order
Citations
PageRank
Jonas Gehring11487.87
Michael Auli2106153.54
David Grangier381641.60
Dauphin, Yann N.497949.26