Title
Sequence Level Training with Recurrent Neural Networks
Abstract
Many natural language processing applications use language models to generate text. These models are typically trained to predict the next word in a sequence, given the previous words and some context such as an image. However, at test time the model is expected to generate the entire sequence from scratch. This discrepancy makes generation brittle, as errors may accumulate along the way. We address this issue by proposing a novel sequence level training algorithm that directly optimizes the metric used at test time, such as BLEU or ROUGE. On three different tasks, our approach outperforms several strong baselines for greedy generation. The method is also competitive when these baselines employ beam search, while being several times faster.
Year
Venue
Field
2015
international conference on learning representations
Scratch,Computer science,Beam search,Recurrent neural network,Artificial intelligence,Language model,Machine learning
DocType
Volume
Citations 
Journal
abs/1511.06732
192
PageRank 
References 
Authors
4.79
19
4
Search Limit
100192
Name
Order
Citations
PageRank
Marc'Aurelio Ranzato15242470.94
Sumit Chopra22835181.37
Michael Auli3106153.54
Wojciech Zaremba42733117.55