Title
Training Language Models Using Target-Propagation.
Abstract
While Truncated Back-Propagation through Time (BPTT) is the most popular approach to training Recurrent Neural Networks (RNNs), it suffers from being inherently sequential (making parallelization difficult) and from truncating gradient flow between distant time-steps. We investigate whether Target Propagation (TPROP) style approaches can address these shortcomings. Unfortunately, extensive experiments suggest that TPROP generally underperforms BPTT, and we end with an analysis of this phenomenon, and suggestions for future work.
Year
Venue
Field
2017
arXiv: Computation and Language
Computer science,Recurrent neural network,Artificial intelligence,Language model,Machine learning
DocType
Volume
Citations 
Journal
abs/1702.04770
1
PageRank 
References 
Authors
0.39
9
7
Name
Order
Citations
PageRank
Sam Wiseman11019.02
Sumit Chopra22835181.37
Marc'Aurelio Ranzato35242470.94
Arthur Szlam4105668.60
Ruoyu Sun529616.15
Soumith Chintala62056102.09
Nicolas Vasilache7414.24