Title
Efficient Transfer Learning for Neural Network Language Models.
Abstract
We apply transfer learning techniques to create topically and/or stylistically biased natural language models from small data samples, given generic long short-term memory (LSTM) language models trained on larger data sets. Although LSTM language models are powerful tools with wide-ranging applications, they require enormous amounts of data and time to train. Thus, we build general purpose language models that take advantage of large standing corpora and computational resources proactively, allowing us to build more specialized analytical tools from smaller data sets on demand. We show that it is possible to construct a language model from a small, focused corpus by first training an LSTM language model on a large corpus (e.g., the text from English Wikipedia) and then retraining only the internal transition model parameters on the smaller corpus. We also show that a single general language model can be reused through transfer learning to create many distinct special purpose language models quickly with modest amounts of data.
Year
DOI
Venue
2018
10.5555/3382225.3382415
ASONAM '18: International Conference on Advances in Social Networks Analysis and Mining Barcelona Spain August, 2018
Field
DocType
ISBN
Data set,Small data,Computer science,Neural network language models,Transfer of learning,Natural language,Artificial intelligence,Machine learning,Language model,Retraining,General-purpose language
Conference
978-1-5386-6051-5
Citations 
PageRank 
References 
0
0.34
0
Authors
5
Name
Order
Citations
PageRank
Jacek Skryzalin131.23
Hamilton E. Link2405.31
Jeremy D. Wendt332.13
Richard Field431.77
Samuel N. Richter501.35