Title
The New Large-Scale RNNLM System Based on Distributed Neuron
Abstract
RNNLM (Recurrent Neural Network Language Model) can save the historical information of the training dataset by the last hidden layer and can also as input for training. It has become an interesting topic in the field of Natural Language Processing research. However, the immense training time overhead is a big problem. The large output layer, hidden layer, last hidden layer and the connections among them will generate enormous matrix in training. It is the main facts to influence the efficiency and scalability. At the same time, output layer class and small hidden layer should decrease the accuracy of RNNLM. In general, the lack of parallel for artificial neuron is main reason for these. We change the structure of RNNLM and design the new large-scale RNNLM by the center of distributed artificial neurons in hidden layer to stimulate the parallel characteristic of biological neuron system. Meanwhile, we change training method, and present the coordination strategy for distributed neuron. At last, the prototype of new large-scale RNNLM system is implemented based on Spark. The testing and analysis results show that the training time overhead is far less than the growth rate of the distributed neuron in hidden layer and size of training dataset. These results show our large-scale RNNLM system has efficiency and scalability advantage.
Year
DOI
Venue
2017
10.1109/IPDPSW.2017.21
2017 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW)
Keywords
Field
DocType
Recurrent Neural Network Language Model,Distributed Computing,Distributed System,Spark
Spark (mathematics),Computer science,Artificial neuron,Parallel computing,Server,Recurrent neural network,Artificial intelligence,Artificial neural network,Machine learning,Language model,Distributed computing,Scalability
Conference
ISSN
ISBN
Citations 
2164-7062
978-1-5386-3409-7
0
PageRank 
References 
Authors
0.34
2
6
Name
Order
Citations
PageRank
DeJiao Niu175.41
Rui Xue200.34
Tao Cai3106.57
Hai Li42435208.37
Kingsley Effah500.34
hang zhang63116.05