Title
Model-Distributed DNN Training for Memory-Constrained Edge Computing Devices
Abstract
We consider a model-distributed learning framework in which layers of a deep learning model is distributed across multiple workers. To achieve consistent gradient updates during the training phase, model-distributed learning requires the storage of multiple versions of the layer parameters at every worker. In this paper, we design mcPipe to reduce the memory cost of model-distributed learning, whi...
Year
DOI
Venue
2021
10.1109/LANMAN52105.2021.9478829
2021 IEEE International Symposium on Local and Metropolitan Area Networks (LANMAN)
Keywords
DocType
ISBN
Training,Learning systems,Performance evaluation,Metropolitan area networks,Computational modeling,Neural networks,Memory management
Conference
978-1-6654-4579-5
Citations 
PageRank 
References 
0
0.34
0
Authors
4
Name
Order
Citations
PageRank
Pengzhen Li100.68
Hulya Seferoglu242628.46
Venkat R. Dasari300.34
Erdem Koyuncu401.35