Title
Exploiting potential of deep neural networks by layer-wise fine-grained parallelism
Abstract
Deep neural networks (DNNs) have become more and more important for big data analysis. They usually use data parallelism or model parallelism for extreme scale computing. However, the two approaches realize the performance improvement mainly by using coarse-grained parallelization schemes. Neither can fully exploit the potentials of the parallelism of many-core systems (such as GPUs) for neural network models. Here, a new fine−grained parallelism strategy (named FiLayer) is presented based on layer-wise parallelization. It has two components: inter-layer parallelism and intra-layer parallelism. The inter-layer parallelism makes several neighboring layers be processed by using a pipeline manner in a network model. For intra-layer parallelism, the operations in one layer are separated into several parts and processed concurrently. To implement above fine-grained parallelism methods, CUDA streams are used. A mathematical analysis is presented for the influence of fragment number on performance of the inter-layer parallelism, and also an analysis for the influence of CUDA stream number on the performance of the intra-layer parallelism is given. The proposed approach is realized based on Caffe. Some representative datasets including CIFAR100 and ImageNet, are applied for experiments. The evaluation results show that it can help Caffe realize remarkable speedups, which makes much sense to big data analysis.
Year
DOI
Venue
2020
10.1016/j.future.2019.07.054
Future Generation Computer Systems
Keywords
Field
DocType
Deep learning,Fine-grained parallelism,CUDA stream
CUDA,Computer science,Parallel computing,Caffè,Data parallelism,Artificial neural network,Big data,Network model,Deep neural networks,Performance improvement,Distributed computing
Journal
Volume
ISSN
Citations 
102
0167-739X
1
PageRank 
References 
Authors
0.34
0
7
Name
Order
Citations
PageRank
Wenbin Jiang135536.55
Yangsong Zhang27511.65
Pai Liu310.68
Jing Peng410.34
Laurence T. Yang56870682.61
Geyan Ye610.34
Hai Jin76544644.63