Title
Parallelizing DNN inference in mobile web browsers on heterogeneous hardware
Abstract
BSTRACTMobile Web apps are emerging to leverage DNN models to provide intelligent user experience. But the limited functionalities for heterogeneous hardware provided by mobile Web browsers challenge the Web apps to perform DNN inference efficiently. In this paper, we propose a novel DNN inference engine, named PipeEngine, to parallelize the DNN inference process on CPU and GPU in mobile Web browsers. The design of PipeEngine enables pipeline parallelism between two adjacent DNN inference tasks with heterogeneous hardware. Evaluation results show that PipeEngine can increase the inference throughput by up to 2.77×.
Year
DOI
Venue
2022
10.1145/3498361.3538763
Mobile Systems, Applications, and Services
DocType
Citations 
PageRank 
Conference
0
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Deyu Tian100.34
Haiyang Shen200.34
Yun Ma321620.25