Title
Distilled Split Deep Neural Networks for Edge-Assisted Real-Time Systems
Abstract
Offloading the execution of complex Deep Neural Networks (DNNs) models to compute-capable devices at the network edge, that is, edge servers, can significantly reduce capture-to-output delay. However, the communication link between the mobile devices and edge servers can become the bottleneck when channel conditions are poor. We propose a framework to split DNNs for image processing and minimize capture-to-output delay in a wide range of network conditions and computing parameters. The core idea is to split the DNN models into head and tail models, where the two sections are deployed at the mobile device and edge server, respectively. Different from prior literature presenting DNN splitting frameworks, we distill the architecture of the head DNN to reduce its computational complexity and introduce a bottleneck, thus minimizing processing load at the mobile device as well as the amount of wirelessly transferred data. Our results show 98% reduction in used bandwidth and 85% in computation load compared to straightforward splitting.
Year
DOI
Venue
2019
10.1145/3349614.3356022
Proceedings of the 2019 Workshop on Hot Topics in Video Analytics and Intelligent Edges
Keywords
Field
DocType
deep neural networks, edge computing, network distillation
Computer science,Real-time computing,Deep neural networks
Conference
ISBN
Citations 
PageRank 
978-1-4503-6928-2
4
0.40
References 
Authors
0
5
Name
Order
Citations
PageRank
Yoshitomo Matsubara1103.20
Sabur Baidya2123.04
Davide Callegaro382.47
Marco Levorato4323.55
Sameer Singh5106071.63