Title
The Art of Getting Deep Neural Networks in Shape.
Abstract
Training a deep neural network (DNN) involves selecting a set of hyperparameters that define the network topology and influence the accuracy of the resulting network. Often, the goal is to maximize prediction accuracy on a given dataset. However, non-functional requirements of the trained network -- such as inference speed, size, and energy consumption -- can be very important as well. In this article, we aim to automate the process of selecting an appropriate DNN topology that fulfills both functional and non-functional requirements of the application. Specifically, we focus on tuning two important hyperparameters, depth and width, which together define the shape of the resulting network and directly affect its accuracy, speed, size, and energy consumption. To reduce the time needed to search the design space, we train a fraction of DNNs and build a model to predict the performances of the remaining ones. We are able to produce tuned ResNets, which are up to 4.22 times faster than original depth-scaled ResNets on a batch of 128 images while matching their accuracy.
Year
DOI
Venue
2019
10.1145/3291053
TACO
Keywords
Field
DocType
Deep neural networks, computer vision, parallel processing
Design space,Hyperparameter,Computer science,Inference,Parallel processing,Real-time computing,Network topology,Artificial intelligence,Artificial neural network,Energy consumption,Deep neural networks,Machine learning
Journal
Volume
Issue
ISSN
15
4
1544-3566
Citations 
PageRank 
References 
1
0.37
20
Authors
3
Name
Order
Citations
PageRank
Rahim Mammadli110.37
Felix Wolf25712.00
Ali Jannesari314920.11