Title
A Small-Footprint Accelerator for Large-Scale Neural Networks
Abstract
Machine-learning tasks are becoming pervasive in a broad range of domains, and in a broad range of systems (from embedded systems to data centers). At the same time, a small set of machine-learning algorithms (especially Convolutional and Deep Neural Networks, i.e., CNNs and DNNs) are proving to be state-of-the-art across many applications. As architectures evolve toward heterogeneous multicores composed of a mix of cores and accelerators, a machine-learning accelerator can achieve the rare combination of efficiency (due to the small number of target algorithms) and broad application scope. Until now, most machine-learning accelerator designs have been focusing on efficiently implementing the computational part of the algorithms. However, recent state-of-the-art CNNs and DNNs are characterized by their large size. In this study, we design an accelerator for large-scale CNNs and DNNs, with a special emphasis on the impact of memory on accelerator design, performance, and energy. We show that it is possible to design an accelerator with a high throughput, capable of performing 452 GOP/s (key NN operations such as synaptic weight multiplications and neurons outputs additions) in a small footprint of 3.02mm<sup>2</sup> and 485mW; compared to a 128-bit 2GHz SIMD processor, the accelerator is 117.87 × faster, and it can reduce the total energy by 21.08 ×. The accelerator characteristics are obtained after layout at 65nm. Such a high throughput in a small footprint can open up the usage of state-of-the-art machine-learning algorithms in a broad set of systems and for a broad set of applications.
Year
DOI
Venue
2015
10.1145/2701417
ACM Transactions on Computer Systems
Keywords
Field
DocType
Architecture,Processor,Hardware
Computer science,Convolutional neural network,Real-time computing,Footprint,Hardware acceleration,Artificial intelligence,Throughput,Deep learning,Artificial neural network,Small set,Synaptic weight,Distributed computing
Journal
Volume
Issue
ISSN
33
2
0734-2071
Citations 
PageRank 
References 
3
0.44
33
Authors
12
Name
Order
Citations
PageRank
Chen Tianshi1120559.29
Shijin Zhang240616.44
Shaoli Liu356027.88
Zidong Du457429.68
Tao Luo560123.27
Yuan Gao6402.12
Junjie Liu799.04
Dongsheng Wang837364.93
Chengyong Wu951526.67
SUN Ning-Hui10126897.37
Yunji Chen11143279.99
Olivier Temam122474148.79