Abstract | ||
---|---|---|
Neural Networks or NNs are widely used for many machine learning applications such as image processing and speech recognition. Since general-purpose processors such as CPUs and GPUs are energy inefficient for computing NNs, application-specific hardware accelerators for NNs (a.k.a. Neural Network Accelerators or NNAs) have been proposed to improve the energy efficiency. However, the existing NNAs are too customized for computing specific NNs, and do not allow to change neuron models or learning algorithms. This limitation prevents machine-learning researchers from exploiting NNAs, so we are developing a general-purpose NNA including reconfigurable logic, which is called a reconfigurable NNA or RNNA. The RNNA is highly tuned for the NN computation but allows end users to customize the hardware to compute desired NNs. This paper introduces the RNNA architecture, and reports the performance analysis of the RNNA with an in-house cycle-level simulator. |
Year | DOI | Venue |
---|---|---|
2016 | 10.1109/CANDAR.2016.106 | 2016 FOURTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING (CANDAR) |
Field | DocType | ISSN |
Computer architecture,End user,Efficient energy use,Computer science,Parallel computing,Image processing,Artificial neural network,Computation | Conference | 2379-1888 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Momoka Ohba | 1 | 0 | 0.34 |
Satoshi Shindo | 2 | 0 | 0.34 |
Shinobu Miwa | 3 | 28 | 13.09 |
Tomoaki Tsumura | 4 | 58 | 14.63 |
Hayato Yamaki | 5 | 0 | 3.04 |
Hiroki Honda | 6 | 2 | 1.09 |