Title | ||
---|---|---|
TrainWare: A Memory Optimized Weight Update Architecture for On-Device Convolutional Neural Network Training |
Abstract | ||
---|---|---|
Training convolutional neural network on device has become essential where it allows applications to consider user's individual environment. Meanwhile, the weight update operation from the training process is the primary factor of high energy consumption due to its substantial memory accesses. We propose a dedicated weight update architecture with two key features: (1) a specialized local buffer for the DRAM access deduction (2) a novel dataflow and its suitable processing element array structure for weight gradient computation to optimize the energy consumed by internal memories. Our scheme achieves 14.3%-30.2% total energy reduction by drastically eliminating the memory accesses.
|
Year | DOI | Venue |
---|---|---|
2018 | 10.1145/3218603.3218625 | ISLPED |
Field | DocType | ISBN |
Dram,Array data structure,Computer architecture,Architecture,Convolutional neural network,Computer science,Real-time computing,Dataflow,Processing element,High energy,Computation | Conference | 978-1-4503-5704-3 |
Citations | PageRank | References |
3 | 0.39 | 9 |
Authors | ||
4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Seungkyu Choi | 1 | 10 | 3.90 |
Jaehyeong Sim | 2 | 52 | 7.63 |
Myeonggu Kang | 3 | 12 | 4.00 |
Lee-Sup Kim | 4 | 707 | 98.58 |