Title
TrainWare: A Memory Optimized Weight Update Architecture for On-Device Convolutional Neural Network Training
Abstract
Training convolutional neural network on device has become essential where it allows applications to consider user's individual environment. Meanwhile, the weight update operation from the training process is the primary factor of high energy consumption due to its substantial memory accesses. We propose a dedicated weight update architecture with two key features: (1) a specialized local buffer for the DRAM access deduction (2) a novel dataflow and its suitable processing element array structure for weight gradient computation to optimize the energy consumed by internal memories. Our scheme achieves 14.3%-30.2% total energy reduction by drastically eliminating the memory accesses.
Year
DOI
Venue
2018
10.1145/3218603.3218625
ISLPED
Field
DocType
ISBN
Dram,Array data structure,Computer architecture,Architecture,Convolutional neural network,Computer science,Real-time computing,Dataflow,Processing element,High energy,Computation
Conference
978-1-4503-5704-3
Citations 
PageRank 
References 
3
0.39
9
Authors
4
Name
Order
Citations
PageRank
Seungkyu Choi1103.90
Jaehyeong Sim2527.63
Myeonggu Kang3124.00
Lee-Sup Kim470798.58