Title
Deep Learning on FPGAs: Past, Present, and Future.
Abstract
The rapid growth of data size and accessibility in recent years has instigated a shift of philosophy in algorithm design for artificial intelligence. Instead of engineering algorithms by hand, the ability to learn composable systems automatically from massive amounts of data has led to ground-breaking performance in important domains such as computer vision, speech recognition, and natural language processing. The most popular class of techniques used in these domains is called deep learning, and is seeing significant attention from industry. However, these models require incredible amounts of data and compute power to train, and are limited by the need for better hardware acceleration to accommodate scaling beyond current data and model sizes. While the current solution has been to use clusters of graphics processing units (GPU) as general purpose processors (GPGPU), the use of field programmable gate arrays (FPGA) provide an interesting alternative. Current trends in design tools for FPGAs have made them more compatible with the high-level software practices typically practiced in the deep learning community, making FPGAs more accessible to those who build and deploy models. Since FPGA architectures are flexible, this could also allow researchers the ability to explore model-level optimizations beyond what is possible on fixed architectures such as GPUs. As well, FPGAs tend to provide high performance per watt of power consumption, which is of particular importance for application scientists interested in large scale server-based deployment or resource-limited embedded applications. This review takes a look at deep learning and FPGAs from a hardware acceleration perspective, identifying trends and innovations that make these technologies a natural fit, and motivates a discussion on how FPGAs may best serve the needs of the deep learning community moving forward.
Year
Venue
Field
2016
arXiv: Distributed, Parallel, and Cluster Computing
Graphics,Computer science,Field-programmable gate array,Real-time computing,Software,Hardware acceleration,General-purpose computing on graphics processing units,Artificial intelligence,Performance per watt,Deep learning,Distributed computing,Reconfigurable computing
DocType
Volume
Citations 
Journal
abs/1602.04283
15
PageRank 
References 
Authors
0.64
17
3
Name
Order
Citations
PageRank
Griffin Lacey1150.64
Graham W. Taylor21523127.22
Shawki Areibi336741.38