Abstract | ||
---|---|---|
Machine learning nowadays is profoundly impacting every aspect of our lives. With the evolution of the machine learning, many techniques, such as deep learning, improve the accuracy and performance of machine learning. Deep learning is a set of ML techniques that use layers of transformation and consist of neural networks. The power consumption of deep learning becomes a serious problem when it comes to edge computing. One of the most computationally demanding operation of DNN is convolution which preserve the image arrangement and obtain partial image as an input feature. Our goal is to find an effective way for programmers to improve the performance of convolution operation. In this paper, we proposed the design of sparse matrix compression schedule primitives in Halide and find a way to improve convolution operation with im2col method. Halide is an image processing programming language that separates algorithm from its schedule. With this design, we can compress the result of im2col matrix to achieve performance improvements. In our experiments, results show that convolution operation can achieve 20X speedup with our implementation.
|
Year | DOI | Keywords |
---|---|---|
2019 | 10.1145/3339186.3339194 | Deep Learning, Halide, OpenCL, Sparse Matrix |
Field | DocType | ISSN |
Compression (physics),Computer science,Parallel computing,Halide,Sparse matrix | Conference | 978-1-4503-7196-4 |
ISBN | Citations | PageRank |
978-1-4503-7196-4 | 0 | 0.34 |
References | Authors | |
0 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Chao-Lin Lee | 1 | 1 | 3.41 |
Chen-Ting Chao | 2 | 0 | 1.69 |
Jenq-Kuen Lee | 3 | 120 | 17.98 |
Ming-Yu Hung | 4 | 53 | 7.68 |
Chung-Wen Huang | 5 | 37 | 5.91 |