Loading [MathJax]/extensions/MathMenu.js
Fast Convolution Algorithm for Convolutional Neural Networks | IEEE Conference Publication | IEEE Xplore

Fast Convolution Algorithm for Convolutional Neural Networks


Abstract:

Recent advances in computing power made possible by developments of faster general-purpose graphics processing units (GPGPUs) have increased the complexity of convolution...Show More

Abstract:

Recent advances in computing power made possible by developments of faster general-purpose graphics processing units (GPGPUs) have increased the complexity of convolutional neural network (CNN) models. However, because of the limited applications of the existing GPGPUs, CNN accelerators are becoming more important. The current accelerators focus on improvement in memory scheduling and architectures. Thus, the number of multiplier-accumulator (MAC) operations is not reduced. In this study, a new convolution layer operation algorithm is proposed using the coarse-to-fine method instead of hardware or architecture approaches. This algorithm is shown to reduce the MAC operations by 33%. However, the accuracy of the Top 1 is decreased only by 3% and the Top 5 only by 1%.
Date of Conference: 18-20 March 2019
Date Added to IEEE Xplore: 25 July 2019
ISBN Information:
Conference Location: Hsinchu, Taiwan
References is not available for this document.

I. Introduction

Computer vision technology using deep learning has recently been utilized in such areas such medicine, entertainment, robots, and vehicles. These deep-learning technologies mimic the neural network within the brain. Among the deep neural networks, the convolutional neural network (CNN), in particular, has shown promising performance in such areas as object recognition and image classification [1] .

Select All
1.
"A Resource-Limited Hardware Accelerator for Convolutional Neural Networks in Embedded Vision Applications".
2.
"A Survey of Model Compression and Acceleration for Deep Neural Networks".
3.
K. Simonyan and A. Zisserman, "Very deep convolutional networks for large-scale image recognition", Int. Conf. Learning Representations, May 2015.
4.
C. Szegedy et al., "Going deeper with convolutions", Proc. IEEE Comput. Vis. Pattern Recognit., Jun. 2015.
5.
K. He et al., "Deep residual learning for image recognition", Proc. IEEE Comput. Vis. Pattern Recognit., Jun. 2016.
6.
A. Krizhevsky, I. Sutskever and G. E. Hinton, "ImageNet classification with deep convolutional neural networks", Proc. Adv. Neural Inf. Process. Syst., vol. 25, pp. 1097-1105, 2012.
7.
A. Ardakani, C. Condo and W. J. Gross, "An Architecture to Accelerate Convolution in Deep Neural Networks", IEEE Transactions on Circuits and Systems - I, vol. 65, no. 4, pp. 1349-1362, 2018.
8.
Y. Ma, Y. Cao, S. Vrudhula and J. Seo, "Optimizing the Convolution Operation to Accelerate Deep Neural Networks on FPGA", IEEE Transactions on Very Large Scale Integration Systems, vol. 26, no. 7, pp. 1354-1367, 2018.
9.
M. Jaderberg, A. Vedaldi and A. Zisserman, "Speeding up convolutional neural networks with low rank expansions", Proceedings of the British Machine Vision Conference, 2014.
10.
V. Lebedev, Y. Ganin, M. Rakhuba, I. V. Oseledets and V. S. Lempitsky, "Speeding-up convolutional neural networks using fine-tuned cp decomposition", CoRR, vol. abs/1412.6553, 2014.
11.
C. Tai, T. Xiao, X. Wang and W. E, "Convolutional neural networks with low-rank regularization", vol. abs/1511.06067, 2015.

Contact IEEE to Subscribe

References

References is not available for this document.