Structured Pruning of Deep Convolutional Neural Networks

作者: Sajid Anwar , Kyuyeon Hwang , Wonyong Sung

DOI: 10.1145/3005348

关键词: Machine learningComputational complexity theoryComputational resourceImportance WeightArtificial intelligenceDeep learningParticle filterComputer scienceConvolutional neural networkKernel (image processing)

摘要: Real-time application of deep learning algorithms is often hindered by high computational complexity and frequent memory accesses. Network pruning a promising technique to solve this problem. However, usually results in irregular network connections that not only demand extra representation efforts but also do fit well on parallel computation. We introduce structured sparsity at various scales for convolutional neural networks: feature map-wise, kernel-wise, intra-kernel strided sparsity. This very advantageous direct resource savings embedded computers, computing environments, hardware-based systems. To decide the importance paths, proposed method uses particle filtering approach. The weight each assigned assessing misclassification rate with corresponding connectivity pattern. pruned retrained compensate losses due pruning. While implementing convolutions as matrix products, we particularly show simple constraint can significantly reduce size kernel map tensors. work shows when granularities are applied combination, prune CIFAR-10 more than 70% less 1% loss accuracy.

参考文章(32)
Matthew Zeiler, Rob Fergus, Li Wan, Yann Le Cun, Sixin Zhang, Regularization of Neural Networks using DropConnect international conference on machine learning. pp. 1058- 1066 ,(2013)
Kumar Chellapilla, Sidd Puri, Patrice Simard, High Performance Convolutional Neural Networks for Document Processing international conference on frontiers in handwriting recognition. ,(2006)
Maxwell D. Collins, Pushmeet Kohli, Memory Bounded Deep Convolutional Networks. arXiv: Computer Vision and Pattern Recognition. ,(2014)
Evan Shelhamer, John Tran, Jonathan Cohen, Sharan Chetlur, Philippe Vandermersch, Cliff Woolley, Bryan Catanzaro, cuDNN: Efficient Primitives for Deep Learning arXiv: Neural and Evolutionary Computing. ,(2014)
Karen Simonyan, Andrew Zisserman, Very Deep Convolutional Networks for Large-Scale Image Recognition computer vision and pattern recognition. ,(2014)
Yoshua Bengio, Matthieu Courbariaux, Jean-Pierre David, BinaryConnect: Training Deep Neural Networks with binary weights during propagations arXiv: Learning. ,(2015)
Yann LeCun, Mikael Henaff, Michael Mathieu, Fast Training of Convolutional Networks through FFTs arXiv: Computer Vision and Pattern Recognition. ,(2013)
Tiancheng Li, Shudong Sun, Tariq Pervez Sattar, Juan Manuel Corchado, Review: Fight sample degeneracy and impoverishment in particle filters: A review of intelligent approaches Expert Systems With Applications. ,vol. 41, pp. 3944- 3954 ,(2014) , 10.1016/J.ESWA.2013.12.031
N.M. Kwok, Gu Fang, W. Zhou, Evolutionary particle filter: re-sampling from the genetic algorithm perspective intelligent robots and systems. pp. 2935- 2940 ,(2005) , 10.1109/IROS.2005.1545119
Daniel Strigl, Klaus Kofler, Stefan Podlipnig, Performance and Scalability of GPU-Based Convolutional Neural Networks parallel, distributed and network-based processing. pp. 317- 324 ,(2010) , 10.1109/PDP.2010.43