How Can Deep Neural Networks Be Generated Efficiently for Devices with Limited Resources

作者: Unai Elordi , Luis Unzueta , Ignacio Arganda-Carreras , Oihana Otaegui

DOI: 10.1007/978-3-319-94544-6_3

关键词: ComputationDistributed computingDeep learningDocumentationPipeline (software)Software deploymentSIMPLE (military communications protocol)Artificial neural networkDeep neural networksArtificial intelligenceComputer science

摘要: Despite the increasing hardware capabilities of embedded devices, running a Deep Neural Network (DNN) in such systems remains challenge. As trend DNNs is to design more complex architectures, computation time low-resource devices increases dramatically due their low memory capabilities. Moreover, physical used store network parameters augments with its complexity, hindering feasible model be deployed target hardware. Although compressed helps reducing RAM consumption, large amount consecutive deep layers time. wide literature about DNN optimization, there lack documentation for practical and efficient deployment these networks. In this paper, we propose an generation by analyzing impact address simple comprehensive pipeline optimal deployment.

参考文章(29)
Karen Simonyan, Andrew Zisserman, Very Deep Convolutional Networks for Large-Scale Image Recognition computer vision and pattern recognition. ,(2014)
Ankur Agrawal, Pritish Narayanan, Kailash Gopalakrishnan, Suyog Gupta, Deep Learning with Limited Numerical Precision arXiv: Learning. ,(2015)
Philipp Fischer, Thomas Brox, None, U-Net: Convolutional Networks for Biomedical Image Segmentation medical image computing and computer assisted intervention. pp. 234- 241 ,(2015) , 10.1007/978-3-319-24574-4_28
Yoshua Bengio, Matthieu Courbariaux, Jean-Pierre David, BinaryConnect: Training Deep Neural Networks with binary weights during propagations arXiv: Learning. ,(2015)
Li Deng, Jinyu Li, Jui-Ting Huang, Kaisheng Yao, Dong Yu, Frank Seide, Michael Seltzer, Geoff Zweig, Xiaodong He, Jason Williams, Yifan Gong, Alex Acero, Recent advances in deep learning for speech research at Microsoft international conference on acoustics, speech, and signal processing. pp. 8604- 8608 ,(2013) , 10.1109/ICASSP.2013.6639345
Yann LeCun, John Denker, Sara Solla, None, Optimal Brain Damage neural information processing systems. ,vol. 2, pp. 598- 605 ,(1989)
Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, Li Fei-Fei, ImageNet Large Scale Visual Recognition Challenge International Journal of Computer Vision. ,vol. 115, pp. 211- 252 ,(2015) , 10.1007/S11263-015-0816-Y
Babak Hassibi, David G. Stork, Second order derivatives for network pruning: Optimal Brain Surgeon neural information processing systems. ,vol. 5, pp. 164- 171 ,(1992)
Ilya Sutskever, Geoffrey E. Hinton, Alex Krizhevsky, ImageNet Classification with Deep Convolutional Neural Networks neural information processing systems. ,vol. 25, pp. 1097- 1105 ,(2012)
Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun, Deep Residual Learning for Image Recognition computer vision and pattern recognition. pp. 770- 778 ,(2016) , 10.1109/CVPR.2016.90