TY - GEN
T1 - Optimal SDRAM Buffer Allocator for Efficient Reuse of Layer IO in CNNs Inference Framework
AU - Miniskar, Narasinga Rao
AU - Pasupuleti, Sirish Kumar
AU - Rajagopal, Vasanthakumar
AU - Vishnoi, Ashok
AU - Ramasamy, Chandra Kumar
AU - Gadde, Raj Narayana
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2018/4/26
Y1 - 2018/4/26
N2 - Deep Learning based applications are becoming increasingly ubiquitous. The new generation smart phones are adapting lot of applications built on deep learning technology. However, adapting complex Deep Neural Network (DNN) applications on embedded processors is a huge challenge not only due to huge computational requirement, but also due to the massive SDRAM memory requirements of network layer IO buffers. Hence, an efficient reuse of layer IO buffers is required. However, it is challenging to reuse layer IO buffers because of complex network topology and large number of layers in the network. In this paper, we present an optimal SDRAM buffer allocator to minimize the overall SDRAM memory requirement of layer IO buffers, which works for any complex networks. The proposed SDRAM buffer allocator is integrated with ICNN (Inference only Convolutional Neural Networks) framework which is an extension of Caffe framework. Our framework with optimal SDRAM buffer allocator is experimented with popular AlexNet, GoogLeNet, ResNet-50 and Inception-ResNet-v2 CNNs. The results show 2× to 30× reduction in SDRAM footprint when compared with Caffe/TensorFlow frameworks and 26% to 47% reduction when compared with MXNet framework.
AB - Deep Learning based applications are becoming increasingly ubiquitous. The new generation smart phones are adapting lot of applications built on deep learning technology. However, adapting complex Deep Neural Network (DNN) applications on embedded processors is a huge challenge not only due to huge computational requirement, but also due to the massive SDRAM memory requirements of network layer IO buffers. Hence, an efficient reuse of layer IO buffers is required. However, it is challenging to reuse layer IO buffers because of complex network topology and large number of layers in the network. In this paper, we present an optimal SDRAM buffer allocator to minimize the overall SDRAM memory requirement of layer IO buffers, which works for any complex networks. The proposed SDRAM buffer allocator is integrated with ICNN (Inference only Convolutional Neural Networks) framework which is an extension of Caffe framework. Our framework with optimal SDRAM buffer allocator is experimented with popular AlexNet, GoogLeNet, ResNet-50 and Inception-ResNet-v2 CNNs. The results show 2× to 30× reduction in SDRAM footprint when compared with Caffe/TensorFlow frameworks and 26% to 47% reduction when compared with MXNet framework.
UR - http://www.scopus.com/inward/record.url?scp=85057093600&partnerID=8YFLogxK
U2 - 10.1109/ISCAS.2018.8351294
DO - 10.1109/ISCAS.2018.8351294
M3 - Conference contribution
AN - SCOPUS:85057093600
T3 - Proceedings - IEEE International Symposium on Circuits and Systems
BT - 2018 IEEE International Symposium on Circuits and Systems, ISCAS 2018 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2018 IEEE International Symposium on Circuits and Systems, ISCAS 2018
Y2 - 27 May 2018 through 30 May 2018
ER -