Kim, Eunhui
(Korea Institute of Science and Technology Information, Daejeon, South Korea)
,
Lee, Kyong-Ha
(Korea Institute of Science and Technology Information, Daejeon, South Korea)
Deep neural networks (DNN) have been applied to numerous artificial-intelligence applications because of their remarkable accuracy. However, computational requirements for deep neural networks are recently skyrocketing far beyond the Moore’s Law. In addition to the importance of accuracy, the...
Deep neural networks (DNN) have been applied to numerous artificial-intelligence applications because of their remarkable accuracy. However, computational requirements for deep neural networks are recently skyrocketing far beyond the Moore’s Law. In addition to the importance of accuracy, the industry’s demand for efficiency in model learning process is increasing. This has led to various attempts to make DNNs more lightweight. Hence, we propose a modeling technique that applies lightweight convolutional neural networks (CNN) to handle the model-learning processes for DNNs. The proposed spatial-shift pointwise quantization (SSPQ) model elegantly combines compact network-design techniques to revitalize DNN quantization efficiency with little accuracy loss. We set the depths of our SSPQ model to 20, 34, and 50 to test against CIFAR10, CIFAR100, and ImageNet datasets, respectively. By applying SSPQ20 to the CIFAR10 dataset, we reduced accuracy degradation by 2.95%, while reducing the number of parameters $8.7\times $. For the same dataset, our “wide” SSPQ20 variant reduced training parameters $1.96\times $, compared with the ResNet20 model architecture, which provided a 0.2% improvement. By applying SSPQ34 to the CIFAR100 dataset, we successfully reduced the number of learning parameters $8\times $, compared with the ResNet34 model with an accuracy degradation of 4.57%. By applying SSPQ50 to the ImageNet dataset, we successfully reduced the number of parameters $10.2\times $ over ResNet50 with an accuracy degradation of only 2.68%. Therefore, using only 9.78MB of learning parameter, the SSPQ50 model guarantees 73.6% accuracy. This is an improved performance by 1.46% in terms of accuracy and 3.62MB in terms of model size compared to the MobileNetV2 lightweight model. The source code is available at https://github.com/Eunhui-Kim/SSPQ.
Deep neural networks (DNN) have been applied to numerous artificial-intelligence applications because of their remarkable accuracy. However, computational requirements for deep neural networks are recently skyrocketing far beyond the Moore’s Law. In addition to the importance of accuracy, the industry’s demand for efficiency in model learning process is increasing. This has led to various attempts to make DNNs more lightweight. Hence, we propose a modeling technique that applies lightweight convolutional neural networks (CNN) to handle the model-learning processes for DNNs. The proposed spatial-shift pointwise quantization (SSPQ) model elegantly combines compact network-design techniques to revitalize DNN quantization efficiency with little accuracy loss. We set the depths of our SSPQ model to 20, 34, and 50 to test against CIFAR10, CIFAR100, and ImageNet datasets, respectively. By applying SSPQ20 to the CIFAR10 dataset, we reduced accuracy degradation by 2.95%, while reducing the number of parameters $8.7\times $. For the same dataset, our “wide” SSPQ20 variant reduced training parameters $1.96\times $, compared with the ResNet20 model architecture, which provided a 0.2% improvement. By applying SSPQ34 to the CIFAR100 dataset, we successfully reduced the number of learning parameters $8\times $, compared with the ResNet34 model with an accuracy degradation of 4.57%. By applying SSPQ50 to the ImageNet dataset, we successfully reduced the number of parameters $10.2\times $ over ResNet50 with an accuracy degradation of only 2.68%. Therefore, using only 9.78MB of learning parameter, the SSPQ50 model guarantees 73.6% accuracy. This is an improved performance by 1.46% in terms of accuracy and 3.62MB in terms of model size compared to the MobileNetV2 lightweight model. The source code is available at https://github.com/Eunhui-Kim/SSPQ.
참고문헌 (34)
Custom-Op feng 2019
Tensorpack wu 2016
Proc Adv Neural Inf Process Syst Normalization helps training of quantized LSTM hou 2019 7346
10.1007/978-3-030-58539-6_26
10.5244/C.30.87
10.1109/CVPR.2018.00951
Proc Adv Neural Inf Process Syst Constructing fast network through deconstruction of convolution jeon 2018 5951
10.1109/CVPR.2019.00741
10.1109/CVPR.2017.243
arXiv 1905 11946 EfficientNet: Rethinking model scaling for convolutional neural networks tan 2019
arXiv 1510 00149 [cs] Deep compression: Compressing deep neural networks with pruning, trained quantization and Huffman coding han 2015
Proc Adv Neural Inf Process Syst Binaryconnect: Training deep neural networks with binary weights during propagations courbariaux 2015 3123
※ AI-Helper는 부적절한 답변을 할 수 있습니다.