TY - GEN
T1 - Regularized Training Framework for Combining Pruning and Quantization to Compress Neural Networks
AU - Ding, Qimin
AU - Zhang, Ruonan
AU - Jiang, Yi
AU - Zhai, Daosen
AU - Li, Bin
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/10
Y1 - 2019/10
N2 - Many convolutional neural networks(CNNs) have been proposed to solve computer vision tasks such as image classification and image segmentation. However the CNNs usually contain a large number of parameters to determine which consumes very high computation and power resources. Thus, it is difficult to deploy the CNNs on resource-limited devices. Network pruning and network quantization are two main methods to compress the CNNs, researchers often apply these methods individually without considering the relationship between them. In this paper, we explore the coupling relationship between network pruning and quantization, as well as the limits of the current network compression training method. Then we propose a new regularized training method that can combine pruning and quantization within a simple training framework. Experiments show that by using the proposed training framework, the finetune process is not needed anymore and hence we can reduce much time for training a network. The simulation results also show that the performance of the network can over-perform the traditional methods. The proposed framework is suitable for the CNNs deployed in portable devices with limited computational resources and power supply.
AB - Many convolutional neural networks(CNNs) have been proposed to solve computer vision tasks such as image classification and image segmentation. However the CNNs usually contain a large number of parameters to determine which consumes very high computation and power resources. Thus, it is difficult to deploy the CNNs on resource-limited devices. Network pruning and network quantization are two main methods to compress the CNNs, researchers often apply these methods individually without considering the relationship between them. In this paper, we explore the coupling relationship between network pruning and quantization, as well as the limits of the current network compression training method. Then we propose a new regularized training method that can combine pruning and quantization within a simple training framework. Experiments show that by using the proposed training framework, the finetune process is not needed anymore and hence we can reduce much time for training a network. The simulation results also show that the performance of the network can over-perform the traditional methods. The proposed framework is suitable for the CNNs deployed in portable devices with limited computational resources and power supply.
KW - convolutional neural networks
KW - coupling relationship
KW - fuzzy rules
KW - network compression
KW - training framework
UR - http://www.scopus.com/inward/record.url?scp=85077780921&partnerID=8YFLogxK
U2 - 10.1109/WCSP.2019.8928083
DO - 10.1109/WCSP.2019.8928083
M3 - 会议稿件
AN - SCOPUS:85077780921
T3 - 2019 11th International Conference on Wireless Communications and Signal Processing, WCSP 2019
BT - 2019 11th International Conference on Wireless Communications and Signal Processing, WCSP 2019
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 11th International Conference on Wireless Communications and Signal Processing, WCSP 2019
Y2 - 23 October 2019 through 25 October 2019
ER -