Regularized Training Framework for Combining Pruning and Quantization to Compress Neural Networks

Qimin Ding, Ruonan Zhang, Yi Jiang, Daosen Zhai, Bin Li

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

Many convolutional neural networks(CNNs) have been proposed to solve computer vision tasks such as image classification and image segmentation. However the CNNs usually contain a large number of parameters to determine which consumes very high computation and power resources. Thus, it is difficult to deploy the CNNs on resource-limited devices. Network pruning and network quantization are two main methods to compress the CNNs, researchers often apply these methods individually without considering the relationship between them. In this paper, we explore the coupling relationship between network pruning and quantization, as well as the limits of the current network compression training method. Then we propose a new regularized training method that can combine pruning and quantization within a simple training framework. Experiments show that by using the proposed training framework, the finetune process is not needed anymore and hence we can reduce much time for training a network. The simulation results also show that the performance of the network can over-perform the traditional methods. The proposed framework is suitable for the CNNs deployed in portable devices with limited computational resources and power supply.

源语言英语
主期刊名2019 11th International Conference on Wireless Communications and Signal Processing, WCSP 2019
出版商Institute of Electrical and Electronics Engineers Inc.
ISBN(电子版)9781728135557
DOI
出版状态已出版 - 10月 2019
活动11th International Conference on Wireless Communications and Signal Processing, WCSP 2019 - Xi'an, 中国
期限: 23 10月 201925 10月 2019

出版系列

姓名2019 11th International Conference on Wireless Communications and Signal Processing, WCSP 2019

会议

会议11th International Conference on Wireless Communications and Signal Processing, WCSP 2019
国家/地区中国
Xi'an
时期23/10/1925/10/19

指纹

探究 'Regularized Training Framework for Combining Pruning and Quantization to Compress Neural Networks' 的科研主题。它们共同构成独一无二的指纹。

引用此