A systematic DNN weight pruning framework based on symmetric accelerated stochastic ADMM

Ming Yuan, Jianchao Bai, Feng Jiang, Lin Du

科研成果: 期刊稿件文章同行评审

2 引用 (Scopus)

摘要

Weight pruning is widely employed in compressing Deep Neural Networks (DNNs) because of the increasing computation and storage requirement. However, related work failed to efficiently combine the structure of the DNN loss function with the Alternating Direction Method of Multipliers (ADMM). This paper presents a systematic weight pruning framework of DNNs using the advanced symmetric accelerated stochastic ADMM (SAS-ADMM). Specifically, the weight pruning problem of DNNs is formulated as an optimization problem that consists of the DNN loss function and a L1 regularization term. SAS-ADMM is widely used to solve the problem by dividing it into two small-dimensional and relatively easier subproblems. Besides, an optimizer based on SAS-ADMM is presented to make the DNNs after pruning converge. Experimental results demonstrate that our method achieves a faster convergence rate in a better or similar weight pruning rate than previous work. For the CIFAR-10 data set, our method reduces the number of ResNet-32 and ResNet-56 parameters by a factor of 6.61× and 9.93 × while maintaining accuracy. In similar experiments of AlexNet on the ImageNet data set, we achieve 20.9× weight reduction, which only takes half of the time compared with prior work.

源语言英语
文章编号127327
期刊Neurocomputing
575
DOI
出版状态已出版 - 28 3月 2024

指纹

探究 'A systematic DNN weight pruning framework based on symmetric accelerated stochastic ADMM' 的科研主题。它们共同构成独一无二的指纹。

引用此