论文标题

神经修剪通过生长正规化

Neural Pruning via Growing Regularization

论文作者

Wang, Huan, Qin, Can, Zhang, Yulun, Fu, Yun

论文摘要

长期以来,正规化已被用来学习深度神经网络修剪的稀疏性。但是,它的作用主要是在小额的惩罚力量制度中探索的。在这项工作中,我们将其应用程序扩展到了一个新的方案,该方案正规化逐渐增长,以解决修剪的两个核心问题:修剪时间表和体重的重要性评分。 (1)以前的主题在这项工作中是新提出的,我们发现这对修剪性能至关重要,而很少受到研究的关注。具体而言,我们提出了一个具有上升惩罚因素的L2正则化变体,并表明它与单发相比相比,即使去除相同的权重,也可以带来显着的准确性提高。 (2)日益增长的惩罚方案还为我们带来了一种利用Hessian信息以进行更准确修剪的方法,而不知道其特定值,因此不受共同的Hessian近似问题的困扰。从经验上讲,所提出的算法易于实现,并且可扩展到结构化和非结构化修剪的大型数据集和网络。与许多最先进的算法相比,CIFAR和Imagenet数据集上的现代深神经网络证明了它们的有效性。我们的代码和训练有素的模型可在https://github.com/mingsuntse/regularization-pruning上公开获取。

Regularization has long been utilized to learn sparsity in deep neural network pruning. However, its role is mainly explored in the small penalty strength regime. In this work, we extend its application to a new scenario where the regularization grows large gradually to tackle two central problems of pruning: pruning schedule and weight importance scoring. (1) The former topic is newly brought up in this work, which we find critical to the pruning performance while receives little research attention. Specifically, we propose an L2 regularization variant with rising penalty factors and show it can bring significant accuracy gains compared with its one-shot counterpart, even when the same weights are removed. (2) The growing penalty scheme also brings us an approach to exploit the Hessian information for more accurate pruning without knowing their specific values, thus not bothered by the common Hessian approximation problems. Empirically, the proposed algorithms are easy to implement and scalable to large datasets and networks in both structured and unstructured pruning. Their effectiveness is demonstrated with modern deep neural networks on the CIFAR and ImageNet datasets, achieving competitive results compared to many state-of-the-art algorithms. Our code and trained models are publicly available at https://github.com/mingsuntse/regularization-pruning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源