论文标题
基于端到端灵敏度的过滤器修剪
End-to-End Sensitivity-Based Filter Pruning
论文作者
论文摘要
在本文中,我们提出了一种基于灵敏度的新型滤波器修剪算法(SBF-Pruner),以了解端到端每一层过滤器的重要性得分。我们的方法从滤波器的权重学习得分,使其能够说明每一层过滤器之间的相关性。此外,通过同时训练所有层的修剪得分,我们的方法可以解释层相互依赖性,这对于找到性能的稀疏子网络至关重要。我们提出的方法可以在不需要预审慎的网络的情况下直接,单阶段的训练过程中从头开始训练和生成修剪的网络。最终,我们不需要特定图层特异性的超参数和预定义的层预算,因为SBF-Pruner可以隐式地确定每个层中适当的通道数。我们对不同网络体系结构的实验结果表明,SBF-Pruner的表现优于先进的修剪方法。值得注意的是,在CIFAR-10上,在不需要预处理的基线网络的情况下,我们获得1.02%和1.19%的RESNET56和RESNET110的准确性增益,与报告的最先进的修剪算法的基线相比。这是当SBF-Pruner将参数计数减少52.3%(对于RESNET56)和54%(对于RESNET101),这比最先进的修剪算法要好9.5%和6.6%。
In this paper, we present a novel sensitivity-based filter pruning algorithm (SbF-Pruner) to learn the importance scores of filters of each layer end-to-end. Our method learns the scores from the filter weights, enabling it to account for the correlations between the filters of each layer. Moreover, by training the pruning scores of all layers simultaneously our method can account for layer interdependencies, which is essential to find a performant sparse sub-network. Our proposed method can train and generate a pruned network from scratch in a straightforward, one-stage training process without requiring a pretrained network. Ultimately, we do not need layer-specific hyperparameters and pre-defined layer budgets, since SbF-Pruner can implicitly determine the appropriate number of channels in each layer. Our experimental results on different network architectures suggest that SbF-Pruner outperforms advanced pruning methods. Notably, on CIFAR-10, without requiring a pretrained baseline network, we obtain 1.02% and 1.19% accuracy gain on ResNet56 and ResNet110, compared to the baseline reported for state-of-the-art pruning algorithms. This is while SbF-Pruner reduces parameter-count by 52.3% (for ResNet56) and 54% (for ResNet101), which is better than the state-of-the-art pruning algorithms with a high margin of 9.5% and 6.6%.