论文标题
MGIAD:在各个方面的跨越。效率和鲁棒性通过在分辨率和渠道维度上的变形而稳健
MGiaD: Multigrid in all dimensions. Efficiency and robustness by coarsening in resolution and channel dimensions
论文作者
论文摘要
当前用于图像分类的最新深层神经网络由10-1亿可学习的权重组成,因此本质上容易过度拟合。重量计数的复杂性可以看作是通道数,输入的空间范围和网络层数的函数。由于使用卷积层,重量复杂性的缩放通常与分辨率尺寸有关,但相对于通道数量保持二次。近年来,就深层神经网络中使用多族灵感的想法而言,积极的研究表明,一方面,可以通过适当的重量共享来节省大量的权重,而另一方面,通道维度中的层次结构可以提高线性的重量复杂性。在这项工作中,我们结合了这些多族想法,以引入一个由多族启发的架构的联合框架,该架构在所有相关维度中利用了多移民结构,以实现线性重量复杂性缩放和大幅度降低的重量计数。我们的实验表明,这种结构化的体重计数能够减少过度拟合,因此在较低的网络复杂性下的典型图像分类基准上显示了最先进的重新系统架构的性能。
Current state-of-the-art deep neural networks for image classification are made up of 10 - 100 million learnable weights and are therefore inherently prone to overfitting. The complexity of the weight count can be seen as a function of the number of channels, the spatial extent of the input and the number of layers of the network. Due to the use of convolutional layers the scaling of weight complexity is usually linear with regards to the resolution dimensions, but remains quadratic with respect to the number of channels. Active research in recent years in terms of using multigrid inspired ideas in deep neural networks have shown that on one hand a significant number of weights can be saved by appropriate weight sharing and on the other that a hierarchical structure in the channel dimension can improve the weight complexity to linear. In this work, we combine these multigrid ideas to introduce a joint framework of multigrid inspired architectures, that exploit multigrid structures in all relevant dimensions to achieve linear weight complexity scaling and drastically reduced weight counts. Our experiments show that this structured reduction in weight count is able to reduce overfitting and thus shows improved performance over state-of-the-art ResNet architectures on typical image classification benchmarks at lower network complexity.