论文标题
通过自适应正则化发展神经选择
Evolving Neural Selection with Adaptive Regularization
论文作者
论文摘要
过度参数化是现代深神经网络的固有特征之一,通常可以通过利用正则化方法(例如辍学)来克服。通常,这些方法是全球应用的,所有输入案例都经过平等处理。但是,鉴于对现实世界任务(例如图像识别和自然语言理解)的输入空间的自然变化,固定的正则化模式不太可能对所有输入案例具有相同的有效性。在这项工作中,我们展示了一种方法,即深层神经网络中神经元的选择会发展,以适应预测的难度。我们提出了自适应神经选择(ANS)框架,该框架演变为在一层中称重神经元以形成适合处理不同输入案例的网络变体。实验结果表明,所提出的方法可以显着提高标准图像识别基准的常用神经网络体系结构的性能。消融研究还验证了拟议框架中每个组件的有效性和贡献。
Over-parameterization is one of the inherent characteristics of modern deep neural networks, which can often be overcome by leveraging regularization methods, such as Dropout. Usually, these methods are applied globally and all the input cases are treated equally. However, given the natural variation of the input space for real-world tasks such as image recognition and natural language understanding, it is unlikely that a fixed regularization pattern will have the same effectiveness for all the input cases. In this work, we demonstrate a method in which the selection of neurons in deep neural networks evolves, adapting to the difficulty of prediction. We propose the Adaptive Neural Selection (ANS) framework, which evolves to weigh neurons in a layer to form network variants that are suitable to handle different input cases. Experimental results show that the proposed method can significantly improve the performance of commonly-used neural network architectures on standard image recognition benchmarks. Ablation studies also validate the effectiveness and contribution of each component in the proposed framework.