论文标题
(持续)的类标准化?广义零射学习
Class Normalization for (Continual)? Generalized Zero-Shot Learning
论文作者
论文摘要
事实证明,标准化技术是在传统监督学习制度中成功培训的关键要素。但是,在零拍的学习(ZSL)世界中,这些想法只受到了边缘的关注。这项工作从理论和实际角度研究了ZSL方案中的归一化。首先,我们对在零拍学习中使用的两个流行技巧进行了理论解释:归一化+刻度和属性归一化,并表明它们通过在向前传球期间保留差异来帮助训练。接下来,我们证明它们不足以使深度ZSL模型正常化并提出类别归一化(CN):一种归一化方案,该方案可以减轻该问题,从而证明并且在实践中。第三,我们表明,与传统分类器相比,ZSL模型通常具有更不规则的损失表面,并且所提出的方法部分补救了此问题。然后,我们在4个标准ZSL数据集上测试了我们的方法,并以简单的MLP优化了无铃声和哨声,并且训练速度更快约50倍。最后,我们将ZSL推广到一个更广泛的问题 - 连续的ZSL,并为此新设置介绍了一些原则上的指标和严格的基线。项目页面位于https://universome.github.io/class-norm。
Normalization techniques have proved to be a crucial ingredient of successful training in a traditional supervised learning regime. However, in the zero-shot learning (ZSL) world, these ideas have received only marginal attention. This work studies normalization in ZSL scenario from both theoretical and practical perspectives. First, we give a theoretical explanation to two popular tricks used in zero-shot learning: normalize+scale and attributes normalization and show that they help training by preserving variance during a forward pass. Next, we demonstrate that they are insufficient to normalize a deep ZSL model and propose Class Normalization (CN): a normalization scheme, which alleviates this issue both provably and in practice. Third, we show that ZSL models typically have more irregular loss surface compared to traditional classifiers and that the proposed method partially remedies this problem. Then, we test our approach on 4 standard ZSL datasets and outperform sophisticated modern SotA with a simple MLP optimized without any bells and whistles and having ~50 times faster training speed. Finally, we generalize ZSL to a broader problem -- continual ZSL, and introduce some principled metrics and rigorous baselines for this new setup. The project page is located at https://universome.github.io/class-norm.