论文标题
质心匹配:一种有效的持续学习方法在嵌入空间中运行
Centroids Matching: an efficient Continual Learning approach operating in the embedding space
论文作者
论文摘要
当神经网络失去以前从不同分布的样本(即新任务)培训一组样本时,发生灾难性遗忘(CF)。现有方法在减轻CF方面取得了显着的结果,尤其是在一种称为任务增量学习的情况下。但是,这种情况是不现实的,并且已经在更现实的情况下完成了有限的工作来取得良好的结果。在本文中,我们提出了一种称为Centroid匹配的新型正则化方法,该方法受到元学习方法的启发,通过在神经网络产生的特征空间中运行,在需要良好的结果中与CF作斗争,同时需要少量的记忆足迹。具体而言,该方法使用神经网络产生的特征向量直接对样品进行了分类,通过将这些向量与代表当前任务中的类或所有任务的质心匹配,直到该点。质心匹配速度比竞争基线更快,并且可以通过在过去的任务结束时保留模型产生的嵌入空间和当前生产的嵌入空间之间的距离,从而有效地减轻CF,从而实现了一种在不使用外部记忆的情况下实现易于实现的方法,或者使用易于实现的情况,则可以实现高准确的方法。广泛的实验表明,匹配的质心在多个数据集和方案上取得了准确的提高。
Catastrophic forgetting (CF) occurs when a neural network loses the information previously learned while training on a set of samples from a different distribution, i.e., a new task. Existing approaches have achieved remarkable results in mitigating CF, especially in a scenario called task incremental learning. However, this scenario is not realistic, and limited work has been done to achieve good results on more realistic scenarios. In this paper, we propose a novel regularization method called Centroids Matching, that, inspired by meta-learning approaches, fights CF by operating in the feature space produced by the neural network, achieving good results while requiring a small memory footprint. Specifically, the approach classifies the samples directly using the feature vectors produced by the neural network, by matching those vectors with the centroids representing the classes from the current task, or all the tasks up to that point. Centroids Matching is faster than competing baselines, and it can be exploited to efficiently mitigate CF, by preserving the distances between the embedding space produced by the model when past tasks were over, and the one currently produced, leading to a method that achieves high accuracy on all the tasks, without using an external memory when operating on easy scenarios, or using a small one for more realistic ones. Extensive experiments demonstrate that Centroids Matching achieves accuracy gains on multiple datasets and scenarios.