论文标题
通过在线差异远程学习的无任务持续学习
Task-Free Continual Learning via Online Discrepancy Distance Learning
论文作者
论文摘要
由于缺乏明确的任务信息,从非平稳数据流学习(也称为无任务持续学习(TFCL))仍然具有挑战性。尽管最近提出了一些用于TFCL的方法,但它们缺乏理论保证。此外,理论上从未对TFCL期间的分析进行忘记分析。本文开发了一个新的理论分析框架,该框架根据访问的样本与可用于训练模型的全部信息提供了基于概括的范围。该分析为分类任务中的遗忘行为提供了新的见解。受这个理论模型的启发,我们提出了一种由混合模型的动态组件扩展机制启用的新方法,即在线差异距离学习(ODDL)。 ODDL估计当前内存缓冲区的概率表示与已经累积的知识之间的差异,并将其用作扩展信号,以确保具有最佳性能的紧凑网络体系结构。然后,我们提出了一种新的样本选择方法,该方法通过基于差异的度量有选择地将最相关的样本存储到存储器缓冲区中,从而进一步改善了性能。我们对提出的方法进行了几个TFCL实验,这些实验表明所提出的方法实现了最先进的性能状态。
Learning from non-stationary data streams, also called Task-Free Continual Learning (TFCL) remains challenging due to the absence of explicit task information. Although recently some methods have been proposed for TFCL, they lack theoretical guarantees. Moreover, forgetting analysis during TFCL was not studied theoretically before. This paper develops a new theoretical analysis framework which provides generalization bounds based on the discrepancy distance between the visited samples and the entire information made available for training the model. This analysis gives new insights into the forgetting behaviour in classification tasks. Inspired by this theoretical model, we propose a new approach enabled by the dynamic component expansion mechanism for a mixture model, namely the Online Discrepancy Distance Learning (ODDL). ODDL estimates the discrepancy between the probabilistic representation of the current memory buffer and the already accumulated knowledge and uses it as the expansion signal to ensure a compact network architecture with optimal performance. We then propose a new sample selection approach that selectively stores the most relevant samples into the memory buffer through the discrepancy-based measure, further improving the performance. We perform several TFCL experiments with the proposed methodology, which demonstrate that the proposed approach achieves the state of the art performance.