论文标题
自动召回机:内部重播,持续学习和大脑
Automatic Recall Machines: Internal Replay, Continual Learning and the Brain
论文作者
论文摘要
在神经网络中的重播涉及对记忆样本进行连续数据的培训,这抵消了忘记由非平稳性引起的先前行为。我们提出了一种方法,其中仅鉴于正在为评估目标训练的模型,而没有外部缓冲液或发电机网络,因此即时生成了这些辅助样品。取而代之的是,利用了评估模型本身中学到的样本的隐式记忆。此外,尽管现有工作重点是增强完整的数据分布,但我们表明,优化而不要忘记呼吁生成专门针对每个真实培训批次的样本的呼吁,这是更有效和可扩展的。我们考虑与大脑的高水平相似之处,特别是将单个模型用于推理和召回,召回样本对当前环境批次的依赖性,激活和学习的自上而下调制,抽象回忆以及学习任务的程度与召回的程度之间的依赖性。这些特征自然而然地从该方法中出现,而无需控制。
Replay in neural networks involves training on sequential data with memorized samples, which counteracts forgetting of previous behavior caused by non-stationarity. We present a method where these auxiliary samples are generated on the fly, given only the model that is being trained for the assessed objective, without extraneous buffers or generator networks. Instead the implicit memory of learned samples within the assessed model itself is exploited. Furthermore, whereas existing work focuses on reinforcing the full seen data distribution, we show that optimizing for not forgetting calls for the generation of samples that are specialized to each real training batch, which is more efficient and scalable. We consider high-level parallels with the brain, notably the use of a single model for inference and recall, the dependency of recalled samples on the current environment batch, top-down modulation of activations and learning, abstract recall, and the dependency between the degree to which a task is learned and the degree to which it is recalled. These characteristics emerge naturally from the method without being controlled for.