论文标题

完全在线元学习没有任务界限

Fully Online Meta-Learning Without Task Boundaries

论文作者

Rajasegaran, Jathushan, Finn, Chelsea, Levine, Sergey

论文摘要

虽然深层网络可以学习复杂的功能,例如分类器,检测器和跟踪器,但许多应用程序都需要不断适应更改输入分布,更改任务和改变环境条件的模型。的确,这种能够连续获得知识并利用过去的经验在持续设置中快速学习新任务的能力是智能系统的关键属性之一。对于复杂且高维的问题,只需通过标准学习算法(例如梯度下降)连续更新模型,可能会导致缓慢的适应性。元学习可以提供一个强大的工具来加速适应,但通常在批处理设置中研究。在本文中,我们研究了如何应用元学习来解决这种性质的在线问题,同时适应了不断变化的任务和输入分布以及对模型的元培训,以便将来更快地适应。将元学习扩展到在线环境中提出了自己的挑战,尽管有几种先前的方法研究了相关问题,但它们通常需要具有已知的基础真相任务范围的任务概念。这样的方法通常会按顺序适应每个任务,重置任务之间的模型,而不是在任务之间连续调整。在许多现实世界中,这种离散的边界不可用,甚至可能不存在。为了解决这些设置,我们提出了一种完全在线的元学习(FOML)算法,该算法不需要关于任务边界的任何基本真实知识,并且在不重新设置预先训练的权重的情况下完全在线。我们的实验表明,FOML能够比Rainbow-Mnist,Cifar100和Celeba数据集的最先进的在线学习方法更快地学习新任务。

While deep networks can learn complex functions such as classifiers, detectors, and trackers, many applications require models that continually adapt to changing input distributions, changing tasks, and changing environmental conditions. Indeed, this ability to continuously accrue knowledge and use past experience to learn new tasks quickly in continual settings is one of the key properties of an intelligent system. For complex and high-dimensional problems, simply updating the model continually with standard learning algorithms such as gradient descent may result in slow adaptation. Meta-learning can provide a powerful tool to accelerate adaptation yet is conventionally studied in batch settings. In this paper, we study how meta-learning can be applied to tackle online problems of this nature, simultaneously adapting to changing tasks and input distributions and meta-training the model in order to adapt more quickly in the future. Extending meta-learning into the online setting presents its own challenges, and although several prior methods have studied related problems, they generally require a discrete notion of tasks, with known ground-truth task boundaries. Such methods typically adapt to each task in sequence, resetting the model between tasks, rather than adapting continuously across tasks. In many real-world settings, such discrete boundaries are unavailable, and may not even exist. To address these settings, we propose a Fully Online Meta-Learning (FOML) algorithm, which does not require any ground truth knowledge about the task boundaries and stays fully online without resetting back to pre-trained weights. Our experiments show that FOML was able to learn new tasks faster than the state-of-the-art online learning methods on Rainbow-MNIST, CIFAR100 and CELEBA datasets.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源