论文标题
元学习和应用的信息理论概括范围
Information-Theoretic Generalization Bounds for Meta-Learning and Applications
论文作者
论文摘要
元学习或“学习学习”是指从与多个相关任务相对应的数据中推断出归纳偏置的技术,目的是提高新的,以前未观察到的任务的样本效率。元学习的关键绩效指标是元将军差距,即在元训练数据上测量的平均损失与新的,随机选择的任务之间的差异。本文介绍了元将军差距上的新信息理论上限。认为两种广泛的元学习算法被认为使用了单独的任务内训练和测试集,例如MAML或联合任务内培训和测试集,例如爬行动物。扩展了用于常规学习的现有工作,对以前的类别的元元化差距上的上限依赖于元学习算法的输出与其输入元训练数据之间的相互信息(MI)。对于后者,派生的界限包括每个任务学习过程的输出与相应的数据集之间的额外MI,以捕获任务内的不确定性。然后,在给定的技术条件下,通过新型的个人任务MI(ITMI)边界开发了更紧密的界限。最终讨论了派生界限的应用,包括用于元学习的一系列嘈杂的迭代算法。
Meta-learning, or "learning to learn", refers to techniques that infer an inductive bias from data corresponding to multiple related tasks with the goal of improving the sample efficiency for new, previously unobserved, tasks. A key performance measure for meta-learning is the meta-generalization gap, that is, the difference between the average loss measured on the meta-training data and on a new, randomly selected task. This paper presents novel information-theoretic upper bounds on the meta-generalization gap. Two broad classes of meta-learning algorithms are considered that uses either separate within-task training and test sets, like MAML, or joint within-task training and test sets, like Reptile. Extending the existing work for conventional learning, an upper bound on the meta-generalization gap is derived for the former class that depends on the mutual information (MI) between the output of the meta-learning algorithm and its input meta-training data. For the latter, the derived bound includes an additional MI between the output of the per-task learning procedure and corresponding data set to capture within-task uncertainty. Tighter bounds are then developed, under given technical conditions, for the two classes via novel Individual Task MI (ITMI) bounds. Applications of the derived bounds are finally discussed, including a broad class of noisy iterative algorithms for meta-learning.