论文标题
辅助任务重新加权,以进行最低数据
Auxiliary Task Reweighting for Minimum-data Learning
论文作者
论文摘要
监督的学习需要大量的培训数据,从而限制了其应用程序稀缺的应用。为了弥补数据稀缺性,一种可能的方法是利用辅助任务为主要任务提供其他监督。分配和优化对不同辅助任务的重要性权重仍然是一个至关重要且在很大程度上研究的研究问题。在这项工作中,我们提出了一种自动重新辅助任务的方法,以减少主要任务的数据要求。具体而言,我们将辅助任务的加权似然函数作为主要任务的先验。通过调整辅助任务权重,以最大程度地减少替代替代品与主要任务的真实先验之间的差异,我们获得了更准确的先验估计,实现了最大程度地减少主要任务所需培训数据并避免昂贵的网格搜索的目标。在多个实验设置(例如,半监督学习,多标签分类)中,我们证明我们的算法可以有效利用与先前的任务重新加权方法相比,具有辅助任务的有限的主要任务标记数据。我们还表明,在极端情况下,只有几个额外的例子(例如,少数域的适应性),我们的算法会显着改善基线。
Supervised learning requires a large amount of training data, limiting its application where labeled data is scarce. To compensate for data scarcity, one possible method is to utilize auxiliary tasks to provide additional supervision for the main task. Assigning and optimizing the importance weights for different auxiliary tasks remains an crucial and largely understudied research question. In this work, we propose a method to automatically reweight auxiliary tasks in order to reduce the data requirement on the main task. Specifically, we formulate the weighted likelihood function of auxiliary tasks as a surrogate prior for the main task. By adjusting the auxiliary task weights to minimize the divergence between the surrogate prior and the true prior of the main task, we obtain a more accurate prior estimation, achieving the goal of minimizing the required amount of training data for the main task and avoiding a costly grid search. In multiple experimental settings (e.g. semi-supervised learning, multi-label classification), we demonstrate that our algorithm can effectively utilize limited labeled data of the main task with the benefit of auxiliary tasks compared with previous task reweighting methods. We also show that under extreme cases with only a few extra examples (e.g. few-shot domain adaptation), our algorithm results in significant improvement over the baseline.