论文标题
学习多任务学习更好的神经机器翻译
Learning to Multi-Task Learn for Better Neural Machine Translation
论文作者
论文摘要
平行句子对的稀缺是训练双层低资源场景中高质量神经机器翻译(NMT)模型的主要挑战,因为NMT是渴望数据的。多任务学习是一种优雅的方法,可以使用辅助句法和语义任务将与语言相关的电感偏置注入NMT,以改善概括。但是,面临的挑战是设计有效的培训时间表,规定何时在培训过程中使用辅助任务,以填补主要翻译任务的知识差距,该设置称为有偏见的MTL。当前的培训时间表方法是基于手工设计的启发式方法,其有效性在不同的MTL设置中有所不同。我们建议一个学习培训时间表的新颖框架,即学习多任务学习以及感兴趣的MTL设置。我们将培训时间表作为马尔可夫决策过程,铺平了采用策略学习方法来学习调度策略的方式。我们使用Oracle策略算法有效,高效地学习模仿学习框架内的培训时间表策略,该算法会根据其对主要NMT任务的普遍性的贡献,动态设置辅助任务的重要性权重。低资源NMT设置的实验表明,由此产生的自动学习的培训调度程序具有最佳的启发式竞争力,并提高了+1.1 BLEU得分。
Scarcity of parallel sentence pairs is a major challenge for training high quality neural machine translation (NMT) models in bilingually low-resource scenarios, as NMT is data-hungry. Multi-task learning is an elegant approach to inject linguistic-related inductive biases into NMT, using auxiliary syntactic and semantic tasks, to improve generalisation. The challenge, however, is to devise effective training schedules, prescribing when to make use of the auxiliary tasks during the training process to fill the knowledge gaps of the main translation task, a setting referred to as biased-MTL. Current approaches for the training schedule are based on hand-engineering heuristics, whose effectiveness vary in different MTL settings. We propose a novel framework for learning the training schedule, ie learning to multi-task learn, for the MTL setting of interest. We formulate the training schedule as a Markov decision process which paves the way to employ policy learning methods to learn the scheduling policy. We effectively and efficiently learn the training schedule policy within the imitation learning framework using an oracle policy algorithm that dynamically sets the importance weights of auxiliary tasks based on their contributions to the generalisability of the main NMT task. Experiments on low-resource NMT settings show the resulting automatically learned training schedulers are competitive with the best heuristics, and lead to up to +1.1 BLEU score improvements.