论文标题
通过负蒸馏使神经对话产生多样化
Diversifying Neural Dialogue Generation via Negative Distillation
论文作者
论文摘要
生成对话模型由于通用响应问题而遭受严重影响,将其应用限制在一些玩具场景中。最近,提出了一种有趣的方法,即负面培训,以减轻该问题,通过提醒该模型在培训期间不要产生高频反应。但是,其性能受到两个问题的阻碍,忽略了低频但通用的反应,而带来了低频但毫无意义的回应。在本文中,我们提出了一种新型的负面训练范式,称为负蒸馏,以使模型远离不良通用响应,同时避免上述问题。首先,我们引入了一个负面的教师模型,该模型可以产生查询的通用响应,然后需要学生模型以最大程度地利用多层次的负面知识来最大化距离。经验结果表明,我们的方法的表现大大优于先前的负训练方法。
Generative dialogue models suffer badly from the generic response problem, limiting their applications to a few toy scenarios. Recently, an interesting approach, namely negative training, has been proposed to alleviate this problem by reminding the model not to generate high-frequency responses during training. However, its performance is hindered by two issues, ignoring low-frequency but generic responses and bringing low-frequency but meaningless responses. In this paper, we propose a novel negative training paradigm, called negative distillation, to keep the model away from the undesirable generic responses while avoiding the above problems. First, we introduce a negative teacher model that can produce query-wise generic responses, and then the student model is required to maximize the distance with multi-level negative knowledge. Empirical results show that our method outperforms previous negative training methods significantly.