论文标题
自我监督的对比学习,以在对话代理中有效的用户满意度预测
Self-Supervised Contrastive Learning for Efficient User Satisfaction Prediction in Conversational Agents
论文作者
论文摘要
转向级用户满意度是对话代理的最重要的性能指标之一。它可用于监视代理商的性能,并提供有关用户体验有缺陷的见解。此外,强大的满意度模型可以用作对话代理不断优化的目标函数。尽管端到端的深度学习显示出令人鼓舞的结果,但可以访问这些方法所需的大量可靠的带注释的样本仍然具有挑战性。在大规模的对话系统中,由于所需的注释成本以及周转时间,使传统的数据收集,注释和建模过程变得不切实际,因此有越来越多的新开发技能。在本文中,我们建议一种自我监督的对比学习方法,该方法利用未标记的数据来学习用户代理交互。我们表明,使用自我监督目标的预训练模型可以转移到用户满意度预测中。此外,我们提出了一种新颖的几次转移学习方法,可确保针对非常小的样本量的更好的传递性。建议的几种方法不需要任何内部循环优化过程,并且可扩展到非常大的数据集和复杂模型。基于我们使用来自大型商业系统的现实世界数据的实验,建议的方法能够显着减少所需的注释数量,同时改善对看不见的室外技能的概括。
Turn-level user satisfaction is one of the most important performance metrics for conversational agents. It can be used to monitor the agent's performance and provide insights about defective user experiences. Moreover, a powerful satisfaction model can be used as an objective function that a conversational agent continuously optimizes for. While end-to-end deep learning has shown promising results, having access to a large number of reliable annotated samples required by these methods remains challenging. In a large-scale conversational system, there is a growing number of newly developed skills, making the traditional data collection, annotation, and modeling process impractical due to the required annotation costs as well as the turnaround times. In this paper, we suggest a self-supervised contrastive learning approach that leverages the pool of unlabeled data to learn user-agent interactions. We show that the pre-trained models using the self-supervised objective are transferable to the user satisfaction prediction. In addition, we propose a novel few-shot transfer learning approach that ensures better transferability for very small sample sizes. The suggested few-shot method does not require any inner loop optimization process and is scalable to very large datasets and complex models. Based on our experiments using real-world data from a large-scale commercial system, the suggested approach is able to significantly reduce the required number of annotations, while improving the generalization on unseen out-of-domain skills.