论文标题

神经对话生成的小组对比度学习

Group-wise Contrastive Learning for Neural Dialogue Generation

论文作者

Cai, Hengyi, Chen, Hongshen, Song, Yonghao, Ding, Zhuoye, Bao, Yongjun, Yan, Weipeng, Zhao, Xiaofang

论文摘要

近年来,神经对话反应的产生已广受欢迎。在现有的对话模型学习中广泛采用了最大似然估计(MLE)目标。但是,在开放域的对话设置方面,低多样性问题困扰了接受MLE目标功能训练的模型。受到观察的启发,即人类不仅从积极的信号中学习,而且还从纠正不良行动的行为中受益,在这项工作中,我们将对比度学习引入对话生成中,该模型明确地感知了精心挑选的积极话语和负面话语之间的差异。具体而言,我们采用验证的基线模型作为参考。在对比度学习过程中,与参考模型相比,训练目标对话模型为阳性样本提供更高的条件概率,并为这些负样本提供较低的条件概率。为了管理人类对话中盛行的多映射关系,我们通过小组双重采样来增强对比对话学习。广泛的实验结果表明,拟议的小组对比学习框架适合训练广泛的神经对话生成模型,其性能比基线训练方法非常有利。

Neural dialogue response generation has gained much popularity in recent years. Maximum Likelihood Estimation (MLE) objective is widely adopted in existing dialogue model learning. However, models trained with MLE objective function are plagued by the low-diversity issue when it comes to the open-domain conversational setting. Inspired by the observation that humans not only learn from the positive signals but also benefit from correcting behaviors of undesirable actions, in this work, we introduce contrastive learning into dialogue generation, where the model explicitly perceives the difference between the well-chosen positive and negative utterances. Specifically, we employ a pretrained baseline model as a reference. During contrastive learning, the target dialogue model is trained to give higher conditional probabilities for the positive samples, and lower conditional probabilities for those negative samples, compared to the reference model. To manage the multi-mapping relations prevailed in human conversation, we augment contrastive dialogue learning with group-wise dual sampling. Extensive experimental results show that the proposed group-wise contrastive learning framework is suited for training a wide range of neural dialogue generation models with very favorable performance over the baseline training approaches.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源