论文标题
持续及时调整对话框跟踪
Continual Prompt Tuning for Dialog State Tracking
论文作者
论文摘要
理想的对话系统应该能够在不忘记旧技能的情况下不断学习新技能,从而适应其生命周期中的新领域或任务。但是,不断训练模型通常会导致众所周知的灾难性遗忘问题。在本文中,我们提出了持续的及时调整,这是一个参数有效的框架,不仅避免忘记,而且还可以在任务之间进行知识转移。为了避免忘记,我们只学习并存储一些迅速令牌的每个任务的嵌入,同时冷冻骨架预训练的模型。为了实现任务之间的双向知识转移,我们提出了几种技术(持续的及时初始化,查询融合和内存重播),从先前任务和内存引导的技术转移知识,以从后续任务转移知识。与最先进的基线相比,广泛的实验证明了我们提出的方法在对话状态跟踪的持续学习中的有效性和效率。
A desirable dialog system should be able to continually learn new skills without forgetting old ones, and thereby adapt to new domains or tasks in its life cycle. However, continually training a model often leads to a well-known catastrophic forgetting issue. In this paper, we present Continual Prompt Tuning, a parameter-efficient framework that not only avoids forgetting but also enables knowledge transfer between tasks. To avoid forgetting, we only learn and store a few prompt tokens' embeddings for each task while freezing the backbone pre-trained model. To achieve bi-directional knowledge transfer among tasks, we propose several techniques (continual prompt initialization, query fusion, and memory replay) to transfer knowledge from preceding tasks and a memory-guided technique to transfer knowledge from subsequent tasks. Extensive experiments demonstrate the effectiveness and efficiency of our proposed method on continual learning for dialog state tracking, compared with state-of-the-art baselines.