论文标题
几次对话状态跟踪的双重提示学习框架
A Dual Prompt Learning Framework for Few-Shot Dialogue State Tracking
论文作者
论文摘要
对话状态跟踪(DST)模块是面向任务的对话系统的重要组件,以了解用户的目标和需求。收集包括插槽和价值在内的对话状态标签可能会很昂贵,尤其是在对话系统中广泛应用在越来越多的新升级域中。在本文中,我们关注如何利用预先训练语言模型的DST的语言理解和产生能力。我们为少数射击DST设计了双重提示学习框架。具体而言,我们将学习插槽生成和价值生成的学习视为双重任务,并且基于这种双重结构设计了两个提示,以分别结合与任务相关的这两个任务的知识。通过这种方式,DST任务可以在几个射击设置下有效地为语言建模任务进行配制。两个面向任务的对话数据集的实验结果表明,所提出的方法不仅胜过现有的最新方法,而且可以生成看不见的插槽。这表明可以从PLM探索与DST相关的知识,并在及时学习的帮助下用来有效地解决低资源DST。
Dialogue state tracking (DST) module is an important component for task-oriented dialog systems to understand users' goals and needs. Collecting dialogue state labels including slots and values can be costly, especially with the wide application of dialogue systems in more and more new-rising domains. In this paper, we focus on how to utilize the language understanding and generation ability of pre-trained language models for DST. We design a dual prompt learning framework for few-shot DST. Specifically, we consider the learning of slot generation and value generation as dual tasks, and two prompts are designed based on such a dual structure to incorporate task-related knowledge of these two tasks respectively. In this way, the DST task can be formulated as a language modeling task efficiently under few-shot settings. Experimental results on two task-oriented dialogue datasets show that the proposed method not only outperforms existing state-of-the-art few-shot methods, but also can generate unseen slots. It indicates that DST-related knowledge can be probed from PLM and utilized to address low-resource DST efficiently with the help of prompt learning.