论文标题

人类在循环抽象的对话摘要

Human-in-the-loop Abstractive Dialogue Summarization

论文作者

Chen, Jiaao, Dodda, Mohan, Yang, Diyi

论文摘要

抽象性对话摘要最近受到了越来越多的关注。尽管事实上,当前大多数对话摘要系统都经过培训,以最大程度地提高人类写的摘要的可能性并取得了重大成果,但产生高质量的摘要仍然存在很大的差距,这是由人类确定的,例如连贯性和忠诚,部分原因是使单个人为概述最大程度地造成了最大程度的摘要。为此,我们建议将不同水平的人类反馈纳入培训过程。这将使我们能够指导模型捕获人类关心摘要的行为。具体来说,我们要求人类强调要在摘要中包含的显着信息,以提供当地的反馈,并在一致性,准确性,覆盖率,简洁和整体质量方面对摘要进行整体比较,以此作为全球反馈。然后,我们将本地和全球反馈结合在一起,以将对话汇总政策与强化学习微调。在多个数据集上进行的实验证明了我们方法对最先进的监督基线的有效性和概括性,尤其是在人类判断方面。

Abstractive dialogue summarization has received increasing attention recently. Despite the fact that most of the current dialogue summarization systems are trained to maximize the likelihood of human-written summaries and have achieved significant results, there is still a huge gap in generating high-quality summaries as determined by humans, such as coherence and faithfulness, partly due to the misalignment in maximizing a single human-written summary. To this end, we propose to incorporate different levels of human feedback into the training process. This will enable us to guide the models to capture the behaviors humans care about for summaries. Specifically, we ask humans to highlight the salient information to be included in summaries to provide the local feedback , and to make overall comparisons among summaries in terms of coherence, accuracy, coverage, concise and overall quality, as the global feedback. We then combine both local and global feedback to fine-tune the dialog summarization policy with Reinforcement Learning. Experiments conducted on multiple datasets demonstrate the effectiveness and generalization of our methods over the state-of-the-art supervised baselines, especially in terms of human judgments.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源