论文标题

学会通过有针对性的扰动欺骗知识增强模型

Learning to Deceive Knowledge Graph Augmented Models via Targeted Perturbation

论文作者

Raman, Mrigank, Chan, Aaron, Agarwal, Siddhant, Wang, Peifeng, Wang, Hansen, Kim, Sungchul, Rossi, Ryan, Zhao, Handong, Lipka, Nedim, Ren, Xiang

论文摘要

知识图(KGS)帮助神经模型改善了各种知识密集任务的性能,例如问答和项目建议。通过对KG上的关注,这样的KG扬名模型也可以“解释”哪种KG信息与做出给定的预测最相关。在本文中,我们质疑这些模型是否真的按照我们的期望。我们表明,通过强化学习政策(甚至简单的启发式方法),可以产生欺骗性的kg,从而保持原始KG的下游性能,同时显着偏离原始KG的语义和结构。我们的发现引起了人们对KG的模型推理KG信息并给出明智的解释的能力的怀疑。

Knowledge graphs (KGs) have helped neural models improve performance on various knowledge-intensive tasks, like question answering and item recommendation. By using attention over the KG, such KG-augmented models can also "explain" which KG information was most relevant for making a given prediction. In this paper, we question whether these models are really behaving as we expect. We show that, through a reinforcement learning policy (or even simple heuristics), one can produce deceptively perturbed KGs, which maintain the downstream performance of the original KG while significantly deviating from the original KG's semantics and structure. Our findings raise doubts about KG-augmented models' ability to reason about KG information and give sensible explanations.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源