论文标题
对图神经网络中的解释评估进行解释
Deconfounding to Explanation Evaluation in Graph Neural Networks
论文作者
论文摘要
图神经网络(GNN)的解释性旨在回答“为什么GNN做出一定的预测?”,这对于解释模型预测至关重要。功能归因框架将GNN的预测分配给其输入特征(例如边缘),将有影响力的子图识别为解释。在评估解释(即子图重要性)时,一种标准方法是仅根据子图审核模型预测。但是,我们认为在完整图和子图之间存在分配转移,从而导致分布外问题。此外,通过深入的因果分析,我们发现OOD效应起着混杂因素的作用,从而带来了子图重要性和模型预测之间的虚假关联,从而使评估的可靠性降低了。在这项工作中,我们提出了反对的子图评估(DSE),该评估评估了解释性子图对模型预测的因果影响。虽然分布转移通常是棘手的,但我们采用了前门调整,并引入了子图的替代变量。具体而言,我们设计了一个生成模型,以生成符合数据分布的合理替代物,从而接近对亚图重要性的无偏估计。经验结果证明了DSE在解释保真度方面的有效性。
Explainability of graph neural networks (GNNs) aims to answer "Why the GNN made a certain prediction?", which is crucial to interpret the model prediction. The feature attribution framework distributes a GNN's prediction to its input features (e.g., edges), identifying an influential subgraph as the explanation. When evaluating the explanation (i.e., subgraph importance), a standard way is to audit the model prediction based on the subgraph solely. However, we argue that a distribution shift exists between the full graph and the subgraph, causing the out-of-distribution problem. Furthermore, with an in-depth causal analysis, we find the OOD effect acts as the confounder, which brings spurious associations between the subgraph importance and model prediction, making the evaluation less reliable. In this work, we propose Deconfounded Subgraph Evaluation (DSE) which assesses the causal effect of an explanatory subgraph on the model prediction. While the distribution shift is generally intractable, we employ the front-door adjustment and introduce a surrogate variable of the subgraphs. Specifically, we devise a generative model to generate the plausible surrogates that conform to the data distribution, thus approaching the unbiased estimation of subgraph importance. Empirical results demonstrate the effectiveness of DSE in terms of explanation fidelity.