论文标题
消除不可能的一切,剩下的一切都必须是真实的
Eliminating The Impossible, Whatever Remains Must Be True
论文作者
论文摘要
AI做出预测和决策的方法的兴起导致对更可解释的人工智能(XAI)方法的迫切需求。 XAI的一种常见方法是产生事后解释,解释了为什么黑匣子ML模型做出一定的预测。正式的事后解释方法为为什么做出预测提供了简洁的原因,以及为什么不进行其他预测。但是这些方法假定特征是独立且均匀分布的。尽管这意味着“为什么”解释是正确的,但它们可能比要求更长。这也意味着“为什么不”解释可能会被怀疑,因为他们所依赖的反例可能没有意义。在本文中,我们展示了人们如何运用背景知识来提供更简洁的“为什么”形式解释,这些解释大概更容易被人类解释,并给出更准确的“为什么不”解释。此外,我们还展示了如何使用现有的规则归纳技术从数据集中有效提取背景信息,以及如何报告使用哪些背景信息来做出解释,如果人类怀疑解释的正确性,则可以对其进行检查。
The rise of AI methods to make predictions and decisions has led to a pressing need for more explainable artificial intelligence (XAI) methods. One common approach for XAI is to produce a post-hoc explanation, explaining why a black box ML model made a certain prediction. Formal approaches to post-hoc explanations provide succinct reasons for why a prediction was made, as well as why not another prediction was made. But these approaches assume that features are independent and uniformly distributed. While this means that "why" explanations are correct, they may be longer than required. It also means the "why not" explanations may be suspect as the counterexamples they rely on may not be meaningful. In this paper, we show how one can apply background knowledge to give more succinct "why" formal explanations, that are presumably easier to interpret by humans, and give more accurate "why not" explanations. In addition, we show how to use existing rule induction techniques to efficiently extract background information from a dataset, and also how to report which background information was used to make an explanation, allowing a human to examine it if they doubt the correctness of the explanation.