论文标题

概念解释器:从概念角度来看深神经网络的交互式解释

ConceptExplainer: Interactive Explanation for Deep Neural Networks from a Concept Perspective

论文作者

Huang, Jinbin, Mishra, Aditi, Kwon, Bum Chul, Bryan, Chris

论文摘要

适合模型用户的传统深度学习可解释性方法无法在全球层面解释网络行为,并且在提供细粒度的解释方面不灵活。作为解决方案,基于概念的解释由于其人类的直觉以及描述全球和本地模型行为的灵活性而引起了人们的关注。概念是类似有意义的像素的群体,它们表达了一个概念,该概念嵌入了网络潜在空间中,并且通常是手工生成的,但最近是通过自动化方法发现的。不幸的是,发现概念的大小和多样性使得难以导航并理解概念空间。视觉分析可以通过启用结构化导航和探索概念空间来为用户提供基于概念的模型行为见解,从而在弥合这些空白方面发挥重要作用。为此,我们设计,开发和验证概念插图,这是一个视觉分析系统,使人们能够交互探测和探索概念空间,以在实例/类别/全球级别上解释模型行为。该系统是通过迭代原型开发的,以应对用户在解释深度学习模型的行为时面临的许多设计挑战。通过严格的用户研究,我们验证了ConceptExplainer如何支持这些挑战。同样,我们进行了一系列用法方案,以演示系统如何支持对各种任务和解释粒度跨模型行为的互动分析,例如识别对分类很重要的概念,识别培训数据中的偏见以及如何在多元化和看似异常的类别中共享概念如何共享概念。

Traditional deep learning interpretability methods which are suitable for model users cannot explain network behaviors at the global level and are inflexible at providing fine-grained explanations. As a solution, concept-based explanations are gaining attention due to their human intuitiveness and their flexibility to describe both global and local model behaviors. Concepts are groups of similarly meaningful pixels that express a notion, embedded within the network's latent space and have commonly been hand-generated, but have recently been discovered by automated approaches. Unfortunately, the magnitude and diversity of discovered concepts makes it difficult to navigate and make sense of the concept space. Visual analytics can serve a valuable role in bridging these gaps by enabling structured navigation and exploration of the concept space to provide concept-based insights of model behavior to users. To this end, we design, develop, and validate ConceptExplainer, a visual analytics system that enables people to interactively probe and explore the concept space to explain model behavior at the instance/class/global level. The system was developed via iterative prototyping to address a number of design challenges that model users face in interpreting the behavior of deep learning models. Via a rigorous user study, we validate how ConceptExplainer supports these challenges. Likewise, we conduct a series of usage scenarios to demonstrate how the system supports the interactive analysis of model behavior across a variety of tasks and explanation granularities, such as identifying concepts that are important to classification, identifying bias in training data, and understanding how concepts can be shared across diverse and seemingly dissimilar classes.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源