论文标题

对伪装医学对抗攻击的分层特征约束

A Hierarchical Feature Constraint to Camouflage Medical Adversarial Attacks

论文作者

Yao, Qingsong, He, Zecheng, Lin, Yi, Ma, Kai, Zheng, Yefeng, Zhou, S. Kevin

论文摘要

用于医学图像的深神经网络(DNN)极易受到对抗性例子(AES)的影响,这引起了对临床决策的安全问题。幸运的是,根据我们的研究,医疗AE在层次特征空间中也很容易检测到。为了更好地理解这一现象,我们彻底研究了特征空间中医学AES的内在特征,为这个问题提供了经验证据和理论解释:为什么医学对抗性攻击易于检测?与自然图像相比,我们首先进行了压力测试,以揭示医学图像的深度表示的脆弱性。然后,我们从理论上证明,对二元疾病诊断网络的典型对抗性攻击通过在固定方向上连续优化脆弱的表示来操纵预测,从而产生了使医疗AES易于检测的异常特征。但是,也可以利用此漏洞将AE隐藏在功能空间中。我们提出了一种新颖的分层特征约束(HFC)作为现有的对抗攻击的附加组件,这鼓励了对抗性表示在正常特征分布中的隐藏。我们在两个公共医疗图像数据集上评估了所提出的方法,即{fundsoscopy}和{胸部X射线}。实验结果证明了我们的对抗性攻击方法的优势,因为它比竞争攻击方法更容易绕过一系列最先进的对抗探测器,这支持了医疗特征的极大脆弱性使攻击者更有空间来操纵对抗表示。

Deep neural networks (DNNs) for medical images are extremely vulnerable to adversarial examples (AEs), which poses security concerns on clinical decision making. Luckily, medical AEs are also easy to detect in hierarchical feature space per our study herein. To better understand this phenomenon, we thoroughly investigate the intrinsic characteristic of medical AEs in feature space, providing both empirical evidence and theoretical explanations for the question: why are medical adversarial attacks easy to detect? We first perform a stress test to reveal the vulnerability of deep representations of medical images, in contrast to natural images. We then theoretically prove that typical adversarial attacks to binary disease diagnosis network manipulate the prediction by continuously optimizing the vulnerable representations in a fixed direction, resulting in outlier features that make medical AEs easy to detect. However, this vulnerability can also be exploited to hide the AEs in the feature space. We propose a novel hierarchical feature constraint (HFC) as an add-on to existing adversarial attacks, which encourages the hiding of the adversarial representation within the normal feature distribution. We evaluate the proposed method on two public medical image datasets, namely {Fundoscopy} and {Chest X-Ray}. Experimental results demonstrate the superiority of our adversarial attack method as it bypasses an array of state-of-the-art adversarial detectors more easily than competing attack methods, supporting that the great vulnerability of medical features allows an attacker more room to manipulate the adversarial representations.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源