论文标题

通过目标标签指导的深神经网络的加速鲁棒性验证

Accelerating Robustness Verification of Deep Neural Networks Guided by Target Labels

论文作者

Wan, Wenjie, Zhang, Zhaodi, Zhu, Yiwei, Zhang, Min, Song, Fu

论文摘要

深度神经网络(DNN)已成为许多安全性应用程序(例如自动驾驶和医疗诊断)的关键组成部分。但是,由于其对对抗性示例的敏感性,因此对DNN的稳健性很差,因此对输入的小扰动会导致错误预测。关于这种关注,已经提出了各种方法来正式验证DNN的鲁棒性。这些方法中的大多数将验证问题降低到为给定输入搜索对抗性示例的优化问题,以使其未正确分类到原始标签。但是,它们的准确性和可伸缩性受到限制。在本文中,我们提出了一种新颖的方法,可以通过引导目标标签来验证来加速鲁棒性验证技术。我们方法的关键见解是,可以通过验证DNN的子问题,每个目标标签一个DNN的鲁棒性验证问题可以解决。在验证过程中固定目标标签可以大大降低搜索空间,从而提高效率。我们还通过利用符号间隔传播和线性松弛技术来提出一种方法,以根据对抗性实例存在的机会对目标标签进行分类。这通常使我们能够快速伪造DNN的鲁棒性,并且可以避免剩余目标标签的验证。我们的方法是正交的,并且可以与许多现有验证技术集成。出于评估目的,我们将其与最近有前途的DNN验证工具(即Mipverify,Deepz和神经化)集成在一起。实验结果表明,当扰动距离设置为合理范围时,我们的方法可以通过36倍加速显着改善这些工具。

Deep Neural Networks (DNNs) have become key components of many safety-critical applications such as autonomous driving and medical diagnosis. However, DNNs have been shown suffering from poor robustness because of their susceptibility to adversarial examples such that small perturbations to an input result in misprediction. Addressing to this concern, various approaches have been proposed to formally verify the robustness of DNNs. Most of these approaches reduce the verification problem to optimization problems of searching an adversarial example for a given input so that it is not correctly classified to the original label. However, they are limited in accuracy and scalability. In this paper, we propose a novel approach that can accelerate the robustness verification techniques by guiding the verification with target labels. The key insight of our approach is that the robustness verification problem of DNNs can be solved by verifying sub-problems of DNNs, one per target label. Fixing the target label during verification can drastically reduce the search space and thus improve the efficiency. We also propose an approach by leveraging symbolic interval propagation and linear relaxation techniques to sort the target labels in terms of chances that adversarial examples exist. This often allows us to quickly falsify the robustness of DNNs and the verification for remaining target labels could be avoided. Our approach is orthogonal to, and can be integrated with, many existing verification techniques. For evaluation purposes, we integrate it with three recent promising DNN verification tools, i.e., MipVerify, DeepZ, and Neurify. Experimental results show that our approach can significantly improve these tools by 36X speedup when the perturbation distance is set in a reasonable range.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源