论文标题

通过补丁处理捍卫对视觉变压器的后门攻击

Defending Backdoor Attacks on Vision Transformer via Patch Processing

论文作者

Doan, Khoa D., Lao, Yingjie, Yang, Peng, Li, Ping

论文摘要

视觉变压器(VIT)具有根本不同的结构,其感应偏置明显少于卷积神经网络。随着绩效的提高,VIT的安全性和鲁棒性对于研究也非常重要。与许多最近的作品相反,这些作品利用了VIT的鲁棒性针对对抗性例子的鲁棒性,本文调查了代表性的病因攻击,即后门。我们首先检查了VIT对各种后门攻击的脆弱性,发现VIT也很容易受到现有攻击的影响。但是,我们观察到,VIT的清洁数据准确性和后门攻击成功率在位置编码之前对贴片变换做出了明显的反应。然后,根据这一发现,我们为VIT提出了一种通过补丁处理来捍卫基于补丁的触发后门攻击的有效方法。在包括CIFAR10,GTSRB和Tinyimagenet在内的几个基准数据集上评估了这些性能,这些数据表明,拟议中的新型防御在减轻VIT的后门攻击方面非常成功。据我们所知,本文提出了第一个防御性策略,该策略利用了反对后门攻击的独特特征。 该论文将出现在AAAI'23会议的会议记录中。这项工作最初于2021年11月提交给CVPR'22,然后将其重新提交到ECCV'22。该论文于2022年6月公开。作者真诚地感谢CVPR'22,ECCV'22和AAAI'23的计划委员会的所有裁判。

Vision Transformers (ViTs) have a radically different architecture with significantly less inductive bias than Convolutional Neural Networks. Along with the improvement in performance, security and robustness of ViTs are also of great importance to study. In contrast to many recent works that exploit the robustness of ViTs against adversarial examples, this paper investigates a representative causative attack, i.e., backdoor. We first examine the vulnerability of ViTs against various backdoor attacks and find that ViTs are also quite vulnerable to existing attacks. However, we observe that the clean-data accuracy and backdoor attack success rate of ViTs respond distinctively to patch transformations before the positional encoding. Then, based on this finding, we propose an effective method for ViTs to defend both patch-based and blending-based trigger backdoor attacks via patch processing. The performances are evaluated on several benchmark datasets, including CIFAR10, GTSRB, and TinyImageNet, which show the proposed novel defense is very successful in mitigating backdoor attacks for ViTs. To the best of our knowledge, this paper presents the first defensive strategy that utilizes a unique characteristic of ViTs against backdoor attacks. The paper will appear in the Proceedings of the AAAI'23 Conference. This work was initially submitted in November 2021 to CVPR'22, then it was re-submitted to ECCV'22. The paper was made public in June 2022. The authors sincerely thank all the referees from the Program Committees of CVPR'22, ECCV'22, and AAAI'23.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源