论文标题

对手训练的神经网络的重点是什么:基于傅立叶领域的研究

What Do Adversarially trained Neural Networks Focus: A Fourier Domain-based Study

论文作者

Huang, Binxiao, Tao, Chaofan, Lin, Rui, Wong, Ngai

论文摘要

尽管许多领域都见证了深度学习带来的优越表现,但神经网络的鲁棒性仍然是一个悬而未决的问题。具体而言,输入的小对抗扰动可能会导致模型产生完全不同的输出。如此差的鲁棒性意味着许多潜在的危害,尤其是在关键安全应用中,例如自动驾驶和移动机器人技术。这项工作研究了对手训练的模型的重点是哪些信息。从经验上讲,我们注意到清洁数据和对抗数据之间的差异主要分布在低频区域中。然后,我们发现,受对抗训练的模型比其自然训练的模型更强大,因为前者更多地关注低频组件中的主要信息。此外,我们考虑了两种改善模型鲁棒性的常见方法,即通过数据增强和使用更强的网络体系结构,并从频域的角度了解这些技术。我们希望这项工作能够阐明更强大的神经网络的设计。

Although many fields have witnessed the superior performance brought about by deep learning, the robustness of neural networks remains an open issue. Specifically, a small adversarial perturbation on the input may cause the model to produce a completely different output. Such poor robustness implies many potential hazards, especially in security-critical applications, e.g., autonomous driving and mobile robotics. This work studies what information the adversarially trained model focuses on. Empirically, we notice that the differences between the clean and adversarial data are mainly distributed in the low-frequency region. We then find that an adversarially-trained model is more robust than its naturally-trained counterpart due to the reason that the former pays more attention to learning the dominant information in low-frequency components. In addition, we consider two common ways to improve model robustness, namely, by data augmentation and by using stronger network architectures, and understand these techniques from a frequency-domain perspective. We are hopeful this work can shed light on the design of more robust neural networks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源