论文标题

学习模型中的边界厚度和稳健性

Boundary thickness and robustness in learning models

论文作者

Yang, Yaoqing, Khanna, Rajiv, Yu, Yaodong, Gholami, Amir, Keutzer, Kurt, Gonzalez, Joseph E., Ramchandran, Kannan, Mahoney, Michael W.

论文摘要

机器学习模型对各种对抗性和非对抗性腐败的鲁棒性仍然引起人们的关注。在本文中,我们介绍了分类器的边界厚度的概念,并描述了其与模型鲁棒性的联系和有用性。厚厚的决策边界会提高性能,而薄的决策边界会导致过度拟合(例如,通过训练和测试之间的稳定概括差距来衡量)和较低的鲁棒性。我们表明,较厚的边界有助于提高针对对抗性示例的鲁棒性(例如,提高对抗训练的鲁棒测试准确性)以及所谓的分布外(OOD)变换,我们表明许多常用的正则化和数据增强程序可以增加边界厚度。在理论方面,我们确定在训练过程中最大化边界厚度类似于所谓的混合训练。使用这些观察结果,我们表明,混合训练中的噪声提示进一步增加了边界厚度,从而打击了对各种形式的对抗攻击和OOD变换的脆弱性。我们还可以证明,最近的几项工作的性能提高与边界较厚的边界结合发生。

Robustness of machine learning models to various adversarial and non-adversarial corruptions continues to be of interest. In this paper, we introduce the notion of the boundary thickness of a classifier, and we describe its connection with and usefulness for model robustness. Thick decision boundaries lead to improved performance, while thin decision boundaries lead to overfitting (e.g., measured by the robust generalization gap between training and testing) and lower robustness. We show that a thicker boundary helps improve robustness against adversarial examples (e.g., improving the robust test accuracy of adversarial training) as well as so-called out-of-distribution (OOD) transforms, and we show that many commonly-used regularization and data augmentation procedures can increase boundary thickness. On the theoretical side, we establish that maximizing boundary thickness during training is akin to the so-called mixup training. Using these observations, we show that noise-augmentation on mixup training further increases boundary thickness, thereby combating vulnerability to various forms of adversarial attacks and OOD transforms. We can also show that the performance improvement in several lines of recent work happens in conjunction with a thicker boundary.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源