论文标题

过度参数化对抗训练:克服维度诅咒的分析

Over-parameterized Adversarial Training: An Analysis Overcoming the Curse of Dimensionality

论文作者

Zhang, Yi, Plevrakis, Orestis, Du, Simon S., Li, Xingguo, Song, Zhao, Arora, Sanjeev

论文摘要

对抗性训练是一种流行的方法,可以使神经网络稳健性针对对抗性扰动。实际上,对抗训练会导致较低的强大训练损失。但是,关于为什么在自然条件下发生这种情况的严格解释仍然缺失。最近,各组对标准(非对抗性)监督培训的收敛理论开发了{\ em非常兼容的}网。目前尚不清楚如何将这些结果扩展到最终最大目标,将这些结果扩展到对抗性训练。最近,Gao等人迈出了第一步。使用在线学习中的工具,但是它们要求在输入尺寸$ d $中的网络宽度为\ emph {指数},并且具有不自然的激活功能。我们的工作证明了\ emph {多项式}宽度,而不是指数级,在自然假设和relu激活下,融合了\ emph {polyenmial}宽度的稳健训练损失。我们证明的关键要素表明,接近初始化的Relu网络可以近似步骤函数,这可能具有独立的关注。

Adversarial training is a popular method to give neural nets robustness against adversarial perturbations. In practice adversarial training leads to low robust training loss. However, a rigorous explanation for why this happens under natural conditions is still missing. Recently a convergence theory for standard (non-adversarial) supervised training was developed by various groups for {\em very overparametrized} nets. It is unclear how to extend these results to adversarial training because of the min-max objective. Recently, a first step towards this direction was made by Gao et al. using tools from online learning, but they require the width of the net to be \emph{exponential} in input dimension $d$, and with an unnatural activation function. Our work proves convergence to low robust training loss for \emph{polynomial} width instead of exponential, under natural assumptions and with the ReLU activation. Key element of our proof is showing that ReLU networks near initialization can approximate the step function, which may be of independent interest.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源