论文标题
对手的例子和指标
Adversarial Examples and Metrics
论文作者
论文摘要
对抗性示例是对机器学习(ML)系统的一种攻击,导致输入错误分类。实现对抗性实例的鲁棒性对于在现实世界中应用ML至关重要。虽然大多数先前在对抗性例子上的工作都是经验的,但最近的工作却建立了基于密码硬度的强大分类的基本局限性。但是,该领域中最积极和负面的结果假设有一个固定的目标度量来限制对手,我们认为这通常是一个不切实际的假设。在这项工作中,如果目标度量尚不确定,我们研究了鲁棒分类的局限性。具体而言,我们构建了一个分类问题,如果在训练模型时已知目标指标,该分类问题将接受小分类器的鲁棒分类,但是如果在事实之后选择了目标度量标准,则小分类器不可能进行鲁棒分类。在此过程中,我们探索了鲁棒分类和有限存储模型加密之间的新颖联系。
Adversarial examples are a type of attack on machine learning (ML) systems which cause misclassification of inputs. Achieving robustness against adversarial examples is crucial to apply ML in the real world. While most prior work on adversarial examples is empirical, a recent line of work establishes fundamental limitations of robust classification based on cryptographic hardness. Most positive and negative results in this field however assume that there is a fixed target metric which constrains the adversary, and we argue that this is often an unrealistic assumption. In this work we study the limitations of robust classification if the target metric is uncertain. Concretely, we construct a classification problem, which admits robust classification by a small classifier if the target metric is known at the time the model is trained, but for which robust classification is impossible for small classifiers if the target metric is chosen after the fact. In the process, we explore a novel connection between hardness of robust classification and bounded storage model cryptography.