论文标题
针对神经二进制功能检测的黑盒攻击
Black-box Attacks Against Neural Binary Function Detection
论文作者
论文摘要
近年来,基于深神经网络(DNN)或神经二进制分析(NBA)的二进制分析已成为近年来经过热门研究的主题。 DNN在推动自然语言和图像处理领域的性能和准确性信封方面取得了巨大成功。因此,DNN对于解决二进制分析问题的高度有希望,这些问题通常由于缺乏损失的汇编过程而难以进行的二进制分析问题。尽管有这样的承诺,但尚不清楚,鉴于二进制分析经常运行的对抗性环境,重新利用嵌入和模型架构的主要策略是听起来的。 在本文中,我们从经验上证明了神经功能边界检测中的当前最新状态很容易受到无意和故意的对抗攻击的影响。我们从洞察力的见解中进行,即当前一代NBA是建立在旨在解决句法问题的嵌入和模型体系结构上的。我们设计了一种简单,可再现和可扩展的黑框方法,用于探索无意攻击的空间 - 可以通过常见编译器工具链和配置发出的指令序列 - 利用了这种句法设计的重点。然后,我们证明这些无意间的错误分类可以由攻击者利用,这是高效的黑盒对抗示例生成过程的基础。我们针对两个最新的神经功能边界检测器评估了这种方法:XDA和DEEPDI。最后,我们对评估数据的分析和建议分析了未来研究如何避免屈服于类似攻击的建议。
Binary analyses based on deep neural networks (DNNs), or neural binary analyses (NBAs), have become a hotly researched topic in recent years. DNNs have been wildly successful at pushing the performance and accuracy envelopes in the natural language and image processing domains. Thus, DNNs are highly promising for solving binary analysis problems that are typically hard due to a lack of complete information resulting from the lossy compilation process. Despite this promise, it is unclear that the prevailing strategy of repurposing embeddings and model architectures originally developed for other problem domains is sound given the adversarial contexts under which binary analysis often operates. In this paper, we empirically demonstrate that the current state of the art in neural function boundary detection is vulnerable to both inadvertent and deliberate adversarial attacks. We proceed from the insight that current generation NBAs are built upon embeddings and model architectures intended to solve syntactic problems. We devise a simple, reproducible, and scalable black-box methodology for exploring the space of inadvertent attacks - instruction sequences that could be emitted by common compiler toolchains and configurations - that exploits this syntactic design focus. We then show that these inadvertent misclassifications can be exploited by an attacker, serving as the basis for a highly effective black-box adversarial example generation process. We evaluate this methodology against two state-of-the-art neural function boundary detectors: XDA and DeepDi. We conclude with an analysis of the evaluation data and recommendations for how future research might avoid succumbing to similar attacks.