论文标题

针对复发性神经网络的模型提取攻击

Model Extraction Attacks against Recurrent Neural Networks

论文作者

Takemura, Tatsuya, Yanai, Naoto, Fujiwara, Toru

论文摘要

模型提取攻击是一种攻击,在这种攻击中,对手获得了一种新模型,其性能与目标模型相当,通过查询对目标模型有效的访问,即比目标模型的数据集和计算资源更少。现有作品仅处理了简单的深神经网络(DNN),例如,只有三层,作为模型提取攻击的目标,因此并不意识到复发性神经网络(RNN)在处理时间序列数据方面的有效性。在这项工作中,我们阐明了针对RNN的模型提取攻击的威胁。我们讨论是否可以使用长期短期内存(LSTM)的简单RNN提取具有较高精度的模型,这是一个更复杂和功能强大的RNN。具体来说,我们解决了以下问题。首先,在分类问题(例如图像识别)的情况下,通过在序列中途使用输出来提出没有最终输出的RNN模型。接下来,在回归问题的情况下。例如在天气预报中,提出了新配置损失功能的新攻击。我们对针对RNN的模型提取攻击和经过公开可用的学术数据集培训的LSTM进行了实验。然后,我们证明可以有效提取具有更高精度的模型,尤其是通过配置损失函数和与目标模型不同的复杂体系结构。

Model extraction attacks are a kind of attacks in which an adversary obtains a new model, whose performance is equivalent to that of a target model, via query access to the target model efficiently, i.e., fewer datasets and computational resources than those of the target model. Existing works have dealt with only simple deep neural networks (DNNs), e.g., only three layers, as targets of model extraction attacks, and hence are not aware of the effectiveness of recurrent neural networks (RNNs) in dealing with time-series data. In this work, we shed light on the threats of model extraction attacks against RNNs. We discuss whether a model with a higher accuracy can be extracted with a simple RNN from a long short-term memory (LSTM), which is a more complicated and powerful RNN. Specifically, we tackle the following problems. First, in a case of a classification problem, such as image recognition, extraction of an RNN model without final outputs from an LSTM model is presented by utilizing outputs halfway through the sequence. Next, in a case of a regression problem. such as in weather forecasting, a new attack by newly configuring a loss function is presented. We conduct experiments on our model extraction attacks against an RNN and an LSTM trained with publicly available academic datasets. We then show that a model with a higher accuracy can be extracted efficiently, especially through configuring a loss function and a more complex architecture different from the target model.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源