论文标题

学会忘记:通过神经元掩盖的机器学习

Learn to Forget: Machine Unlearning via Neuron Masking

论文作者

Liu, Yang, Ma, Zhuo, Liu, Ximeng, Liu, Jian, Jiang, Zhongyuan, Ma, Jianfeng, Yu, Philip, Ren, Kui

论文摘要

如今,机器学习模型,尤其是神经网络,在许多现实世界应用中都普遍存在。这些模型是根据用户数据的单向旅行进行培训的:只要用户贡献他们的数据,就无法撤回;众所周知,神经网络记住其培训数据。这与GDPR的“被遗忘的权利”相矛盾,这可能导致违反法律。为此,机器的学习成为一个流行的研究主题,该主题允许用户从训练有素的机器学习模型中消除其私人数据的记忆。在本文中,我们提出了第一个称为“伪造率”的统一指标,以衡量机器未学习方法的有效性。它基于会员推理的概念,并描述了在进行学习后从“记忆”到“未知”的消除数据的转换率。我们还提出了一种称为Forsaken的新颖学习方法。它比以前的效用或效率(达到相同的遗忘率时)优越。我们使用八个标准数据集进行基准验证,以评估其性能。实验结果表明,它可以平均达到超过90 \%遗忘的速度,而无要点仅比5 \%的精度损失。

Nowadays, machine learning models, especially neural networks, become prevalent in many real-world applications.These models are trained based on a one-way trip from user data: as long as users contribute their data, there is no way to withdraw; and it is well-known that a neural network memorizes its training data. This contradicts the "right to be forgotten" clause of GDPR, potentially leading to law violations. To this end, machine unlearning becomes a popular research topic, which allows users to eliminate memorization of their private data from a trained machine learning model.In this paper, we propose the first uniform metric called for-getting rate to measure the effectiveness of a machine unlearning method. It is based on the concept of membership inference and describes the transformation rate of the eliminated data from "memorized" to "unknown" after conducting unlearning. We also propose a novel unlearning method calledForsaken. It is superior to previous work in either utility or efficiency (when achieving the same forgetting rate). We benchmark Forsaken with eight standard datasets to evaluate its performance. The experimental results show that it can achieve more than 90\% forgetting rate on average and only causeless than 5\% accuracy loss.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源