论文标题

部分可观测时空混沌系统的无模型预测

DPIS: An Enhanced Mechanism for Differentially Private SGD with Importance Sampling

论文作者

Wei, Jianxin, Bao, Ergute, Xiao, Xiaokui, Yang, Yin

论文摘要

如今,差异隐私(DP)已成为隐私保护的良好标准,深层神经网络(DNN)在机器学习方面已取得了巨大成功。这两种技术(即深度学习与差异隐私)的结合承诺,保留具有敏感数据(例如医疗记录)训练的高意外模型的隐私释放。为此目的的经典机制是DP-SGD,它是常用于DNN训练的随机梯度下降(SGD)优化器的差异私有版本。随后的方法改善了模型训练过程的各个方面,包括噪声衰减时间表,模型架构,功能工程和超参数调整。但是,自从最初的DP-SGD算法以来,在SGD优化器中执行DP的核心机制一直保持不变,该算法越来越成为限制DP兼容机器学习解决方案性能的基本障碍。 在此激励的情况下,我们提出了DPI,这是一种用于差异私有SGD训练的新型机制,可以用作DP-SGD核心优化器的倒数替换,在后者中具有一致且显着的准确性提高。主要思想是在每种SGD迭代中采用重要的采样(IS)进行迷你批次选择,从而减少了采样方差和注入满足DP所需梯度的随机噪声量。整合到DP-SGD的复杂数学机制中是高度不平凡的。 DPI通过新颖的机制设计,细粒度的隐私分析,提高效率和自适应梯度剪接优化来应对挑战。在四个基准数据集(即MNIST,FMNIST,CIFAR-10和IMDB)上进行了广泛的实验,证明了DPI与现有解决方案相对于具有不同隐私的深度学习的效率。

Nowadays, differential privacy (DP) has become a well-accepted standard for privacy protection, and deep neural networks (DNN) have been immensely successful in machine learning. The combination of these two techniques, i.e., deep learning with differential privacy, promises the privacy-preserving release of high-utility models trained with sensitive data such as medical records. A classic mechanism for this purpose is DP-SGD, which is a differentially private version of the stochastic gradient descent (SGD) optimizer commonly used for DNN training. Subsequent approaches have improved various aspects of the model training process, including noise decay schedule, model architecture, feature engineering, and hyperparameter tuning. However, the core mechanism for enforcing DP in the SGD optimizer remains unchanged ever since the original DP-SGD algorithm, which has increasingly become a fundamental barrier limiting the performance of DP-compliant machine learning solutions. Motivated by this, we propose DPIS, a novel mechanism for differentially private SGD training that can be used as a drop-in replacement of the core optimizer of DP-SGD, with consistent and significant accuracy gains over the latter. The main idea is to employ importance sampling (IS) in each SGD iteration for mini-batch selection, which reduces both sampling variance and the amount of random noise injected to the gradients that is required to satisfy DP. Integrating IS into the complex mathematical machinery of DP-SGD is highly non-trivial. DPIS addresses the challenge through novel mechanism designs, fine-grained privacy analysis, efficiency enhancements, and an adaptive gradient clipping optimization. Extensive experiments on four benchmark datasets, namely MNIST, FMNIST, CIFAR-10 and IMDb, demonstrate the superior effectiveness of DPIS over existing solutions for deep learning with differential privacy.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源