论文标题
一阶方法的素描:低带宽通道和漏洞的有效算法
Sketching for First Order Method: Efficient Algorithm for Low-Bandwidth Channel and Vulnerability
论文作者
论文摘要
素描是大规模机器学习中最基本的工具之一。它通过将原始大问题随机压缩到较低维度来实现运行时和内存节省。在本文中,我们在大规模分布式学习设置中为第一阶方法提出了一种新颖的草图方案,以便在仍然保证算法的收敛时节省分布式代理之间的通信成本。 Given gradient information in a high dimension $d$, the agent passes the compressed information processed by a sketching matrix $R\in \mathbb{R}^{s\times d}$ with $s\ll d$, and the receiver de-compressed via the de-sketching matrix $R^\top$ to ``recover'' the information in original dimension.使用这样的框架,我们开发了以较低的沟通成本来开发用于联合学习的算法。但是,这种随机素描不能直接保护本地数据的隐私。我们表明,通过提出特定的梯度攻击方法应用草图技术后仍然存在梯度泄漏问题。作为一种补救措施,我们严格证明该算法将通过在渐变信息中添加其他随机噪声来差异化,从而导致沟通效率和差异化的私人一阶方法,用于联合学习任务。我们的草图方案可以进一步推广到其他学习环境,并且可能具有独立的兴趣本身。
Sketching is one of the most fundamental tools in large-scale machine learning. It enables runtime and memory saving via randomly compressing the original large problem into lower dimensions. In this paper, we propose a novel sketching scheme for the first order method in large-scale distributed learning setting, such that the communication costs between distributed agents are saved while the convergence of the algorithms is still guaranteed. Given gradient information in a high dimension $d$, the agent passes the compressed information processed by a sketching matrix $R\in \mathbb{R}^{s\times d}$ with $s\ll d$, and the receiver de-compressed via the de-sketching matrix $R^\top$ to ``recover'' the information in original dimension. Using such a framework, we develop algorithms for federated learning with lower communication costs. However, such random sketching does not protect the privacy of local data directly. We show that the gradient leakage problem still exists after applying the sketching technique by presenting a specific gradient attack method. As a remedy, we prove rigorously that the algorithm will be differentially private by adding additional random noises in gradient information, which results in a both communication-efficient and differentially private first order approach for federated learning tasks. Our sketching scheme can be further generalized to other learning settings and might be of independent interest itself.