论文标题
自适应联合辍学:提高联邦学习的沟通效率和概括
Adaptive Federated Dropout: Improving Communication Efficiency and Generalization for Federated Learning
论文作者
论文摘要
近年来,随着越来越多的法规解决了用户对隐私敏感的数据保护,因此获取此类数据的访问变得越来越受到限制和争议。为了利用生成并位于分布式实体(例如手机)的数据,一种革命性的分散的机器学习设置(称为联合学习),使位于不同地理位置的多个客户能够协作学习机器学习模型,同时保留所有数据的数据。但是,联邦学习的规模和分散化提出了新的挑战。客户与服务器之间的通信被认为是联合学习的融合时间中的主要瓶颈。 在本文中,我们提出和研究自适应联合辍学(AFD),这是一种降低与联邦学习相关的沟通成本的新技术。它通过允许客户在全球模型的选定子集上本地培训来优化服务器客户通信和计算成本。我们从经验上表明,这种策略与现有的压缩方法相结合,共同提供了融合时间最多可减少57倍。它还优于最先进的沟通效率解决方案。此外,它将模型概括提高了1.7%。
With more regulations tackling users' privacy-sensitive data protection in recent years, access to such data has become increasingly restricted and controversial. To exploit the wealth of data generated and located at distributed entities such as mobile phones, a revolutionary decentralized machine learning setting, known as Federated Learning, enables multiple clients located at different geographical locations to collaboratively learn a machine learning model while keeping all their data on-device. However, the scale and decentralization of federated learning present new challenges. Communication between the clients and the server is considered a main bottleneck in the convergence time of federated learning. In this paper, we propose and study Adaptive Federated Dropout (AFD), a novel technique to reduce the communication costs associated with federated learning. It optimizes both server-client communications and computation costs by allowing clients to train locally on a selected subset of the global model. We empirically show that this strategy, combined with existing compression methods, collectively provides up to 57x reduction in convergence time. It also outperforms the state-of-the-art solutions for communication efficiency. Furthermore, it improves model generalization by up to 1.7%.