论文标题

部分可观测时空混沌系统的无模型预测

Exploiting Personalized Invariance for Better Out-of-distribution Generalization in Federated Learning

论文作者

Tang, Xueyang, Guo, Song, Zhang, Jie

论文摘要

最近,培训数据集之间的数据异质性(又称非IID数据)引起了人们对联邦学习(FL)的强烈兴趣,并且已经提出了许多个性化的联合学习方法来处理它。但是,尽管在现实世界中,培训数据集和测试数据集之间的分布变化从未在FL中考虑到。我们注意到,由于个性化和虚假信息之间的纠缠,在非IID联合设置下的分配转移(又称分布概括)问题变得更具挑战性。为了解决上述问题,我们详细阐述了一个普通的双重规范学习框架,以探索个性化的不变性,与由单个基线(通常是全球模型)正规化的个性化联合学习方法相比。利用个性化不变功能,开发的个性化模型可以有效利用最相关的信息,同时消除虚假信息,从而提高每个客户的分布泛滥性能。关于收敛性和OOD概括性能的理论分析以及广泛的实验结果都证明了我们方法比现有的联合学习和不变学习方法的优越性,在不同的分布和非IID数据案例中。

Recently, data heterogeneity among the training datasets on the local clients (a.k.a., Non-IID data) has attracted intense interest in Federated Learning (FL), and many personalized federated learning methods have been proposed to handle it. However, the distribution shift between the training dataset and testing dataset on each client is never considered in FL, despite it being general in real-world scenarios. We notice that the distribution shift (a.k.a., out-of-distribution generalization) problem under Non-IID federated setting becomes rather challenging due to the entanglement between personalized and spurious information. To tackle the above problem, we elaborate a general dual-regularized learning framework to explore the personalized invariance, compared with the exsiting personalized federated learning methods which are regularized by a single baseline (usually the global model). Utilizing the personalized invariant features, the developed personalized models can efficiently exploit the most relevant information and meanwhile eliminate spurious information so as to enhance the out-of-distribution generalization performance for each client. Both the theoretical analysis on convergence and OOD generalization performance and the results of extensive experiments demonstrate the superiority of our method over the existing federated learning and invariant learning methods, in diverse out-of-distribution and Non-IID data cases.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源