论文标题

通过智能数据交换嵌入无监督联盟学习的对齐

Embedding Alignment for Unsupervised Federated Learning via Smart Data Exchange

论文作者

Wagle, Satyavrat, Hosseinalipour, Seyyedali, Khosravan, Naji, Chiang, Mung, Brinton, Christopher G.

论文摘要

联邦学习(FL)被认为是分布式机器学习(ML)最有希望的解决方案之一。在当前的大多数文献中,FL已被研究用于监督的ML任务,其中边缘设备收集标记的数据。然而,在许多应用中,假设存在跨设备标记的数据是不切实际的。为此,我们开发了一种新颖的方法论,合作联合无监督的对比学习(CF-CL),用于使用未标记的数据集的跨边缘设备的FL。 CF-CL采用本地设备合作,其中通过设备到设备(D2D)通信在设备之间进行数据交换,以避免由非独立且相同分布的(非I.I.I.D。)本地数据集引起的本地模型偏差。 CF-CL引入了针对无监督的FL设置量身定制的推扣智能数据共享机制,在该设置中,每种设备将其本地数据点的子集推向其邻居,作为保留数据点,并通过概率的重要性采样技术从其邻居中提取一组数据点。我们证明,CF-CL导致(i)跨设备的无监督的潜在空间对齐,(ii)更快的全局收敛速度,允许较低的全局模型聚合; (iii)在极端非i.i.d中有效。跨设备的数据设置。

Federated learning (FL) has been recognized as one of the most promising solutions for distributed machine learning (ML). In most of the current literature, FL has been studied for supervised ML tasks, in which edge devices collect labeled data. Nevertheless, in many applications, it is impractical to assume existence of labeled data across devices. To this end, we develop a novel methodology, Cooperative Federated unsupervised Contrastive Learning (CF-CL), for FL across edge devices with unlabeled datasets. CF-CL employs local device cooperation where data are exchanged among devices through device-to-device (D2D) communications to avoid local model bias resulting from non-independent and identically distributed (non-i.i.d.) local datasets. CF-CL introduces a push-pull smart data sharing mechanism tailored to unsupervised FL settings, in which, each device pushes a subset of its local datapoints to its neighbors as reserved data points, and pulls a set of datapoints from its neighbors, sampled through a probabilistic importance sampling technique. We demonstrate that CF-CL leads to (i) alignment of unsupervised learned latent spaces across devices, (ii) faster global convergence, allowing for less frequent global model aggregations; and (iii) is effective in extreme non-i.i.d. data settings across the devices.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源