Guo Shunxin, Wang Hongsong, Lin Shuxia, Kou Zhiqiang, Geng Xin
IEEE Trans Neural Netw Learn Syst. 2025 May;36(5):8442-8454. doi: 10.1109/TNNLS.2024.3438281. Epub 2025 May 2.
Federated learning (FL) is an efficient framework designed to facilitate collaborative model training across multiple distributed devices while preserving user data privacy. A significant challenge of FL is data-level heterogeneity, i.e., skewed or long-tailed distribution of private data. Although various methods have been proposed to address this challenge, most of them assume that the underlying global data are uniformly distributed across all clients. This article investigates data-level heterogeneity FL with a brief review and redefines a more practical and challenging setting called skewed heterogeneous FL (SHFL). Accordingly, we propose a novel federated prototype rectification with personalization (FedPRP) which consists of two parts: federated personalization and federated prototype rectification. The former aims to construct balanced decision boundaries between dominant and minority classes based on private data, while the latter exploits both interclass discrimination and intraclass consistency to rectify empirical prototypes. Experiments on three popular benchmarks show that the proposed approach outperforms current state-of-the-art methods and achieves balanced performance in both personalization and generalization.
联邦学习(FL)是一种高效的框架,旨在促进跨多个分布式设备的协作模型训练,同时保护用户数据隐私。联邦学习的一个重大挑战是数据级异质性,即私有数据的倾斜或长尾分布。尽管已经提出了各种方法来应对这一挑战,但大多数方法都假设基础全局数据在所有客户端上均匀分布。本文通过简要回顾研究了数据级异质性联邦学习,并重新定义了一种更实际、更具挑战性的设置,称为倾斜异构联邦学习(SHFL)。相应地,我们提出了一种新颖的带个性化的联邦原型校正(FedPRP),它由两部分组成:联邦个性化和联邦原型校正。前者旨在基于私有数据在主导类和少数类之间构建平衡的决策边界,而后者利用类间区分和类内一致性来校正经验原型。在三个流行基准上的实验表明,所提出的方法优于当前的先进方法,并在个性化和泛化方面都实现了平衡的性能。