Zhu Zhuangdi, Hong Junyuan, Zhou Jiayu
Department of Computer Science and Engineering, Michigan State University, Michigan, USA.
Proc Mach Learn Res. 2021 Jul;139:12878-12889.
Federated Learning (FL) is a decentralized machine-learning paradigm in which a global server iteratively aggregates the model parameters of local users without accessing their data. User has imposed significant challenges to FL, which can incur drifted global models that are slow to converge. has recently emerged to tackle this issue, by refining the server model using aggregated knowledge from heterogeneous users, other than directly aggregating their model parameters. This approach, however, depends on a proxy dataset, making it impractical unless such prerequisite is satisfied. Moreover, the ensemble knowledge is not fully utilized to guide local model learning, which may in turn affect the quality of the aggregated model. In this work, we propose a approach to address heterogeneous FL, where the server learns a lightweight generator to ensemble user information in a data-free manner, which is then broadcasted to users, regulating local training using the learned knowledge as an inductive bias. Empirical studies powered by theoretical implications show that, our approach facilitates FL with better generalization performance using fewer communication rounds, compared with the state-of-the-art.
联邦学习(FL)是一种去中心化的机器学习范式,其中全局服务器在不访问本地用户数据的情况下迭代聚合其模型参数。用户给联邦学习带来了重大挑战,这可能导致全局模型漂移且收敛缓慢。最近出现了一种方法来解决这个问题,即通过利用来自异构用户的聚合知识来优化服务器模型,而不是直接聚合他们的模型参数。然而,这种方法依赖于一个代理数据集,除非满足这样的前提条件,否则它是不切实际的。此外,集成知识没有被充分利用来指导本地模型学习,这反过来可能会影响聚合模型的质量。在这项工作中,我们提出了一种解决异构联邦学习的方法,其中服务器学习一个轻量级生成器,以无数据的方式整合用户信息,然后将其广播给用户,使用学到的知识作为归纳偏差来调节本地训练。由理论启示驱动的实证研究表明,与现有技术相比,我们的方法使用更少的通信轮次就能促进联邦学习具有更好的泛化性能。