Mai Van Sy, La Richard J, Zhang Tao
National Institute of Standards and Technology (NIST), Gaithersburg, MD 20899, USA.
NIST, University of Maryland, College Park, MD 20742, USA.
IEEE Trans Artif Intell. 2024 Jul;Early Access. doi: 10.1109/tai.2024.3430250.
Federated Learning (FL) has emerged as a means of distributed learning using local data stored at clients with a coordinating server. Recent studies showed that FL can suffer from poor performance and slower convergence when training data at the clients are not independent and identically distributed (IID). Here, we consider auxiliary server learning as a approach to improving the performance of FL on non-IID data. Our analysis and experiments show that this approach can achieve significant improvements in both model accuracy and convergence time even when the dataset utilized by the server is small and its distribution differs from that of the clients' aggregate data. Moreover, experimental results suggest that auxiliary server learning delivers benefits when employed together with other techniques proposed to mitigate the performance degradation of FL on non-IID data.
联邦学习(FL)已成为一种分布式学习方式,它利用存储在客户端的本地数据,并通过一个协调服务器进行学习。最近的研究表明,当客户端的训练数据不是独立同分布(IID)时,联邦学习可能会出现性能不佳和收敛速度较慢的问题。在此,我们将辅助服务器学习视为一种提高联邦学习在非IID数据上性能的方法。我们的分析和实验表明,即使服务器使用的数据集较小且其分布与客户端聚合数据的分布不同,这种方法在模型准确性和收敛时间方面都能实现显著提升。此外,实验结果表明,辅助服务器学习与为减轻联邦学习在非IID数据上的性能退化而提出的其他技术一起使用时会带来益处。