Gao Huimin, Wu Qingtao, Zhao Xuhui, Zhu Junlong, Zhang Mingchuan
School of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China.
Intelligent System Science and Technology Innovation Center, Longmen Laboratory, Luoyang 471023, China.
Sensors (Basel). 2023 Jun 29;23(13):6034. doi: 10.3390/s23136034.
Federated learning is served as a novel distributed training framework that enables multiple clients of the internet of things to collaboratively train a global model while the data remains local. However, the implement of federated learning faces many problems in practice, such as the large number of training for convergence due to the size of model and the lack of adaptivity by the stochastic gradient-based update at the client side. Meanwhile, it is sensitive to noise during the optimization process that can affect the performance of the final model. For these reasons, we propose Federated Adaptive learning based on Derivative Term, called FedADT in this paper, which incorporates adaptive step size and difference of gradient in the update of local model. To further reduce the influence of noise on the derivative term that is estimated by difference of gradient, we use moving average decay on the derivative term. Moreover, we analyze the convergence performance of the proposed algorithm for non-convex objective function, i.e., the convergence rate of 1/nT can be achieved by choosing appropriate hyper-parameters, where is the number of clients and is the number of iterations, respectively. Finally, various experiments for the image classification task are conducted by training widely used convolutional neural network on MNIST and Fashion MNIST datasets to verify the effectiveness of FedADT. In addition, the receiver operating characteristic curve is used to display the result of the proposed algorithm by predicting the categories of clothing on the Fashion MNIST dataset.
联邦学习是一种新型的分布式训练框架,它能使多个物联网客户端在数据保持本地化的同时协作训练一个全局模型。然而,联邦学习的实现实际上面临许多问题,例如由于模型规模导致收敛所需的大量训练,以及客户端基于随机梯度的更新缺乏适应性。同时,它在优化过程中对噪声敏感,这会影响最终模型的性能。基于这些原因,我们提出了基于导数项的联邦自适应学习,在本文中称为FedADT,它在局部模型更新中纳入了自适应步长和梯度差。为了进一步减少噪声对由梯度差估计的导数项的影响,我们对导数项使用移动平均衰减。此外,我们分析了所提出算法对于非凸目标函数的收敛性能,即通过选择合适的超参数可以实现1/nT的收敛速率,其中n是客户端数量,T是迭代次数。最后,通过在MNIST和Fashion MNIST数据集上训练广泛使用的卷积神经网络来进行图像分类任务的各种实验,以验证FedADT的有效性。此外,通过在Fashion MNIST数据集上预测服装类别,使用接收者操作特征曲线来展示所提出算法的结果。