Department of Electrical and Electronic Engineering, Lee Kong Chian Faculty of Engineering and Science, Universiti Tunku Abdul Rahman, Kajang 43000, Malaysia.
Department of Mechatronics and Biomedical Engineering, Lee Kong Chian Faculty of Engineering and Science, Universiti Tunku Abdul Rahman, Kajang 43000, Malaysia.
Sensors (Basel). 2023 Feb 23;23(5):2494. doi: 10.3390/s23052494.
Federated learning (FL) is a technique that allows multiple clients to collaboratively train a global model without sharing their sensitive and bandwidth-hungry data. This paper presents a joint early client termination and local epoch adjustment for FL. We consider the challenges of heterogeneous Internet of Things (IoT) environments including non-independent and identically distributed (non-IID) data as well as diverse computing and communication capabilities. The goal is to strike the best tradeoff among three conflicting objectives, namely global model accuracy, training latency and communication cost. We first leverage the balanced-MixUp technique to mitigate the influence of non-IID data on the FL convergence rate. A weighted sum optimization problem is then formulated and solved via our proposed FL double deep reinforcement learning (FedDdrl) framework, which outputs a dual action. The former indicates whether a participating FL client is dropped, whereas the latter specifies how long each remaining client needs to complete its local training task. Simulation results show that FedDdrl outperforms the existing FL scheme in terms of overall tradeoff. Specifically, FedDdrl achieves higher model accuracy by about 4% while incurring 30% less latency and communication costs.
联邦学习(FL)是一种技术,允许多个客户端在不共享其敏感和带宽密集型数据的情况下协同训练一个全局模型。本文提出了一种用于 FL 的联合早期客户端终止和本地时期调整。我们考虑了包括非独立同分布(non-IID)数据以及不同计算和通信能力的异构物联网(IoT)环境的挑战。目标是在三个相互冲突的目标之间取得最佳权衡,即全局模型准确性、训练延迟和通信成本。我们首先利用平衡 MixUp 技术来减轻非 IID 数据对 FL 收敛速度的影响。然后通过我们提出的 FL 双深度强化学习(FedDdrl)框架来制定并解决加权和优化问题,该框架输出双动作。前者表示参与 FL 的客户端是否被丢弃,后者则指定每个剩余客户端完成其本地训练任务所需的时间。仿真结果表明,FedDdrl 在总体权衡方面优于现有的 FL 方案。具体来说,FedDdrl 实现了约 4%的更高模型准确性,同时减少了 30%的延迟和通信成本。