• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

FedDdrl:具有自适应早期客户端终止和本地 epoch 调整的异构物联网联邦双深度强化学习。

FedDdrl: Federated Double Deep Reinforcement Learning for Heterogeneous IoT with Adaptive Early Client Termination and Local Epoch Adjustment.

机构信息

Department of Electrical and Electronic Engineering, Lee Kong Chian Faculty of Engineering and Science, Universiti Tunku Abdul Rahman, Kajang 43000, Malaysia.

Department of Mechatronics and Biomedical Engineering, Lee Kong Chian Faculty of Engineering and Science, Universiti Tunku Abdul Rahman, Kajang 43000, Malaysia.

出版信息

Sensors (Basel). 2023 Feb 23;23(5):2494. doi: 10.3390/s23052494.

DOI:10.3390/s23052494
PMID:36904696
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10006882/
Abstract

Federated learning (FL) is a technique that allows multiple clients to collaboratively train a global model without sharing their sensitive and bandwidth-hungry data. This paper presents a joint early client termination and local epoch adjustment for FL. We consider the challenges of heterogeneous Internet of Things (IoT) environments including non-independent and identically distributed (non-IID) data as well as diverse computing and communication capabilities. The goal is to strike the best tradeoff among three conflicting objectives, namely global model accuracy, training latency and communication cost. We first leverage the balanced-MixUp technique to mitigate the influence of non-IID data on the FL convergence rate. A weighted sum optimization problem is then formulated and solved via our proposed FL double deep reinforcement learning (FedDdrl) framework, which outputs a dual action. The former indicates whether a participating FL client is dropped, whereas the latter specifies how long each remaining client needs to complete its local training task. Simulation results show that FedDdrl outperforms the existing FL scheme in terms of overall tradeoff. Specifically, FedDdrl achieves higher model accuracy by about 4% while incurring 30% less latency and communication costs.

摘要

联邦学习(FL)是一种技术,允许多个客户端在不共享其敏感和带宽密集型数据的情况下协同训练一个全局模型。本文提出了一种用于 FL 的联合早期客户端终止和本地时期调整。我们考虑了包括非独立同分布(non-IID)数据以及不同计算和通信能力的异构物联网(IoT)环境的挑战。目标是在三个相互冲突的目标之间取得最佳权衡,即全局模型准确性、训练延迟和通信成本。我们首先利用平衡 MixUp 技术来减轻非 IID 数据对 FL 收敛速度的影响。然后通过我们提出的 FL 双深度强化学习(FedDdrl)框架来制定并解决加权和优化问题,该框架输出双动作。前者表示参与 FL 的客户端是否被丢弃,后者则指定每个剩余客户端完成其本地训练任务所需的时间。仿真结果表明,FedDdrl 在总体权衡方面优于现有的 FL 方案。具体来说,FedDdrl 实现了约 4%的更高模型准确性,同时减少了 30%的延迟和通信成本。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/7cb0a8e92560/sensors-23-02494-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/e36d53cdb9ec/sensors-23-02494-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/721ace916dc9/sensors-23-02494-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/19dc12d862b3/sensors-23-02494-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/039cf9f03999/sensors-23-02494-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/de03a1e948af/sensors-23-02494-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/f34c2510a609/sensors-23-02494-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/85b51067890d/sensors-23-02494-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/1794bf02495e/sensors-23-02494-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/4f64c2c63b19/sensors-23-02494-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/991cefae8b20/sensors-23-02494-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/b7acac4222a3/sensors-23-02494-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/18fc2767e37e/sensors-23-02494-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/7cb0a8e92560/sensors-23-02494-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/e36d53cdb9ec/sensors-23-02494-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/721ace916dc9/sensors-23-02494-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/19dc12d862b3/sensors-23-02494-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/039cf9f03999/sensors-23-02494-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/de03a1e948af/sensors-23-02494-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/f34c2510a609/sensors-23-02494-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/85b51067890d/sensors-23-02494-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/1794bf02495e/sensors-23-02494-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/4f64c2c63b19/sensors-23-02494-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/991cefae8b20/sensors-23-02494-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/b7acac4222a3/sensors-23-02494-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/18fc2767e37e/sensors-23-02494-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7da9/10006882/7cb0a8e92560/sensors-23-02494-g013.jpg

相似文献

1
FedDdrl: Federated Double Deep Reinforcement Learning for Heterogeneous IoT with Adaptive Early Client Termination and Local Epoch Adjustment.FedDdrl:具有自适应早期客户端终止和本地 epoch 调整的异构物联网联邦双深度强化学习。
Sensors (Basel). 2023 Feb 23;23(5):2494. doi: 10.3390/s23052494.
2
An Optimization Method for Non-IID Federated Learning Based on Deep Reinforcement Learning.一种基于深度强化学习的非独立同分布联邦学习优化方法。
Sensors (Basel). 2023 Nov 16;23(22):9226. doi: 10.3390/s23229226.
3
Trust-Augmented Deep Reinforcement Learning for Federated Learning Client Selection.用于联邦学习客户端选择的信任增强深度强化学习
Inf Syst Front. 2022 Jul 18:1-18. doi: 10.1007/s10796-022-10307-z.
4
An EMD-Based Adaptive Client Selection Algorithm for Federated Learning in Heterogeneous Data Scenarios.一种基于经验模态分解的异构数据场景下联邦学习自适应客户端选择算法
Front Plant Sci. 2022 Jun 9;13:908814. doi: 10.3389/fpls.2022.908814. eCollection 2022.
5
Joint Client Selection and CPU Frequency Control in Wireless Federated Learning Networks with Power Constraints.具有功率约束的无线联邦学习网络中的联合客户端选择与CPU频率控制
Entropy (Basel). 2023 Aug 9;25(8):1183. doi: 10.3390/e25081183.
6
A Cluster-Driven Adaptive Training Approach for Federated Learning.一种基于簇的联邦学习自适应训练方法。
Sensors (Basel). 2022 Sep 18;22(18):7061. doi: 10.3390/s22187061.
7
Federated learning using model projection for multi-center disease diagnosis with non-IID data.使用模型投影的联邦学习用于非独立同分布数据的多中心疾病诊断
Neural Netw. 2024 Oct;178:106409. doi: 10.1016/j.neunet.2024.106409. Epub 2024 May 24.
8
Personalized federated learning for heterogeneous data: A distributed edge clustering approach.面向异构数据的个性化联邦学习:一种分布式边缘聚类方法。
Math Biosci Eng. 2023 Apr 17;20(6):10725-10740. doi: 10.3934/mbe.2023475.
9
Exploring personalization via federated representation Learning on non-IID data.探索非独立同分布数据上的联邦表示学习个性化。
Neural Netw. 2023 Jun;163:354-366. doi: 10.1016/j.neunet.2023.04.007. Epub 2023 Apr 11.
10
New Generation Federated Learning.新一代联邦学习。
Sensors (Basel). 2022 Nov 3;22(21):8475. doi: 10.3390/s22218475.

本文引用的文献

1
Facing small and biased data dilemma in drug discovery with enhanced federated learning approaches.面对药物发现中因小而偏的数据困境,采用增强型联邦学习方法。
Sci China Life Sci. 2022 Mar;65(3):529-539. doi: 10.1007/s11427-021-1946-0. Epub 2021 Jul 26.
2
Federated learning in medicine: facilitating multi-institutional collaborations without sharing patient data.医学中的联邦学习:在不共享患者数据的情况下促进多机构合作。
Sci Rep. 2020 Jul 28;10(1):12598. doi: 10.1038/s41598-020-69250-1.
3
Grandmaster level in StarCraft II using multi-agent reinforcement learning.
星际争霸 II 中的大师级水平使用多智能体强化学习。
Nature. 2019 Nov;575(7782):350-354. doi: 10.1038/s41586-019-1724-z. Epub 2019 Oct 30.
4
A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play.一种通过自我对弈掌握国际象棋、将棋和围棋的通用强化学习算法。
Science. 2018 Dec 7;362(6419):1140-1144. doi: 10.1126/science.aar6404.