• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

FedMed:一种用于语言模型的联邦学习框架。

FedMed: A Federated Learning Framework for Language Modeling.

机构信息

School of Computer Engineering and Science, Shanghai University, Shanghai 200444, China.

Shanghai Institute for Advanced Communication and Data Science, Shanghai University, Shanghai 200444, China.

出版信息

Sensors (Basel). 2020 Jul 21;20(14):4048. doi: 10.3390/s20144048.

DOI:10.3390/s20144048
PMID:32708152
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7412048/
Abstract

Federated learning (FL) is a privacy-preserving technique for training a vast amount of decentralized data and making inferences on mobile devices. As a typical language modeling problem, mobile keyboard prediction aims at suggesting a probable next word or phrase and facilitating the human-machine interaction in a virtual keyboard of the smartphone or laptop. Mobile keyboard prediction with FL hopes to satisfy the growing demand that high-level data privacy be preserved in artificial intelligence applications even with the distributed models training. However, there are two major problems in the federated optimization for the prediction: (1) aggregating model parameters on the server-side and (2) reducing communication costs caused by model weights collection. To address the above issues, traditional FL methods simply use averaging aggregation or ignore communication costs. We propose a novel Federated Mediation (FedMed) framework with the adaptive aggregation, mediation incentive scheme, and topK strategy to address the model aggregation and communication costs. The performance is evaluated in terms of perplexity and communication rounds. Experiments are conducted on three datasets (i.e., Penn Treebank, WikiText-2, and Yelp) and the results demonstrate that our FedMed framework achieves robust performance and outperforms baseline approaches.

摘要

联邦学习(FL)是一种保护隐私的技术,用于训练大量分散的数据,并在移动设备上进行推断。作为一个典型的语言建模问题,移动键盘预测旨在建议一个可能的下一个单词或短语,并在智能手机或笔记本电脑的虚拟键盘中促进人机交互。带有 FL 的移动键盘预测希望满足人工智能应用程序中保护高级别数据隐私的不断增长的需求,即使使用分布式模型进行训练也是如此。然而,在预测的联邦优化中存在两个主要问题:(1)在服务器端聚合模型参数,(2)减少模型权重收集引起的通信成本。为了解决上述问题,传统的 FL 方法简单地使用平均聚合或忽略通信成本。我们提出了一种新颖的联邦调解(FedMed)框架,具有自适应聚合、调解激励方案和 topK 策略,以解决模型聚合和通信成本问题。性能根据困惑度和通信轮数进行评估。在三个数据集(即 Penn Treebank、WikiText-2 和 Yelp)上进行了实验,结果表明我们的 FedMed 框架实现了稳健的性能,优于基线方法。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/40d261ff27e9/sensors-20-04048-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/3999d2a32648/sensors-20-04048-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/d5d98d8f7c6c/sensors-20-04048-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/9d9427f2cd93/sensors-20-04048-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/b507d07b956a/sensors-20-04048-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/708ab5dc9161/sensors-20-04048-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/40d261ff27e9/sensors-20-04048-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/3999d2a32648/sensors-20-04048-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/d5d98d8f7c6c/sensors-20-04048-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/9d9427f2cd93/sensors-20-04048-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/b507d07b956a/sensors-20-04048-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/708ab5dc9161/sensors-20-04048-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d9b1/7412048/40d261ff27e9/sensors-20-04048-g006.jpg

相似文献

1
FedMed: A Federated Learning Framework for Language Modeling.FedMed:一种用于语言模型的联邦学习框架。
Sensors (Basel). 2020 Jul 21;20(14):4048. doi: 10.3390/s20144048.
2
Ternary Compression for Communication-Efficient Federated Learning.用于通信高效联邦学习的三元压缩
IEEE Trans Neural Netw Learn Syst. 2022 Mar;33(3):1162-1176. doi: 10.1109/TNNLS.2020.3041185. Epub 2022 Feb 28.
3
Securing federated learning with blockchain: a systematic literature review.利用区块链保障联邦学习安全:一项系统文献综述
Artif Intell Rev. 2023;56(5):3951-3985. doi: 10.1007/s10462-022-10271-9. Epub 2022 Sep 16.
4
Boosted federated learning based on improved Particle Swarm Optimization for healthcare IoT devices.基于改进粒子群优化算法的联邦学习在医疗保健物联网设备中的应用。
Comput Biol Med. 2023 Sep;163:107195. doi: 10.1016/j.compbiomed.2023.107195. Epub 2023 Jun 22.
5
Evaluating Federated Learning Simulators: A Comparative Analysis of Horizontal and Vertical Approaches.评估联邦学习模拟器:水平方法与垂直方法的比较分析
Sensors (Basel). 2024 Aug 9;24(16):5149. doi: 10.3390/s24165149.
6
Federated Learning With Privacy-Preserving Ensemble Attention Distillation.联邦学习中的隐私保护集成注意力蒸馏
IEEE Trans Med Imaging. 2023 Jul;42(7):2057-2067. doi: 10.1109/TMI.2022.3213244. Epub 2023 Jun 30.
7
Robust and Communication-Efficient Federated Learning From Non-i.i.d. Data.来自非独立同分布数据的稳健且通信高效的联邦学习
IEEE Trans Neural Netw Learn Syst. 2020 Sep;31(9):3400-3413. doi: 10.1109/TNNLS.2019.2944481. Epub 2019 Nov 1.
8
The FeatureCloud Platform for Federated Learning in Biomedicine: Unified Approach.FeatureCloud 平台在生物医学领域的联邦学习:统一方法。
J Med Internet Res. 2023 Jul 12;25:e42621. doi: 10.2196/42621.
9
DACFL: Dynamic Average Consensus-Based Federated Learning in Decentralized Sensors Network.DACFL:分散式传感器网络中的基于动态平均共识的联邦学习。
Sensors (Basel). 2022 Apr 26;22(9):3317. doi: 10.3390/s22093317.
10
Robust and Privacy-Preserving Decentralized Deep Federated Learning Training: Focusing on Digital Healthcare Applications.稳健且保护隐私的去中心化深度联邦学习训练:专注于数字医疗保健应用。
IEEE/ACM Trans Comput Biol Bioinform. 2024 Jul-Aug;21(4):890-901. doi: 10.1109/TCBB.2023.3243932. Epub 2024 Aug 8.

引用本文的文献

1
Heuristic based federated learning with adaptive hyperparameter tuning for households energy prediction.基于启发式的联邦学习与自适应超参数调整用于家庭能源预测。
Sci Rep. 2025 Apr 12;15(1):12564. doi: 10.1038/s41598-025-96443-3.
2
Open challenges and opportunities in federated foundation models towards biomedical healthcare.联合基础模型在生物医学医疗保健领域面临的公开挑战与机遇。
BioData Min. 2025 Jan 4;18(1):2. doi: 10.1186/s13040-024-00414-9.
3
Decades in the Making: The Evolution of Digital Health Research Infrastructure Through Synthetic Data, Common Data Models, and Federated Learning.

本文引用的文献

1
Two-stage Federated Phenotyping and Patient Representation Learning.两阶段联合表型分析与患者表征学习
Proc Conf Assoc Comput Linguist Meet. 2019 Aug;2019:283-291. doi: 10.18653/v1/W19-5030.
2
Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning.平均法可能不是联邦学习中聚合参数的最佳方式。
Entropy (Basel). 2020 Mar 11;22(3):314. doi: 10.3390/e22030314.
3
Federated Reinforcement Learning for Training Control Policies on Multiple IoT Devices.用于在多个物联网设备上训练控制策略的联邦强化学习
数十年磨一剑:通过合成数据、通用数据模型和联邦学习实现数字健康研究基础设施的演进
J Med Internet Res. 2024 Dec 20;26:e58637. doi: 10.2196/58637.
4
Deep Federated Adaptation: An Adaptative Residential Load Forecasting Approach with Federated Learning.深度联邦自适应:一种具有联邦学习的自适应住宅负荷预测方法。
Sensors (Basel). 2022 Apr 24;22(9):3264. doi: 10.3390/s22093264.
Sensors (Basel). 2020 Mar 2;20(5):1359. doi: 10.3390/s20051359.
4
Robust and Communication-Efficient Federated Learning From Non-i.i.d. Data.来自非独立同分布数据的稳健且通信高效的联邦学习
IEEE Trans Neural Netw Learn Syst. 2020 Sep;31(9):3400-3413. doi: 10.1109/TNNLS.2019.2944481. Epub 2019 Nov 1.
5
Long short-term memory.长短期记忆
Neural Comput. 1997 Nov 15;9(8):1735-80. doi: 10.1162/neco.1997.9.8.1735.