• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种用于数据库资源分配的可扩展机器学习策略。

A scalable machine learning strategy for resource allocation in database.

作者信息

Manhary Fady Nashat, Mohamed Marghny H, Farouk Mamdouh

机构信息

Department of Computer Science, Faculty of Computers and Information, Assiut University, Assiut, Egypt.

Computer and Information Technology, Egypt-Japan University of Science and Technology (E-JUST), New Borg El-Arab City, 21934, Alexandria, Egypt.

出版信息

Sci Rep. 2025 Aug 20;15(1):30567. doi: 10.1038/s41598-025-14962-5.

DOI:10.1038/s41598-025-14962-5
PMID:40835668
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12368247/
Abstract

Modern cloud computing systems require intelligent resource allocation strategies that balance quality-of-service (QoS), operational costs, and energy sustainability. Existing deep Q-learning (DQN) methods suffer from sample inefficiency, centralization bottlenecks, and reactive decision-making during workload spikes. Transformer-based forecasting models such as Temporal Fusion Transformer (TFT) offer improved accuracy but introduce computational overhead, limiting real-time deployment. We propose LSTM-MARL-Ape-X, a novel framework integrating bidirectional Long Short-Term Memory (BiLSTM) for workload forecasting with Multi-Agent Reinforcement Learning (MARL) in a distributed Ape-X architecture. This approach enables proactive, decentralized, and scalable resource management through three innovations: high-accuracy forecasting using BiLSTM with feature-wise attention, variance-regularized credit assignment for stable multi-agent coordination, and faster convergence via adaptive prioritized replay. Experimental validation on real-world traces demonstrates 94.6% SLA compliance, 22% reduction in energy consumption, and linear scalability to over 5,000 nodes with sub-100 ms decision latency. The framework converges 3.2× faster than uniform sampling baselines and outperforms transformer-based models in both accuracy and inference speed. Unlike decoupled prediction-action frameworks, our method provides end-to-end optimization, enabling robust and sustainable cloud orchestration at scale.

摘要

现代云计算系统需要智能资源分配策略,以平衡服务质量(QoS)、运营成本和能源可持续性。现有的深度Q学习(DQN)方法存在样本效率低下、集中化瓶颈以及在工作负载高峰期间的反应式决策等问题。基于Transformer的预测模型,如时间融合Transformer(TFT),虽然提高了准确性,但引入了计算开销,限制了实时部署。我们提出了LSTM-MARL-Ape-X,这是一个新颖的框架,它在分布式Ape-X架构中,将用于工作负载预测的双向长短期记忆(BiLSTM)与多智能体强化学习(MARL)相结合。这种方法通过三项创新实现了主动、分散和可扩展的资源管理:使用具有特征级注意力的BiLSTM进行高精度预测、用于稳定多智能体协调的方差正则化信用分配,以及通过自适应优先重放实现更快的收敛。对实际跟踪数据的实验验证表明,该方法的服务水平协议(SLA)合规率达到94.6%,能耗降低22%,并且在决策延迟低于100毫秒的情况下,可线性扩展至超过5000个节点。该框架的收敛速度比均匀采样基线快3.2倍,在准确性和推理速度方面均优于基于Transformer的模型。与解耦的预测-行动框架不同,我们的方法提供了端到端的优化,能够在大规模上实现强大且可持续的云编排。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e00/12368247/e2dbb41c92fc/41598_2025_14962_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e00/12368247/91eba06707b8/41598_2025_14962_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e00/12368247/0695bd7666b1/41598_2025_14962_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e00/12368247/e2dbb41c92fc/41598_2025_14962_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e00/12368247/91eba06707b8/41598_2025_14962_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e00/12368247/0695bd7666b1/41598_2025_14962_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0e00/12368247/e2dbb41c92fc/41598_2025_14962_Figa_HTML.jpg

相似文献

1
A scalable machine learning strategy for resource allocation in database.一种用于数据库资源分配的可扩展机器学习策略。
Sci Rep. 2025 Aug 20;15(1):30567. doi: 10.1038/s41598-025-14962-5.
2
Representation-driven sampling and adaptive policy resetting for improving multi-Agent reinforcement learning.用于改进多智能体强化学习的表征驱动采样与自适应策略重置
Neural Netw. 2025 Jul 15;192:107875. doi: 10.1016/j.neunet.2025.107875.
3
DRL-Driven Intelligent SFC Deployment in MEC Workload for Dynamic IoT Networks.用于动态物联网网络的MEC工作负载中基于深度强化学习驱动的智能软件定义网络编排部署
Sensors (Basel). 2025 Jul 8;25(14):4257. doi: 10.3390/s25144257.
4
Multivariate multi-horizon time-series forecasting for real-time patient monitoring based on cascaded fine tuning of attention-based models.基于注意力模型的级联微调用于实时患者监测的多变量多时间跨度时间序列预测
Comput Biol Med. 2025 Aug;194:110406. doi: 10.1016/j.compbiomed.2025.110406. Epub 2025 Jun 10.
5
Optimization and benefit evaluation model of a cloud computing-based platform for power enterprises.基于云计算的电力企业平台优化与效益评估模型
Sci Rep. 2025 Jul 21;15(1):26366. doi: 10.1038/s41598-025-10314-5.
6
Privacy-Preserving Glycemic Management in Type 1 Diabetes: Development and Validation of a Multiobjective Federated Reinforcement Learning Framework.1型糖尿病中保护隐私的血糖管理:多目标联邦强化学习框架的开发与验证
JMIR Diabetes. 2025 Jul 4;10:e72874. doi: 10.2196/72874.
7
An novel cloud task scheduling framework using hierarchical deep reinforcement learning for cloud computing.一种用于云计算的基于分层深度强化学习的新型云任务调度框架。
PLoS One. 2025 Aug 21;20(8):e0329669. doi: 10.1371/journal.pone.0329669. eCollection 2025.
8
Short-Term Memory Impairment短期记忆障碍
9
Design of an intelligent AI-based multi-layer optimization framework for grid-tied solar PV-fuel cell hybrid energy systems.基于智能人工智能的并网太阳能光伏-燃料电池混合能源系统多层优化框架设计
MethodsX. 2025 Jul 30;15:103540. doi: 10.1016/j.mex.2025.103540. eCollection 2025 Dec.
10
Integrated neural network framework for multi-object detection and recognition using UAV imagery.用于使用无人机图像进行多目标检测与识别的集成神经网络框架。
Front Neurorobot. 2025 Jul 30;19:1643011. doi: 10.3389/fnbot.2025.1643011. eCollection 2025.

本文引用的文献

1
Auto-Scaling Techniques in Cloud Computing: Issues and Research Directions.云计算中的自动缩放技术:问题与研究方向。
Sensors (Basel). 2024 Aug 28;24(17):5551. doi: 10.3390/s24175551.
2
Human-level control through deep reinforcement learning.通过深度强化学习实现人类水平的控制。
Nature. 2015 Feb 26;518(7540):529-33. doi: 10.1038/nature14236.
3
Long short-term memory.长短期记忆
Neural Comput. 1997 Nov 15;9(8):1735-80. doi: 10.1162/neco.1997.9.8.1735.