• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于学习作业车间调度调度的离线强化学习。

Offline reinforcement learning for learning to dispatch for job shop scheduling.

作者信息

Remmerden Jesse van, Bukhsh Zaharah, Zhang Yingqian

机构信息

Information Systems IE&IS, Eindhoven University of Technology, De Zaale, Eindhoven, 5600 MB Netherlands.

出版信息

Mach Learn. 2025;114(8):191. doi: 10.1007/s10994-025-06826-w. Epub 2025 Jul 15.

DOI:10.1007/s10994-025-06826-w
PMID:40677903
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12263752/
Abstract

The Job Shop Scheduling Problem (JSSP) is a complex combinatorial optimization problem. While online Reinforcement Learning (RL) has shown promise by quickly finding acceptable solutions for JSSP, it faces key limitations: it requires extensive training interactions from scratch leading to sample inefficiency, cannot leverage existing high-quality solutions from traditional methods like Constraint Programming (CP), and require simulated environments to train in, which are impracticable to build for complex scheduling environments. We introduce Offline Learned Dispatching (Offline-LD), an offline reinforcement learning approach for JSSP, which addresses these limitations by learning from historical scheduling data. Our approach is motivated by scenarios where historical scheduling data and expert solutions are available or scenarios where online training of RL approaches with simulated environments is impracticable. Offline-LD introduces maskable variants of two Q-learning methods, namely, Maskable Quantile Regression DQN (mQRDQN) and discrete maskable Soft Actor-Critic (d-mSAC), that are able to learn from historical data, through Conservative Q-Learning (CQL), whereby we present a novel entropy bonus modification for d-mSAC, for maskable action spaces. Moreover, we introduce a novel reward normalization method for JSSP in an offline RL setting. Our experiments demonstrate that Offline-LD outperforms online RL on both generated and benchmark instances when trained on only 100 solutions generated by CP. Notably, introducing noise to the expert dataset yields comparable or superior results to using the expert dataset, with the same amount of instances, a promising finding for real-world applications, where data is inherently noisy and imperfect.

摘要

作业车间调度问题(JSSP)是一个复杂的组合优化问题。虽然在线强化学习(RL)通过快速为JSSP找到可接受的解决方案显示出了前景,但它面临着关键限制:它需要从头开始进行广泛的训练交互,导致样本效率低下,无法利用约束规划(CP)等传统方法现有的高质量解决方案,并且需要在模拟环境中进行训练,而对于复杂的调度环境来说构建这样的环境是不切实际的。我们引入了离线学习调度(Offline-LD),一种用于JSSP的离线强化学习方法,它通过从历史调度数据中学习来解决这些限制。我们的方法适用于有历史调度数据和专家解决方案可用的场景,或者RL方法在模拟环境中进行在线训练不切实际的场景。Offline-LD引入了两种Q学习方法的可掩码变体,即可掩码分位数回归深度Q网络(mQRDQN)和离散可掩码软演员评论家(d-mSAC),它们能够通过保守Q学习(CQL)从历史数据中学习,为此我们针对可掩码动作空间为d-mSAC提出了一种新颖的熵奖励修正。此外,我们在离线RL设置中为JSSP引入了一种新颖的奖励归一化方法。我们的实验表明,当仅在由CP生成的100个解决方案上进行训练时,Offline-LD在生成的实例和基准实例上均优于在线RL。值得注意的是,在专家数据集中引入噪声会产生与使用专家数据集相当或更好的结果,在相同数量的实例下,这对于实际应用是一个有前景的发现,因为实际应用中的数据本质上是有噪声和不完美的。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/65d925b9b9e6/10994_2025_6826_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/653171e0d9d8/10994_2025_6826_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/921e08973e30/10994_2025_6826_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/0902b49c27c6/10994_2025_6826_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/bea7c77cac47/10994_2025_6826_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/65d925b9b9e6/10994_2025_6826_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/653171e0d9d8/10994_2025_6826_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/921e08973e30/10994_2025_6826_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/0902b49c27c6/10994_2025_6826_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/bea7c77cac47/10994_2025_6826_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/aba4/12263752/65d925b9b9e6/10994_2025_6826_Fig4_HTML.jpg

相似文献

1
Offline reinforcement learning for learning to dispatch for job shop scheduling.用于学习作业车间调度调度的离线强化学习。
Mach Learn. 2025;114(8):191. doi: 10.1007/s10994-025-06826-w. Epub 2025 Jul 15.
2
Actor critic with experience replay-based automatic treatment planning for prostate cancer intensity modulated radiotherapy.基于经验回放的演员-评论家算法用于前列腺癌调强放射治疗的自动治疗计划
Med Phys. 2025 Jul;52(7):e17915. doi: 10.1002/mp.17915. Epub 2025 May 31.
3
EKG-AC: A New Paradigm for Process Industrial Optimization Based on Offline Reinforcement Learning With Expert Knowledge Guidance.心电图交流电:一种基于离线强化学习与专家知识指导的过程工业优化新范式。
IEEE Trans Cybern. 2025 Jul 17;PP. doi: 10.1109/TCYB.2025.3579361.
4
Comparison of Two Modern Survival Prediction Tools, SORG-MLA and METSSS, in Patients With Symptomatic Long-bone Metastases Who Underwent Local Treatment With Surgery Followed by Radiotherapy and With Radiotherapy Alone.两种现代生存预测工具 SORG-MLA 和 METSSS 在接受手术联合放疗和单纯放疗治疗有症状长骨转移患者中的比较。
Clin Orthop Relat Res. 2024 Dec 1;482(12):2193-2208. doi: 10.1097/CORR.0000000000003185. Epub 2024 Jul 23.
5
Privacy-Preserving Glycemic Management in Type 1 Diabetes: Development and Validation of a Multiobjective Federated Reinforcement Learning Framework.1型糖尿病中保护隐私的血糖管理:多目标联邦强化学习框架的开发与验证
JMIR Diabetes. 2025 Jul 4;10:e72874. doi: 10.2196/72874.
6
Sexual Harassment and Prevention Training性骚扰与预防培训
7
Short-Term Memory Impairment短期记忆障碍
8
Decoupled Prioritized Resampling for Offline RL.用于离线强化学习的解耦优先重采样
IEEE Trans Neural Netw Learn Syst. 2025 Jul;36(7):13094-13108. doi: 10.1109/TNNLS.2024.3488358.
9
Leveraging a foundation model zoo for cell similarity search in oncological microscopy across devices.利用基础模型库进行跨设备肿瘤显微镜检查中的细胞相似性搜索。
Front Oncol. 2025 Jun 18;15:1480384. doi: 10.3389/fonc.2025.1480384. eCollection 2025.
10
The Lived Experience of Autistic Adults in Employment: A Systematic Search and Synthesis.成年自闭症患者的就业生活经历:系统检索与综述
Autism Adulthood. 2024 Dec 2;6(4):495-509. doi: 10.1089/aut.2022.0114. eCollection 2024 Dec.

本文引用的文献

1
Flexible Job Shop Scheduling via Dual Attention Network-Based Reinforcement Learning.基于双注意力网络强化学习的柔性作业车间调度
IEEE Trans Neural Netw Learn Syst. 2024 Mar;35(3):3091-3102. doi: 10.1109/TNNLS.2023.3306421. Epub 2024 Feb 29.
2
A Survey on Offline Reinforcement Learning: Taxonomy, Review, and Open Problems.离线强化学习综述:分类、回顾与开放问题
IEEE Trans Neural Netw Learn Syst. 2024 Aug;35(8):10237-10257. doi: 10.1109/TNNLS.2023.3250269. Epub 2024 Aug 5.