• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于深度Q网络(DQN)算法的大型货运列车编组站调车作业计划优化

Optimization of shunting operation plan in large freight train depot based on DQN algorithm.

作者信息

Qiu Jiandong, Xu Shusheng, Tang Minan, Liu Jiaxuan, Song Hailong

机构信息

School of Mechanical Engineering, Lanzhou Jiaotong University, Lanzhou, China.

School of Automation and Electrical Engineering, Lanzhou Jiaotong University, Lanzhou, China.

出版信息

PLoS One. 2025 Apr 8;20(4):e0320762. doi: 10.1371/journal.pone.0320762. eCollection 2025.

DOI:10.1371/journal.pone.0320762
PMID:40198688
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11978120/
Abstract

Shunting operation plan is the main daily work of the freight train depot, the optimization of shunting operation plan is of great significance to improve the efficiency of railway operation and production and transportation. In this paper, the deep reinforcement learning (DRL) environment and model of shunting operation problem are constructed by three elements: action, state and reward, taking shunting locomotive as the agent, the lane number of the fall-down train group as the action, the fall-down conditions of the train group as the state, and design the reward function based on the total number of shunting hooks generated after the group's descent and reorganization. The model is solved using the Deep Q network (DQN) algorithm with the objective of minimizing the number of shunting hooks, the optimal shunting operation plan can be solved after sufficient training. DQN is verified to be effective through example simulations: Compared to the overall planning and coordinating (OPC) method, DQN produces a shunting operation plan that occupies fewer lanes and produces 10% fewer total shunting hooks. Compared to the binary search tree (BST) algorithm, DQN produces 5% fewer total shunting hooks. Compared with the branch and bound (B&B) algorithm, DQN takes less time to solve, and the number of freight train removed by the coupling and slipping operations is reduced by 5.3% and 2.9%, respectively, and the quality of the shunting operation plan is better. Therefore, this paper provides a new solution for the intelligentization of shunting operations in large freight train depot.

摘要

调车作业计划是货运列车段的主要日常工作,调车作业计划的优化对于提高铁路运营生产和运输效率具有重要意义。本文通过动作、状态和奖励三个要素构建调车作业问题的深度强化学习(DRL)环境和模型,以调车机车为智能体,以解体列车组的股道号为动作,以列车组的解体条件为状态,并基于列车组解体和重新编组后产生的调车钩总数设计奖励函数。采用深度Q网络(DQN)算法求解该模型,以调车钩数量最少为目标,经过充分训练后可求解出最优调车作业计划。通过实例仿真验证了DQN的有效性:与统筹规划(OPC)方法相比,DQN生成的调车作业计划占用股道更少,总调车钩数减少10%。与二叉搜索树(BST)算法相比,DQN的总调车钩数减少5%。与分支定界(B&B)算法相比,DQN求解时间更短,连挂和溜放作业解编的货运列车数量分别减少5.3%和2.9%,调车作业计划质量更好。因此,本文为大型货运列车段调车作业智能化提供了一种新的解决方案。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/df00578913d2/pone.0320762.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/a4b31aac4fec/pone.0320762.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/c0aa7c7ea6c0/pone.0320762.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/f4fa295d08d4/pone.0320762.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/eed9f72da2c5/pone.0320762.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/fdaf28d488c1/pone.0320762.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/b29012591fd4/pone.0320762.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/f12fa64485c9/pone.0320762.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/df00578913d2/pone.0320762.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/a4b31aac4fec/pone.0320762.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/c0aa7c7ea6c0/pone.0320762.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/f4fa295d08d4/pone.0320762.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/eed9f72da2c5/pone.0320762.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/fdaf28d488c1/pone.0320762.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/b29012591fd4/pone.0320762.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/f12fa64485c9/pone.0320762.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/09a7/11978120/df00578913d2/pone.0320762.g008.jpg

相似文献

1
Optimization of shunting operation plan in large freight train depot based on DQN algorithm.基于深度Q网络(DQN)算法的大型货运列车编组站调车作业计划优化
PLoS One. 2025 Apr 8;20(4):e0320762. doi: 10.1371/journal.pone.0320762. eCollection 2025.
2
[Occupational stress situation analysis of different types of train drivers].[不同类型火车司机的职业压力状况分析]
Zhonghua Yu Fang Yi Xue Za Zhi. 2014 Nov;48(11):1001-6.
3
Optimizing operation of delivering and fetching wagons at a railway station with mixed-shaped goods operation sites.优化具有混合货物作业站点的火车站的货车取送作业。
PLoS One. 2022 Jan 31;17(1):e0263029. doi: 10.1371/journal.pone.0263029. eCollection 2022.
4
Deep reinforcement learning for automated radiation adaptation in lung cancer.深度强化学习在肺癌放射自适应中的应用。
Med Phys. 2017 Dec;44(12):6690-6705. doi: 10.1002/mp.12625. Epub 2017 Nov 14.
5
A two-stage approach to the depot shunting driver assignment problem with workload balance considerations.一种考虑工作量平衡的车厂调车司机指派问题的两阶段方法。
PLoS One. 2017 Jul 13;12(7):e0181165. doi: 10.1371/journal.pone.0181165. eCollection 2017.
6
Robust optimization model and algorithm for railway freight center location problem in uncertain environment.不确定环境下铁路货运中心选址问题的鲁棒优化模型与算法
Comput Intell Neurosci. 2014;2014:607159. doi: 10.1155/2014/607159. Epub 2014 Nov 4.
7
Optimizing the Shunting Schedule of Electric Multiple Units Depot Using an Enhanced Particle Swarm Optimization Algorithm.基于改进粒子群优化算法的电动车组运用所调车作业计划优化
Comput Intell Neurosci. 2016;2016:5804626. doi: 10.1155/2016/5804626. Epub 2016 Jun 29.
8
Teleconsultation dynamic scheduling with a deep reinforcement learning approach.基于深度强化学习的远程会诊动态调度。
Artif Intell Med. 2024 Mar;149:102806. doi: 10.1016/j.artmed.2024.102806. Epub 2024 Feb 9.
9
Approximate Policy-Based Accelerated Deep Reinforcement Learning.基于近似策略的加速深度强化学习
IEEE Trans Neural Netw Learn Syst. 2020 Jun;31(6):1820-1830. doi: 10.1109/TNNLS.2019.2927227. Epub 2019 Aug 6.
10
[Effects of occupational stress and related factors to the mood of speed train drivers].[职业压力及相关因素对高速列车司机情绪的影响]
Zhonghua Yu Fang Yi Xue Za Zhi. 2014 Apr;48(4):281-5.

本文引用的文献

1
Deep Q-network-based traffic signal control models.基于深度 Q 网络的交通信号控制模型。
PLoS One. 2021 Sep 2;16(9):e0256405. doi: 10.1371/journal.pone.0256405. eCollection 2021.
2
A study of the car-to-train assignment problem for rail express cargos in the scheduled and unscheduled train services network.研究计划和非计划铁路快运服务网络中汽车到火车的分配问题。
PLoS One. 2018 Oct 10;13(10):e0204598. doi: 10.1371/journal.pone.0204598. eCollection 2018.
3
Machine learning: Trends, perspectives, and prospects.机器学习:趋势、观点和展望。
Science. 2015 Jul 17;349(6245):255-60. doi: 10.1126/science.aaa8415.
4
Human-level control through deep reinforcement learning.通过深度强化学习实现人类水平的控制。
Nature. 2015 Feb 26;518(7540):529-33. doi: 10.1038/nature14236.