• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种基于强化学习的受电弓控制策略,用于提高高速铁路的受流质量

A Reinforcement Learning-Based Pantograph Control Strategy for Improving Current Collection Quality in High-Speed Railways.

作者信息

Wang Hui, Han Zhiwei, Liu Wenqiang, Wu Yanbo

出版信息

IEEE Trans Neural Netw Learn Syst. 2024 May;35(5):5915-5928. doi: 10.1109/TNNLS.2022.3219814. Epub 2024 May 2.

DOI:10.1109/TNNLS.2022.3219814
PMID:36374889
Abstract

In high-speed railways, the pantograph-catenary system (PCS) is a critical subsystem of the train power supply system. In particular, when the double-PCS (DPCS) is in operation, the passing of the leading pantograph (LP) causes the contact force of the trailing pantograph (TP) to fluctuate violently, affecting the power collection quality of the electric multiple units (EMUs). The actively controlled pantograph is the most promising technique for reducing the pantograph-catenary contact force (PCCF) fluctuation and improving the current collection quality. Based on the Nash equilibrium framework, this study proposes a multiagent reinforcement learning (MARL) algorithm for active pantograph control called cooperative proximity policy optimization (Coo-PPO). In the algorithm implementation, the heterogeneous agents play a unique role in a cooperative environment guided by the global value function. Then, a novel reward propagation channel is proposed to reveal implicit associations between agents. Furthermore, a curriculum learning approach is adopted to strike a balance between reward maximization and rational movement patterns. An existing MARL algorithm and a traditional control strategy are compared in the same scenario to validate the proposed control strategy's performance. The experimental results show that the Coo-PPO algorithm obtains more rewards, significantly suppresses the fluctuation in PCCF (up to 41.55%), and dramatically decreases the TP's offline rate (up to 10.77%). This study adopts MARL technology for the first time to address the coordinated control of double pantographs in DPCS.

摘要

在高速铁路中,受电弓-接触网系统(PCS)是列车供电系统的关键子系统。特别是,当双受电弓-接触网系统(DPCS)运行时,前导受电弓(LP)的通过会导致后随受电弓(TP)的接触力剧烈波动,影响动车组(EMU)的集电质量。主动控制受电弓是降低受电弓-接触网接触力(PCCF)波动并提高集流质量最具前景的技术。基于纳什均衡框架,本研究提出了一种用于主动受电弓控制的多智能体强化学习(MARL)算法,称为协作近端策略优化(Coo-PPO)。在算法实现中,异构智能体在由全局价值函数引导的协作环境中发挥独特作用。然后,提出了一种新颖的奖励传播通道,以揭示智能体之间的隐含关联。此外,采用课程学习方法在奖励最大化和合理运动模式之间取得平衡。在相同场景下将现有的MARL算法和传统控制策略进行比较,以验证所提出控制策略的性能。实验结果表明,Coo-PPO算法获得了更多奖励,显著抑制了PCCF的波动(高达41.55%),并大幅降低了TP的离线率(高达10.77%)。本研究首次采用MARL技术来解决DPCS中双受电弓的协调控制问题。

相似文献

1
A Reinforcement Learning-Based Pantograph Control Strategy for Improving Current Collection Quality in High-Speed Railways.一种基于强化学习的受电弓控制策略,用于提高高速铁路的受流质量
IEEE Trans Neural Netw Learn Syst. 2024 May;35(5):5915-5928. doi: 10.1109/TNNLS.2022.3219814. Epub 2024 May 2.
2
Rapid Adaptation for Active Pantograph Control in High-Speed Railway via Deep Meta Reinforcement Learning.基于深度元强化学习的高速铁路受电弓主动控制快速自适应方法
IEEE Trans Cybern. 2024 May;54(5):2811-2823. doi: 10.1109/TCYB.2023.3271900. Epub 2024 Apr 16.
3
Dataset of measured and commented pantograph electric arcs in DC railways.直流铁路中受电弓电弧测量与注释数据集。
Data Brief. 2020 Jul 3;31:105978. doi: 10.1016/j.dib.2020.105978. eCollection 2020 Aug.
4
Hardware-in-the-Loop Test Bench for Simulation of Catenary-Pantograph Interaction (CPI) with Linear Camera Measurement.受电弓-接触网系统(CPI)硬件在环测试台及线性相机测量仿真
Sensors (Basel). 2023 Feb 4;23(4):1773. doi: 10.3390/s23041773.
5
Pantograph Detection Algorithm with Complex Background and External Disturbances.带复杂背景和外部干扰的受电弓检测算法。
Sensors (Basel). 2022 Nov 2;22(21):8425. doi: 10.3390/s22218425.
6
Pantograph Slider Detection Architecture and Solution Based on Deep Learning.基于深度学习的受电弓滑板检测架构与解决方案
Sensors (Basel). 2024 Aug 8;24(16):5133. doi: 10.3390/s24165133.
7
Accelerating Multiagent Reinforcement Learning by Equilibrium Transfer.通过均衡转移加速多智能体强化学习。
IEEE Trans Cybern. 2015 Jul;45(7):1289-302. doi: 10.1109/TCYB.2014.2349152. Epub 2014 Aug 29.
8
Learning Automata-Based Multiagent Reinforcement Learning for Optimization of Cooperative Tasks.基于学习自动机的多智能体强化学习用于协作任务优化
IEEE Trans Neural Netw Learn Syst. 2021 Oct;32(10):4639-4652. doi: 10.1109/TNNLS.2020.3025711. Epub 2021 Oct 5.
9
Data sets of measured pantograph voltage and current of European AC railways.欧洲交流铁路受电弓电压和电流的测量数据集。
Data Brief. 2020 Apr 20;30:105477. doi: 10.1016/j.dib.2020.105477. eCollection 2020 Jun.
10
Multiagent reinforcement learning with unshared value functions.多智能体强化学习与非共享价值函数。
IEEE Trans Cybern. 2015 Apr;45(4):647-62. doi: 10.1109/TCYB.2014.2332042. Epub 2014 Jul 2.