Jamil Umar, Malmir Mostafa, Chen Alan, Filipovska Monika, Xie Mimi, Ding Caiwen, Jin Yu-Fang
Department of Electrical and Computer Engineering, The University of Texas at San Antonio, San Antonio, TX, USA.
Westlake High School, Austin, TX, USA.
Sci Prog. 2024 Jul-Sep;107(3):368504241263406. doi: 10.1177/00368504241263406.
Eco-driving has garnered considerable research attention owing to its potential socio-economic impact, including enhanced public health and mitigated climate change effects through the reduction of greenhouse gas emissions. With an expectation of more autonomous vehicles (AVs) on the road, an eco-driving strategy in hybrid traffic networks encompassing AV and human-driven vehicles (HDVs) with the coordination of traffic lights is a challenging task. The challenge is partially due to the insufficient infrastructure for collecting, transmitting, and sharing real-time traffic data among vehicles, facilities, and traffic control centers, and the following decision-making of agents involved in traffic control. Additionally, the intricate nature of the existing traffic network, with its diverse array of vehicles and facilities, contributes to the challenge by hindering the development of a mathematical model for accurately characterizing the traffic network. In this study, we utilized the Simulation of Urban Mobility (SUMO) simulator to tackle the first challenge through computational analysis. To address the second challenge, we employed a model-free reinforcement learning (RL) algorithm, proximal policy optimization, to decide the actions of AV and traffic light signals in a traffic network. A novel eco-driving strategy was proposed by introducing different percentages of AV into the traffic flow and collaborating with traffic light signals using RL to control the overall speed of the vehicles, resulting in improved fuel consumption efficiency. Average rewards with different penetration rates of AV (5%, 10%, and 20% of total vehicles) were compared to the situation without any AV in the traffic flow (0% penetration rate). The 10% penetration rate of AV showed a minimum time of convergence to achieve average reward, leading to a significant reduction in fuel consumption and total delay of all vehicles.
生态驾驶因其潜在的社会经济影响而受到了大量研究关注,这些影响包括通过减少温室气体排放来改善公众健康和减轻气候变化影响。随着道路上自动驾驶车辆(AV)数量的增加,在包含自动驾驶车辆和人类驾驶车辆(HDV)的混合交通网络中,结合交通信号灯协调制定生态驾驶策略是一项具有挑战性的任务。这一挑战部分源于车辆、设施和交通控制中心之间收集、传输和共享实时交通数据的基础设施不足,以及交通控制中相关主体的后续决策。此外,现有交通网络的复杂性,以及其种类繁多的车辆和设施,也给准确描述交通网络的数学模型的开发带来阻碍,进而加剧了这一挑战。在本研究中,我们利用城市交通仿真(SUMO)模拟器,通过计算分析来应对第一个挑战。为了应对第二个挑战,我们采用了一种无模型强化学习(RL)算法——近端策略优化,来决定交通网络中自动驾驶车辆和交通信号灯的行动。通过在交通流中引入不同比例的自动驾驶车辆,并利用强化学习与交通信号灯协作以控制车辆的整体速度,我们提出了一种新颖的生态驾驶策略,从而提高了燃油消耗效率。我们将不同自动驾驶车辆渗透率(占总车辆数的5%、10%和20%)下的平均奖励与交通流中没有任何自动驾驶车辆的情况(渗透率为0%)进行了比较。10%的自动驾驶车辆渗透率显示出达到平均奖励的收敛时间最短,从而显著降低了所有车辆的燃油消耗和总延误。