Loading…

Reasoning graph-based reinforcement learning to cooperate mixed connected and autonomous traffic at unsignalized intersections

•Proposed a Reasoning Graph-based Reinforcement Learning (RGRL) framework to cooperate conflicting platoons with complex interactions.•Introduced the Graph Convolution Network and Graph Attention to capture the feature of the multi-interaction scenarios.•A situation-aware optimization to optimize th...

Full description

Saved in:
Bibliographic Details
Published in:Transportation research. Part C, Emerging technologies Emerging technologies, 2024-10, Vol.167, p.104807, Article 104807
Main Authors: Zhou, Donghao, Hang, Peng, Sun, Jian
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:•Proposed a Reasoning Graph-based Reinforcement Learning (RGRL) framework to cooperate conflicting platoons with complex interactions.•Introduced the Graph Convolution Network and Graph Attention to capture the feature of the multi-interaction scenarios.•A situation-aware optimization to optimize the systemic utility in mixed traffic environments.•Did a comparative analysis with methods based on the Multi-Agent Reinforcement Learning and the optimization.•Conducted Hardware-In-Loop and Vehicle-In-Loop experiments to validate the effectiveness of RGRL. Cooperation at unsignalized intersections in mixed traffic environments, where Connected and Autonomous Vehicles (CAVs) and Manually Driving Vehicles (MVs) coexist, holds promise for improving safety, efficiency, and energy savings. However, the mixed traffic at unsignalized intersections present huge challenges like MVs’ uncertainties, the chain reaction and diverse interactions. Following the thought of the situation-aware cooperation, this paper proposes a Reasoning Graph-based Reinforcement Learning (RGRL) method, which integrates a Graph Neural Network (GNN) based policy and an environment providing mixed traffic with uncertain behaviors. Firstly, it graphicly represents the observed scenario as a situation using the interaction graph with connected but uncertain (bi-directional) edges. The situation reasoning process is formulated as a Reasoning Graph-based Markov Decision Process which infers the vehicle sequence stage by stage so as to sequentially depict the entire situation. Then, a GNN-based policy is constructed, which uses Graph Convolution Networks (GCN) to capture the interrelated chain reactions and Graph Attentions Networks (GAT) to measure the attention of diverse interactions. Furthermore, an environment block is developed for training the policy, which provides trajectory generators for both CAVs and MVs. A reward function that considers social compliance, collision avoidance, efficiency and energy savings is also provided in this block. Finally, three Reinforcement Learning methods, D3QN, PPO and SAC, are implemented for comparative tests to explore the applicability and strength of the framework. The test results demonstrate that the D3QN outperformed the other two methods with a larger converged reward while maintaining a similar converged speed. Compared to multi-agent RL (MARL), the RGRL approach showed superior performance statistically, reduced the number of severe conflicts by 77.7
ISSN:0968-090X
DOI:10.1016/j.trc.2024.104807