Generalized Single-Vehicle-Based Graph Reinforcement Learning for Decision-Making in Autonomous Driving

Journal Article (2022)
Author(s)

Fan Yang (Beijing Institute of Technology)

Xueyuan Li (Beijing Institute of Technology)

Qi Liu (Beijing Institute of Technology)

Z. Li (TU Delft - Transport and Planning, Beijing Institute of Technology)

Xin Gao (Beijing Institute of Technology)

Transport and Planning
Copyright
© 2022 Fan Yang, Xueyuan Li, Qi Liu, Z. Li, Xin Gao
DOI related publication
https://doi.org/10.3390/s22134935
More Info
expand_more
Publication Year
2022
Language
English
Copyright
© 2022 Fan Yang, Xueyuan Li, Qi Liu, Z. Li, Xin Gao
Transport and Planning
Issue number
13
Volume number
22
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

In the autonomous driving process, the decision-making system is mainly used to provide macro-control instructions based on the information captured by the sensing system. Learning-based algorithms have apparent advantages in information processing and understanding for an increasingly complex driving environment. To incorporate the interactive information between agents in the environment into the decision-making process, this paper proposes a generalized single-vehicle-based graph neural network reinforcement learning algorithm (SGRL algorithm). The SGRL algorithm introduces graph convolution into the traditional deep neural network (DQN) algorithm, adopts the training method for a single agent, designs a more explicit incentive reward function, and significantly improves the dimension of the action space. The SGRL algorithm is compared with the traditional DQN algorithm (NGRL) and the multi-agent training algorithm (MGRL) in the highway ramp scenario. Results show that the SGRL algorithm has outstanding advantages in network convergence, decision-making effect, and training efficiency.