SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change Decision

With the advancement of autonomous driving technology, vehicle lane-change decision (LCD) has become a critical issue for improving driving safety and efficiency. Traditional deep reinforcement learning (DRL) methods face challenges such as slow convergence, unstable decisions, and low accuracy when...

Full description

Saved in:
Bibliographic Details
Main Authors: Yang Liu, Tianxing Yang, Liwei Tian, Jianbiao Pei
Format: Article
Language:English
Published: MDPI AG 2025-01-01
Series:Mathematics
Subjects:
Online Access:https://www.mdpi.com/2227-7390/13/2/235
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832588043479941120
author Yang Liu
Tianxing Yang
Liwei Tian
Jianbiao Pei
author_facet Yang Liu
Tianxing Yang
Liwei Tian
Jianbiao Pei
author_sort Yang Liu
collection DOAJ
description With the advancement of autonomous driving technology, vehicle lane-change decision (LCD) has become a critical issue for improving driving safety and efficiency. Traditional deep reinforcement learning (DRL) methods face challenges such as slow convergence, unstable decisions, and low accuracy when dealing with complex traffic environments. To address these issues, this paper proposes a novel integrated deep reinforcement learning model called “SGD-TripleQNet” for autonomous vehicle lane-change decision-making. This method integrates three types of deep Q-learning networks (DQN, DDQN, and Dueling DDQN) and uses the Stochastic Gradient Descent (SGD) optimization algorithm to dynamically adjust the network weights. This dynamic weight adjustment process fine-tunes the weights based on gradient information to minimize the target loss function. The approach effectively addresses key challenges in autonomous driving lane-change decisions, including slow convergence, low accuracy, and unstable decision-making. The experiment shows that the proposed method, SGD-TripleQNet, has significant advantages over single models: In terms of convergence speed, it is approximately 25% faster than DQN, DDQN, and Dueling DDQN, achieving stability within 150 epochs; in terms of decision stability, the Q-value fluctuations are reduced by about 40% in the later stages of training; in terms of final performance, the average reward exceeds that of DQN (by 6.85%), DDQN (by 6.86%), and Dueling DDQN (by 6.57%), confirming the effectiveness of the proposed method. It also provides a theoretical foundation and practical guidance for the design and optimization of future autonomous driving systems.
format Article
id doaj-art-1ccb32a34e1149208b9198de234b8c5f
institution Kabale University
issn 2227-7390
language English
publishDate 2025-01-01
publisher MDPI AG
record_format Article
series Mathematics
spelling doaj-art-1ccb32a34e1149208b9198de234b8c5f2025-01-24T13:39:50ZengMDPI AGMathematics2227-73902025-01-0113223510.3390/math13020235SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change DecisionYang Liu0Tianxing Yang1Liwei Tian2Jianbiao Pei3College of Intelligent Science and Information Engineering, Shenyang University, Shenyang 110044, ChinaCollege of Intelligent Science and Information Engineering, Shenyang University, Shenyang 110044, ChinaCollege of Intelligent Science and Information Engineering, Shenyang University, Shenyang 110044, ChinaCollege of Intelligent Science and Information Engineering, Shenyang University, Shenyang 110044, ChinaWith the advancement of autonomous driving technology, vehicle lane-change decision (LCD) has become a critical issue for improving driving safety and efficiency. Traditional deep reinforcement learning (DRL) methods face challenges such as slow convergence, unstable decisions, and low accuracy when dealing with complex traffic environments. To address these issues, this paper proposes a novel integrated deep reinforcement learning model called “SGD-TripleQNet” for autonomous vehicle lane-change decision-making. This method integrates three types of deep Q-learning networks (DQN, DDQN, and Dueling DDQN) and uses the Stochastic Gradient Descent (SGD) optimization algorithm to dynamically adjust the network weights. This dynamic weight adjustment process fine-tunes the weights based on gradient information to minimize the target loss function. The approach effectively addresses key challenges in autonomous driving lane-change decisions, including slow convergence, low accuracy, and unstable decision-making. The experiment shows that the proposed method, SGD-TripleQNet, has significant advantages over single models: In terms of convergence speed, it is approximately 25% faster than DQN, DDQN, and Dueling DDQN, achieving stability within 150 epochs; in terms of decision stability, the Q-value fluctuations are reduced by about 40% in the later stages of training; in terms of final performance, the average reward exceeds that of DQN (by 6.85%), DDQN (by 6.86%), and Dueling DDQN (by 6.57%), confirming the effectiveness of the proposed method. It also provides a theoretical foundation and practical guidance for the design and optimization of future autonomous driving systems.https://www.mdpi.com/2227-7390/13/2/235autonomous drivingdeep reinforcement learninglane-change decisionstochastic gradient descent (SGD) optimization algorithm
spellingShingle Yang Liu
Tianxing Yang
Liwei Tian
Jianbiao Pei
SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change Decision
Mathematics
autonomous driving
deep reinforcement learning
lane-change decision
stochastic gradient descent (SGD) optimization algorithm
title SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change Decision
title_full SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change Decision
title_fullStr SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change Decision
title_full_unstemmed SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change Decision
title_short SGD-TripleQNet: An Integrated Deep Reinforcement Learning Model for Vehicle Lane-Change Decision
title_sort sgd tripleqnet an integrated deep reinforcement learning model for vehicle lane change decision
topic autonomous driving
deep reinforcement learning
lane-change decision
stochastic gradient descent (SGD) optimization algorithm
url https://www.mdpi.com/2227-7390/13/2/235
work_keys_str_mv AT yangliu sgdtripleqnetanintegrateddeepreinforcementlearningmodelforvehiclelanechangedecision
AT tianxingyang sgdtripleqnetanintegrateddeepreinforcementlearningmodelforvehiclelanechangedecision
AT liweitian sgdtripleqnetanintegrateddeepreinforcementlearningmodelforvehiclelanechangedecision
AT jianbiaopei sgdtripleqnetanintegrateddeepreinforcementlearningmodelforvehiclelanechangedecision