Next Article in Journal
Analytical Calculation and Verification of Radial Electromagnetic Force Under Multi-Type Air Gap Eccentricity of Hub Motor
Previous Article in Journal
The Soft Fixed Route Hybrid Electric Aircraft Charging Problem with Variable Speed
Previous Article in Special Issue
A Proactive Collision Avoidance Model for Connected and Autonomous Vehicles in Mixed Traffic Flow
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Optimizing Autonomous Vehicle Navigation through Reinforcement Learning in Dynamic Urban Environments

by
Mohammed Abdullah Alsuwaiket
Department of Computer Science and Engineering Technology, University Of Hafr Al Batin, Hafr Al Batin 39524, Saudi Arabia
World Electr. Veh. J. 2025, 16(8), 472; https://doi.org/10.3390/wevj16080472
Submission received: 10 July 2025 / Revised: 10 August 2025 / Accepted: 11 August 2025 / Published: 18 August 2025
(This article belongs to the Special Issue Modeling for Intelligent Vehicles)

Abstract

Autonomous vehicle (AV) navigation in dynamic urban environments faces challenges such as unpredictable traffic conditions, varying road user behaviors, and complex road networks. This study proposes a novel reinforcement learning-based framework that enhances AV decision making through spatial-temporal context awareness. The framework integrates Proximal Policy Optimization (PPO) and Graph Neural Networks (GNNs) to effectively model urban features like intersections, traffic density, and pedestrian zones. A key innovation is the urban context-aware reward mechanism (UCARM), which dynamically adapts the reward structure based on traffic rules, congestion levels, and safety considerations. Additionally, the framework incorporates a Dynamic Risk Assessment Module (DRAM), which uses Bayesian inference combined with Markov Decision Processes (MDPs) to proactively evaluate collision risks and guide safer navigation. The framework’s performance was validated across three datasets—Argoverse, nuScenes, and CARLA. Results demonstrate significant improvements: an average travel time of 420 ± 20 s, a collision rate of 3.1%, and energy consumption of 11,833 ± 550 J in Argoverse; 410 ± 20 s, 2.5%, and 11,933 ± 450 J in nuScenes; and 450 ± 25 s, 3.6%, and 13,000 ± 600 J in CARLA. The proposed method achieved an average navigation success rate of 92.5%, consistently outperforming baseline models in safety, efficiency, and adaptability. These findings indicate the framework’s robustness and practical applicability for scalable AV deployment in real-world urban traffic conditions.
Keywords: autonomous vehicles; urban traffic optimization; reinforcement learning autonomous vehicles; urban traffic optimization; reinforcement learning

Share and Cite

MDPI and ACS Style

Alsuwaiket, M.A. Optimizing Autonomous Vehicle Navigation through Reinforcement Learning in Dynamic Urban Environments. World Electr. Veh. J. 2025, 16, 472. https://doi.org/10.3390/wevj16080472

AMA Style

Alsuwaiket MA. Optimizing Autonomous Vehicle Navigation through Reinforcement Learning in Dynamic Urban Environments. World Electric Vehicle Journal. 2025; 16(8):472. https://doi.org/10.3390/wevj16080472

Chicago/Turabian Style

Alsuwaiket, Mohammed Abdullah. 2025. "Optimizing Autonomous Vehicle Navigation through Reinforcement Learning in Dynamic Urban Environments" World Electric Vehicle Journal 16, no. 8: 472. https://doi.org/10.3390/wevj16080472

APA Style

Alsuwaiket, M. A. (2025). Optimizing Autonomous Vehicle Navigation through Reinforcement Learning in Dynamic Urban Environments. World Electric Vehicle Journal, 16(8), 472. https://doi.org/10.3390/wevj16080472

Article Metrics

Back to TopTop