Next Article in Journal
Braking Force Coordination Control for In-Wheel Motor Drive Electric Vehicles with Electro-Hydraulic Composite Braking System
Previous Article in Journal
Local Vehicle Density Estimation on Highways Using Awareness Messages and Broadcast Reliability of Vehicular Communications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization Method of Energy Saving Strategy for Networked Driving in Road Sections with Frequent Traffic Flow Changes

1
School of Mechanical and Automotive Engineering, Qingdao University of Technology, Qingdao 266520, China
2
School of Transportation, Shandong University of Science and Technology, Qingdao 266590, China
3
Intelligent Manufacturing Institute, Qingdao Huanghai University, Qingdao 266427, China
*
Author to whom correspondence should be addressed.
Vehicles 2025, 7(4), 118; https://doi.org/10.3390/vehicles7040118
Submission received: 30 August 2025 / Revised: 8 October 2025 / Accepted: 15 October 2025 / Published: 16 October 2025

Abstract

It is of great significance to construct a networked energy-saving driving strategy method and application framework to solve the problems of traffic disorder, speed fluctuations, and high energy consumption caused by frequent acceleration, deceleration, and lane changing of vehicles in road sections with variable traffic flow. Considering the mixed traffic scenario where autonomous vehicles and manually driven vehicles interact and infiltrate, a hybrid traffic flow vehicle energy-saving driving model was established, and the Dueling Double Deep Q-Network (D3QN) was used to optimize and solve the energy-saving driving model; Selecting Qingdao urban intersections as application scenarios, energy-saving driving strategy application facilities were constructed in simulation experiments to carry out simulation verification of energy-saving driving strategies for mixed traffic flow in the context of vehicle networking. The simulation results show that in different scenarios with different proportions of CAVs, the energy-saving strategy based on D3QN deep reinforcement learning algorithm can achieve fuel savings of 8.41%~6.67% compared to conventional strategies. Compared with the ordinary reinforcement learning algorithm Q-learning, its fuel saving rate is increased by 1.94%~1.5%, and the energy-saving effect becomes more significant with the increase of traffic density; From the perspective of dynamic characteristics, the speed stability under the control of D3QN algorithm is superior to Q-learning algorithm, and significantly better than conventional strategies, further highlighting the comprehensive advantages of D3QN algorithm in optimizing traffic flow status and energy consumption control. The energy-saving driving strategy in the networked environment can reduce fuel consumption caused by speed fluctuations and traffic flow frequency disturbances, and optimize the stability of traffic flow operation.

1. Introduction

Under the background of global energy crisis and increasingly severe environmental problems, energy conservation and emission reduction in the field of transportation has become a key link to achieve sustainable development. Especially in the sections with frequent traffic flow [1], such as the intersections of urban roads, suburban junctions and other areas, the driving state of vehicles changes frequently, and the energy consumption rises sharply. It is urgent to explore the energy-saving driving strategy in this scenario.
The optimization method of energy-saving driving of connected vehicles is a hot research content in recent years. Literature [2] designed a vehicle flow dynamic allocation system based on vehicle road collaboration based on new technologies such as c-v2x and edge computing, which reduced the operation delay in the weaving bottleneck area and fuel consumption. Wang [3] described a collaborative ecological driving (CED) system for the signal corridor, developed a longitudinal control model based on its role and distance from the intersection for artificial vehicles and different Cavs, and achieved the maximization of vehicle energy efficiency. Xu [4] for the ecological driving scenario of connected vehicles at urban signalized intersections, Legendre pseudospectral method is used to solve the optimal speed curve, and the system output is close to the global optimal solution, highlighting the excellent energy-saving potential of vehicles. Sun Xiaochong [5] proposed an ecological driving strategy for intelligent Internet connected vehicles considering secondary queuing. By improving IDM car following model and combining Internet connected information, an ecological driving strategy was constructed. Litiezhu [6] proposed a rule-based two-stage energy-saving driving optimization method. Dong [7] proposed an enhanced ecological control method (EEAC) strategy, designed the EEAC strategy using a hierarchical framework, and verified the energy efficiency improvement of the strategy. Literature [8,9] established an energy-saving control algorithm based on CACC. At this stage, the energy-saving optimization methods mainly focus on connected driving vehicles, and the research on mixed traffic flow is less.
In recent years, reinforcement learning algorithm to solve the problem of path optimization and energy-saving driving strategy in complex traffic environment has gradually increased. Chen [10] proposed a deep reinforcement learning method for end-to-end automatic driving, which achieved good results in complex urban driving scenes by introducing the sequential potential environment model. Saxena [11] successfully introduced a driving benchmark in dense traffic by modeling with deep reinforcement learning and implementing a continuous control strategy for the action space of autonomous vehicles. Feng Yao [12] proposed a lane changing trajectory planning method for intelligent connected vehicles based on deep reinforcement learning, which improved the safety and efficiency of intelligent connected vehicles and reduced fuel consumption. Jiang Han [13] established the CAV driving energy consumption model by constructing the state space of the reinforcement learning framework, trained the algorithm model with the deep certainty strategy, and verified the feasibility of the algorithm. Guo [14] designed a deep deterministic strategy gradient (DDPG) algorithm, which enables controlled vehicles to learn a perfect longitudinal fuel saving strategy and perform appropriate lane changing operations at an appropriate time to avoid Lane congestion. Zeng Xiaoqing [15] optimized the energy-saving driving model through reinforcement learning algorithm Q-learning. Some reinforcement learning algorithms have high complexity, large amount of calculation, and the Q value is overestimated.
To sum up, there is relatively little research on traditional energy-saving strategies for mixed traffic flows, and the reinforcement learning algorithms used are not suitable. Thus, the application framework of energy-saving driving strategy on the road with frequent traffic flow changes is constructed for the mixed traffic flow. A hybrid traffic flow driving model for the Internet of vehicles is established, the model refers to the experiment conducted by Alonso et al. in Santander City [16], and uses multi-source data (such as FCD) to calibrate the key parameters in the mixed traffic flow following and energy consumption model, which more realistically reflects the interaction and congestion dynamics between vehicles, and improves the effectiveness and robustness of energy-saving driving strategies in real complex environments. Then, the D3QN deep reinforcement learning algorithm is used to optimize the energy-saving driving model. Finally, the effectiveness of the energy-saving driving strategy is verified by simulation, and the fluctuation of traffic flow in frequency varying sections is greatly reduced.

2. Analysis of Mixed Traffic Flow Driving Scenarios

2.1. Analysis of Energy Saving Driving Strategies

In the upstream functional area of a signal controlled intersection, vehicles need to make lane selection in advance to match downstream turning requirements. In this dynamic process, the lane change decision of manually driven vehicles highly depends on the individual cognitive ability, risk preference, and environmental perception level of the driver. This highly subjective and heterogeneous decision-making mechanism is prone to frequent and non-collaborative lane changing behaviors, significantly exacerbating traffic flow disturbances and potentially leading to a decrease in local traffic efficiency or even potential conflict risks. In contrast, networked autonomous vehicles operate in a vehicle road collaborative environment. It relies on the vehicle networking platform to transmit high-precision, low latency real-time traffic information, including but not limited to signal phase and timing, road geometry topology, surrounding vehicle status, and global traffic situation. Based on these multi-source heterogeneous data, the collaborative decision-making module of CAVs can generate globally or locally optimal lane changing strategies, thereby achieving collaborative optimization of path planning and lane keeping/changing.
To optimize driving strategies for this mixed traffic flow, an energy-saving driving model for mixed traffic flow needs to be constructed. Firstly, the lane following characteristics of the mixed traffic fleet need to be identified from the perspective of traffic flow. This is because the lane following behavior between different vehicle models in mixed traffic flow can affect the overall operation status of the fleet. Therefore, accurately characterizing the lane following behavior within mixed traffic flow is crucial. Secondly, it is necessary to control the energy consumption of each vehicle model in the mixed traffic flow. In order to achieve unified measurement of energy consumption evaluation, it is necessary to establish a parameter unified mixed traffic flow energy consumption model based on the powertrain characteristics and driving features of different vehicle models. This model will provide a consistent benchmark for the quantitative evaluation of the effectiveness of energy-saving driving strategies. Finally, simulate and verify the energy-saving driving strategy. Figure 1 is a decision-making flowchart for networked autonomous vehicles. The blue dashed rectangle in the figure represents the three possible paths that CAV vehicles may choose.

2.2. Energy Saving Framework for Connected Driving

The vehicle networking energy-saving driving application framework constructed in this article includes four parts: roadside detection unit, roadside indication unit, vehicle networking information platform, hybrid fleet, and lane. The road test detection unit consists of a video detector, a laser radar, and a magnetic frequency detector. The lanes are divided into sections with stable traffic flow, sections guided by CAV vehicles, and sections with variable traffic flow frequency. The framework diagram of energy-saving driving application for connected vehicles is shown in Figure 2. The arrows in the figure indicate the direction of information transmission and indication.
The roadside detection unit detects road conditions and reports the number, location, and other information of the detected vehicles to the vehicle networking platform through the information transmission module. After receiving vehicle information on the road, the vehicle networking platform publishes the information to the connected vehicles in the road section. The connected vehicles control the vehicles to guide the fleet on the CAV vehicle guidance road section according to the integrated decision-making and control method of (HRL), and control the deceleration, acceleration, and lane changing behavior of the fleet. The energy-saving strategy control flowchart is shown in Figure 3.

3. A Driving Model for Networked Mixed Traffic Flow

3.1. Vehicle Following and Lane Changing Model

In the Internet of Vehicles environment, autonomous vehicle, equipped with laser radar, infrared video ranging and other on-board sensing equipment, can achieve braking control with shorter response time by virtue of the information interaction and collaboration capabilities of the Internet of Vehicles, so as to obtain higher precision and better numerical safety distance threshold. For manually driven vehicles, based on the differences in driver driving styles, they can be divided into three categories: aggressive, conservative, and cautious. Therefore, the driver adventure coefficient a is introduced. The acceleration of the preceding vehicle can reflect its motion state trend. When the current vehicle is decelerating, the following vehicle needs to reserve a larger safety distance. Based on this, the acceleration influence factor b of the preceding vehicle is introduced to improve and optimize the Gipps safety distance following model. The optimized model is shown in Equation (1):
D s a f e , n = v n ( t ) τ n + v n ( t ) 2 2 b n β v n + 1 ( t ) 2 2 b n + 1 + γ a n + 1 ( t ) τ n 2
In the formula: D s a f e , n represents the minimum safe following distance of vehicle n; v n ( t ) represents the speed of vehicle n at time t; τ n represents the reaction time required for vehicle n to apply emergency braking; b n represents the maximum deceleration of vehicle n, and a n + 1 ( t ) represents the acceleration of the preceding vehicle; For autonomous vehicles, the risk factor can be set as a fixed value β = 1 ; The risk factor of manually driven vehicles β ( 0 , 1 ) ; Sensitivity coefficient γ ( 0.5 , 1 ) , when a n + 1 ( t ) < 0 , this term was positive, increasing the safety distance; when a n + 1 ( t ) > 0 , this item was negative and the distance could be appropriately reduced.
After considering the acceleration of the preceding vehicle, the velocity of the preceding vehicle is corrected to the predicted value, as shown in Equation (2):
v n + 1 p r e d ( t + τ n ) = v n + 1 ( t ) + a n + 1 ( t ) τ n
In the formula: v n + 1 p r e d ( t + τ n ) indicate the predicted speed of the preceding vehicle.
Enter the formula for safe speed, as shown in Equation (3):
v s a f e , n ( t ) = b n τ n + β v n + 1 p r e d ( t + τ n ) 2 + b n 2 τ n 2 + 2 b n d n ( t )
In the formula: v s a f e , n ( t ) represents the safe speed at which vehicle n does not collide with the preceding vehicle at time t; d n ( t ) Indicate the distance between vehicle n and the preceding vehicle at time t.
The above artificial driving model belongs to the ideal physical model. In actual driving scenarios, when in a high-density traffic environment, the psychological pressure of drivers tends to increase, which may lead to unnecessary deceleration behaviors without the interference of the preceding vehicle (such as unconsciously releasing the accelerator pedal). To accurately characterize the probability characteristics of the random slowing behavior, this paper introduces the Richards growth curve model from the field of biology and constructs a functional mapping relationship between road traffic density and driver random slowing probability. Through this model, it is possible to simulate the dynamic process where the psychological burden on drivers gradually increases as road traffic density continues to increase, ultimately leading to a corresponding increase in the probability of implementing deceleration operations during the driving process. This is more in line with the complex characteristics of actual human driving behavior. Parametric setting for achieving expected model behavior. According to relevant research findings [17], set the initial value (basic slowing tendency under low density) to 0.2, the maximum slowing probability A under high density to 0.4, the growth efficiency (sensitivity of density to pressure) K to 0.05, the metabolic rate (improvement of psychological tolerance after high pressure adaptation) m to 0.95, and the random slowing probability curve equation to Equation (4):
P s l o w = 0.2 + 0.2 1 B e 0.05 ρ 1 0.95
B is the theoretical capacity of a road, which is the maximum number of vehicles that a unit length (1 km) of road can carry under ideal conditions; ρ represents the density of road traffic flow, which characterizes the distribution of vehicles within a unit length of the road; N is the actual total number of vehicles on the road; L is the length of the road in kilometers. In actual traffic scenarios, as the traffic density ρ continues to increase, the psychological burden on drivers exhibits dynamic changes: in the initial stage, the increase in traffic density increases the complexity of the driving environment, and the psychological burden on drivers accelerates and accumulates with the increase in density, directly manifested as the growth rate of the probability of slowing down with the machine gradually increasing (slowing down behaviors such as unnecessary deceleration, delayed operation response, etc.); When the traffic density exceeds a certain critical threshold (i.e., the inflection point of the curve), due to the high stress psychological state of the driver, the marginal stimulus effect of increasing traffic density on their psychology gradually decreases. At this time, the growth rate of the random slowing probability changes from increasing to decreasing, and finally approaches the theoretical probability upper limit of 0.4 asymptotically.
The lane change behavior for mixed traffic flow can only be triggered when the lane change conditions are met. Consider free lane changing behavior. When the speed on this lane is insufficient and the safety distance between the front and rear is within the constraint conditions, lane changing can be carried out.
The triggering condition is shown in Equation (5):
v l e a d < η v e g o ( η = 0.85 )
In the formula: v l e a d indicate the speed of the preceding vehicle in this lane, v e g o indicate the current speed of the vehicle, η representing the speed ratio threshold (triggering lane change demand), the value is 0.85.
The forward safety distance constraint is shown in Equation (6):
d l e a d > τ l e a d max ( v e g o , v l e a d ) + L v e h
In the formula: d l e a d indicate the distance to the vehicle in front of this lane; τ l e a d indicating forward safe time interval, 1.8 (aggressive)~2.2 (conservative); L v e h indicates standard vehicle length, ranging from 4.5 (sedan) to 5 (SUV).
The backward safety distance constraint is shown in Equation (7):
d l a g > τ l a g ( v l a g v e g o ) + γ ( v l a g v e g o ) 2 2 a l a g , m a x
In the formula: d l a g represents the distance to the vehicle behind the target lane; τ l a g indicate the safe time interval for the rear vehicle, ranging from 1.2 (aggressive) to 1.8 (conservative); γ representing the driver’s risk preference factor, ranging from 0.8 (aggressive) to 1.2 (conservative); a l a g , m a x indicate the maximum braking deceleration of the rear vehicle; v l a g indicates the speed of the vehicle behind the target lane.
The comprehensive decision function for lane changing is shown in Equation (8):
l a n e   c h a n g e = i f   ( v l e a d < η v e g o ) ( d l e a d > τ l e a d v r e f 1 , + L v e h ) ( d l a g > τ l a g Δ v + γ Δ v 2 2 a l a g , m a x ) 0 , o t h e r w i s e
where: v r e f = max ( v e g o , v l e a d ) , Δ v = v l a g v e g o .

3.2. Hybrid Fleet Energy Consumption Model

Existing studies usually establish energy consumption models for autonomous vehicles (AV) and manual vehicles (HDV). However, in mixed traffic flow, this separate modeling method will lead to high computational complexity and difficulty in collaborative optimization. Therefore, it is necessary to establish a unified energy consumption model for mixed fleet, considering that the existing vehicle types mainly include pure fuel vehicles, pure electric vehicles, and hybrid vehicles, and hybrid vehicles also include plug-in hybrid and extended range hybrid, it is possible to consider setting a power type coefficient a to divide hybrid vehicles into a combination of pure fuel vehicles and pure electric vehicles. The energy consumption of electric vehicles is mainly related to the power of the motor, while that of fuel vehicles is affected by the thermal efficiency of the engine. In order to facilitate the analysis of the energy consumption of the fleet from the same dimension, this article represents the energy consumption of both electric vehicles and fuel vehicles as kj. Specifically, in the energy consumption calculation of electric vehicles, the energy unit of the battery is converted from kw·h to kj units of energy value; In the energy consumption calculation of fuel vehicles, the calorific value of the consumed fuel is converted into an energy value in units of kj, assuming 1 kw = 3600 kj.
Existing research usually establishes energy consumption models for AV and HDV separately, which leads to the need to calculate and stack different vehicle models for hybrid fleets, making the calculations complex and difficult to optimize collaborative strategies. For the convenience of calculation, a unified energy consumption model is established by separately modeling the acceleration/uniform speed (a ≥ 0) and deceleration (a < 0) operating conditions.
Step 1: In the acceleration/uniform speed stage, the theoretical basis adopted by the electric part is: instantaneous power of the electric vehicle = base load + driving resistance power + acceleration power,
Driving resistance power v + v 3 (rolling resistance + air resistance), v + v 2 is simplified to reduce the risk of overfitting. The acceleration power a 2 reflects the nonlinear variation of motor efficiency with load (experimental calibration shows that the square term is better than the linear term). The specific expression is Equation (9):
P A V = k 0 + k 1 v + k 2 v 2 + k 3 a 2
In the formula: k 0 represents the basic power consumption of vehicle electronic devices, air conditioning, etc., set to 0.05 kw; k 1 v is the linear term of speed, represents the energy consumption to overcome rolling resistance (tire deformation, transmission friction), k 1 set to 0.003; k 2 v 2 is the square term of speed, representing the energy consumption of customer service air resistance, k 2 set to 1.8 × 10−4; k 3 a 2 is the square term of acceleration, representing the energy cost of kinetic energy changes (additional load on the motor/engine), k 3 set to 0.0012.
The basic theory used in the fuel section is that the fuel consumption rate is related to the effective fuel consumption rate (BSFC) of the engine, which is exponentially related to the speed (v) and load (a). The specific expression is Equation (10):
P H D V = exp ( m 0 + m 1 v + m 2 a )
In the formula: m 0 represents the basic fuel consumption of fuel vehicles at idle speed, m 0 set to −1.85; m 1 v representing the relationship between combustion efficiency and load, affected by speed, m 1 set to 0.04; m 2 a represents the non-linear relationship between throttle opening and fuel injection quantity, influenced by acceleration, m 2 is set to 0.08;
Step 2: During the deceleration phase, the basic theory used for the electric part is: The regenerative braking power is brake torque × speed. The experiment shows that torque is correlated with deceleration to the power of 0.8 (nonlinear energy recovery efficiency). The specific expression is Equation (11):
P A V _ b r a k e = k 0 + β | a | 0.8 β = 0.65 v > 5 m / s 0.35 v 5 m / s
In the formula: β | a | 0.8 represents the regenerative braking rate of electric vehicles, β represents the proportion of recovered kinetic energy, and the higher the speed, the greater the motor reaction force, and the higher the proportion of kinetic energy recovered.
The basic theory used in the fuel section is that the fuel injection amount during deceleration of a fuel vehicle is approximately equal to the idle fuel consumption (only maintaining engine operation), which is weakly correlated with speed (intake compensation). The specific expression is Equation (12):
P H D V _ b r a k e = exp ( m 0 + m 1 v )
Step 3: Substitute the power type coefficient a and obtain the unified form of the energy consumption model for the hybrid fleet as Equation (13):
P ( t ) = λ [ k 0 + k 1 v ( t ) + k 2 v ( t ) 2 + k 3 a ( t ) 2 ] + ( 1 λ ) e x p [ m 0 + m 1 v ( t ) + m 2 a ( t ) ] a 0 λ [ k 0 + β ( | a | 0.8 ) ] + ( 1 λ ) e x p [ m 0 + m 1 v ( t ) ] a < 0
Establishing a unified energy consumption model for hybrid fleets avoids the need for separately calculating energy consumption by vehicle type and solves the problem of complex calculations and difficult collaborative optimization strategies.

3.3. Energy Saving Driving Model for Hybrid Fleet

To find the optimal energy-saving driving strategy for road sections with variable traffic flow frequency, the energy-saving driving strategy can be divided into two types of implementation paths based on the traffic status (distribution of vehicles in the lane) of the nearby intersection obtained in the vehicle networking environment. When the traffic scene features perceive that the traffic flow ahead is in a steady state and the intersection signal light is green, guide the vehicle to follow the vehicle ahead in a stable mode within the vehicle guidance area; When a waiting queue vehicle is detected ahead and the adjacent lane belongs to the target available lane, guide the vehicle to perform a lane change operation within the vehicle guidance area. The above two scenarios were both carried out under the condition of satisfying the car following and lane changing model, ultimately resulting in a driving strategy that guides the safe and energy-saving operation of the vehicle.
Based on dynamic programming theory, the road driving space is subjected to spatially uniform discretization, and then deconstructed into a continuous multi-stage decision space. At the same time, the driving state of the vehicle is divided into three dimensions: distance, speed, and relative state. Based on the characteristics of actual driving scenarios, the safety distance model is extended to adapt to the constraints of vehicle driving behavior. Among them, the definition of moderate following distance is that the current following distance can meet the safety requirements for the following vehicle to follow at a constant speed. Therefore, the extended formula for calculating the relative following distance is Equations (14) and (15):
D m i d , n = v n ( t ) v n + 1 ( t ) 2 b n + 1 + v n ( t ) 2 2 b n v n + 1 ( t ) 2 2 b n + 1
D f a r , n = v n ( t ) + a n v n + 1 ( t ) b n + 1 + v n ( t ) 2 2 b n v n + 1 ( t ) 2 2 b n + 1
In the formula: D m i d , n is the moderate following distance of vehicle n; D f a r , n is the longer following distance of vehicle n; v n ( t ) is the following speed of vehicle n at time t; a n is the maximum acceleration of vehicle n; b n is the maximum deceleration of vehicle n.
Segmented solution is applied to the driving space ahead. Assuming the straight space length of the driving space ahead is L, the driving space is divided into n sub intervals, each with a length of l g a p . Generally, the value of l g a p is small, and the running speed within the interval can be simplified to take the average of the end of stage speeds. For vehicles that maintain a following state, the calculation formula for the running time of the subinterval can be obtained as shown in Equation (16):
T k = 2 l g a p v k + v k + 1
For vehicles engaged in lane changing behavior, let the lane width be l w i d t h , the approximate length traveled by the vehicle is l g a p 2 + l w i d t h 2 , At this time, the running time of the subinterval is shown in Equation (17):
T k = 2 l gap 2 + l width 2 v k + v k + 1
Assuming constant acceleration within the subinterval, combined with the energy consumption model calculation formula, the energy consumption calculation formula for each stage of state transition is shown in Equation (18):
E k = T k T k + 1 P ( t ) d t
In the formula: E k is the total energy consumption of stage k; P ( t ) is the instantaneous energy consumption of vehicles; T k is the running time for phase k.
The final determination of parameters for the mixed traffic flow following and lane changing model and unified energy consumption model constructed in this article will follow a data-driven calibration process. We will draw on the data fusion approach adopted by Alonso et al. and use floating car data collected from actual roads to invert and optimize model parameters. For example, by comparing the real vehicle trajectory provided by FCD with the trajectory output by the simulation model, adjusting the sensitivity parameters of the car following model, key indicators such as acceleration distribution and headway distribution of the simulated vehicle are matched with real data, thereby ensuring that the model can more accurately reproduce the interaction and energy consumption characteristics of vehicles in frequency variable flow sections.

4. Energy Saving Driving Strategy Based on Deep Reinforcement Learning

Adopting deep reinforcement learning algorithms to solve the optimal strategy for achieving energy-efficient driving. The flowchart of the deep reinforcement learning algorithm is shown in Figure 4. Traditional reinforcement learning algorithms are mostly suitable for discrete and low dimensional state spaces, which store each state and action in a table (such as Q-table). When the state space is large or continuous (such as continuous parameters such as speed and distance in autonomous driving), the table cannot be stored due to the explosion of dimensions, resulting in algorithm failure. The D3QN deep reinforcement learning algorithm introduces deep neural networks (CNN or fully connected networks) to process high-dimensional continuous state spaces, which can directly extract features from raw data (such as sensor data, image pixels) without the need for manual design of state features. For example, in energy-saving road scenarios, multi-dimensional continuous data such as vehicle speed, relative distance, and position of vehicles ahead can be directly input, and the state representation can be automatically learned through the network.

4.1. D3QN Algorithm

The D3QN algorithm model is shown in Figure 5. In the initial state, the agent inputs the environmental state matrix St into the prediction network, and outputs the Q values of each action after calculation by the prediction network. Then, using strategy ε g r e e d y , randomly select an action at to execute and interact with the environment to receive a reward r t . ε g r e e d y strategy refers to randomly selecting an action with a probability of ε and selecting the action corresponding to the maximum Q value calculated by the current prediction network with a probability of 1 ε , as shown in Equation (19):
a t = a i , ε arg   max Q s t , a t ; θ , 1 ε
The interaction between intelligent agents and the environment will generate a series of experience sequences < s t , r t , a t , s t + 1 > . Store it in the experience replay pool as training samples, and uniformly sample a certain batch size of sequences from it during each training session. As the algorithm continues to run, the number of samples in the experience replay pool increases, and the sampling of the prediction network becomes more diverse. The experience replay mechanism breaks the strong correlation between consecutive experiences, making each training independent and the results more reliable.
In deep reinforcement learning, the D3QN algorithm integrates two key technologies, Double and Dueling, to enhance the stability and performance of the model.
(1)
Double Q Network
Two networks are used in the model, one is the current network with an initial parameterization of Q , whose value changes with the operation of the algorithm, and the other is the target network with an initialization parameter of Q . The structures of the two networks are the same to facilitate parameter assignment.
Current network Q online ( s , a ; θ ) :
h 1 = R e L U ( W 1 s t + b 1 ) h 2 = R e L U ( W 2 h 1 + b 2 ) V ( s ; θ v ) = W v h 2 + b v A ( s , a ; θ a ) = W a h 2 + b a Q o n l i n e ( s , a ; θ ) = V ( s ; θ v ) + A ( s , a ; θ a ) 1 | A | a A ( s , a ; θ a )
In the formula: Q o n l i n e represents the final value predicted by the online network; s t state vector representing time step t; θ representing online network parameters; h representing the output of the hidden layer; W 1 , b 1 representing the weights and biases of the first layer; W 2 , b 2 representing the weights and biases of the second layer; V ( s ; θ v ) representing state value scalar; A ( s , a ; θ a ) representing the action advantage vector.
Target network Q t a r g e t ( s , a ; θ ) :
Q t a r g e t ( s , a ; θ ) = V ( s ; θ v ) + A ( s , a ; θ a ) 1 | A | a A ( s , a ; θ a )
The target network structure is the same as the online network, parameter θ Regular replication θ .
The prediction network is used to select actions, and the parameter Q is constantly updated; The target network is used to calculate the temporal difference value y, with the parameter Q fixed and replaced with the latest prediction network Q at regular intervals. The calculation formula for the target value Q is shown in Equation (22):
y t = r t + γ Q t a r g e t s t + 1 , arg   max a Q o n l i n e s t + 1 , a ; θ ; θ
(2)
Dueling Q Network
In the Dueling network architecture design, the Q network is divided into two parts. The first part is only related to state s and is independent of the specific action a used, denoted as the state value function v ( s ) . The second part is related to both state s and action a, denoted as the advantage function A ( s , a ) .
Q s ,   a = V s + A s , a
The D3QN algorithm uses a network based on Convolutional Neural Network (CNN), and its specific neural network structure is shown in Figure 6:

4.2. Loss Function

The loss function of D3QN is based on the parameter θ of the prediction network when selecting the target Q-value action, that is, selecting the action corresponding to the maximum Q-value in the prediction network in the current state, and then calculating the Q-value in the target network. This reduces the correlation between action selection and target Q-value calculation, effectively avoiding the problem of overestimation. In D3QN (Dueling Double DQN), Mean Squared Error (MSE) is used as the loss function by default, as shown in Equation (24):
L D 3 Q N = E r t + γ Q s t + 1 , arg max a Q ( s t + 1 , a ; θ ) ; θ Q ( s t , a t ; θ ) 2
In the formula: E represents the expected operation, the expected distribution of training data (experience replay pool samples); The instant reward obtained by agent r t after performing an action at time t; Q represents the target Q-value calculation term, reflecting the core logic of the “Double Q-Network”; Q ( s t , a t ; θ ) represents the Q value prediction of the current main Q network (parameter θ ) for executing action a t in state s t at time t; T represents the actual action performed by the agent at time t, based on sample data from the experience replay pool.

4.3. Reward Function

The reward function needs to evaluate the fuel consumption of sub intervals and the entire driving process, taking into account the fuel consumption optimality during each state transition and overall strategy learning. Therefore, the relative distance to the preceding vehicle in this lane is introduced as a feedback value in the reward function of the algorithm. Reduce the reward values in order of relative distance to indicate the promotion of following vehicles at a longer distance. The specific function is shown in Equation (25):
R ( s t , a t ) = α R P ( v t , a t ) T ( s t , a t ) + β h ( s t , a t ) + λ s a f e
In the formula: R ( s t , a t ) represents the reward value obtained by the agent when executing action a t in state s t ; α , β and λ are both weight coefficients used to balance the impact of different components in the reward function on the total reward; R represents the benchmark reward under ideal conditions; P ( v t , a t ) represents a function related to the feasibility of the action; T ( s t , a t ) represents the state transition constraint loss; h ( s t , a t ) represents the penalty function, used to characterize the special correlation benefits between state s t and action a t ; s a f e represents a security indicator function used to reinforce rewards for security behaviors or punishments for violations.

5. Simulation Experiment and Result Analysis

5.1. Simulation Experiment Design

The simulation experiment uses the open-source micro traffic simulation software SUMO (Simulation of Urban MObility) 1.19.0, which interacts with Python 3.9 through an external interface (TraCI) to obtain real-time simulation data, the Python script uses the traci library (version matched with the SUMO version) to connect to the SUMO simulation instance through a local socket (default port 8873). Each experimental scenario is independently repeated 30 times and initialized with pre-defined, different random seeds to ensure statistical significance and reproducibility of the results.
The simulation section of energy-saving driving strategy in this article is selected from the intersection of Qingdao Haigang Road and Binhai Avenue to the intersection of Dawan Port Road and Binhai Avenue, with a total length of 1610 m. The distance between the upstream and downstream of this section of intersection is relatively far, and the traffic flow is moderate. Selecting such a representative complex scene for in-depth verification aims to prove the potential of this method in solving such problems, rather than immediately pursuing the universality of its conclusion under all geographical and road conditions. Controlling the variables of the simulation scenario helps to analyze the effect of the strategy itself more clearly.
The simulation is studied in three scenarios, scenario one is a traffic scenario in a natural traffic state without setting any energy-saving strategies; Scenario 2 is a scenario where energy-saving driving strategies are set up, and the energy-saving driving strategy method is Q-Learning reinforcement learning; Scenario three is a scenario layout with energy-saving driving strategies set, and the energy-saving driving strategy method is the D3QN deep reinforcement learning method. According to the current Technical Standards for Highway Engineering, four lane first-class highway can adapt to an annual average daily traffic volume of 15,000 to 30,000 vehicles, so the upper limit of the simulation of vehicle flow density in this paper is set as 30 pcu·km−1. Considering that the proportion of autonomous connected vehicles in traffic flow varies, it will have a significant impact on the results of simulation experiments. In order to comprehensively verify the applicability of the proposed energy-saving driving strategy, simulation experiments will be conducted for various scenarios with different proportions of autonomous connected vehicles. By conducting simulation tests in mixed traffic flow environments with different proportions (gradient settings from low to high proportions), the performance of the strategy in various scenarios is analyzed in depth, and its adaptability to mixed traffic flow environments of different degrees is clarified to ensure that the strategy can cope with complex and changing traffic conditions in practical applications. The following is the reference data used in this simulation, as shown in Table 1:

5.2. Solution of Energy Saving Driving Strategy

The solution process uses the D3QN deep reinforcement learning algorithm constructed in Section 3 to iteratively learn the energy-saving driving strategy. The initial parameter settings of the D3QN energy-saving algorithm are shown in Table 2 below:
Firstly, the convergence of the algorithm in the iterative learning process is verified. Because the environmental state quantity discussed in this paper is large, and the scale of Q table and R table is large, through the analysis of convergence, it can be clear whether the obtained Q table is globally optimal, so as to avoid falling into the local optimal situation. In the simulation process, the convergence of the algorithm in the generation selection learning process is calculated, and the results are shown in Figure 6. It can be seen from the convergence curve in Figure 7 that after 2000 iterations of the cycle, the calculated vehicle fuel consumption becomes relatively fixed and has the characteristics of convergence.

5.3. Simulation Result

In the simulation process, for mixed traffic flow, the vehicle flow density takes 0 as the initial value and gradually increases according to the gradient of 5 pcu·km−1 until it reaches 30 pcu·km−1. For the proportion of automatic connected vehicles, it is initially set to 0 (that is, the simulation of the conventional vehicle scene with all manual driving in the road section), and then increases in order with a gradient of 20% to 100% (the scene of fully realizing automatic connected driving in the corresponding road section). Through iterative learning of the algorithm, the fuel consumption results of the controlled vehicle under each environment are finally obtained. The specific data are shown in Table 3 and Figure 8.
Under the same CAV proportion, with the increase of traffic density on the road, the driving space of controlled vehicles will be more restrained, resulting in the fluctuation of driving strategies. Therefore, the fuel consumption of different driving strategies is increasing to varying degrees, and the rise of scenario 2 and scenario 3 using energy-saving driving strategy is more gentle, it shows that when the road section is more crowded, dueling network can better evaluate the overall value of congestion, which is a “bad state”, and guide vehicles to preferentially choose actions that can avoid or alleviate congestion (such as slow down smoothly in advance), reflecting the excellent effect of the energy-saving driving strategy proposed in this paper.
The energy-saving effects of scenario 2 and scenario 3 are significantly higher than those of the conventional driving strategy in scenario 1. Table 3 further compares the energy-saving effects of D3QN energy-saving driving strategy and Q-learning energy-saving driving strategy, specifically the fuel consumption of the energy-saving strategy and the energy-saving effect of the conventional strategy under different CAV proportions and traffic densities.
It can be seen from the data results in Table 3 that the energy-saving effect of D3QN energy-saving strategy is always better than that of Q-learning energy-saving strategy. Through calculation, the fuel consumption of D3QN energy-saving strategy is 8.41%~6.67% lower than that of conventional strategy, and the fuel consumption of D3QN energy-saving strategy is 1.94%~1.5% lower than that of Q-learning energy-saving strategy.
Analyze the fluctuation of vehicle speed on the road with frequent traffic flow, count the overall vehicle speed under the three scenarios, and quantify the fluctuation amplitude by using the standard deviation. The comparison diagram of vehicle speed time series is shown in Figure 9, and the distribution diagram of vehicle speed probability density is shown in Figure 10.
It can be seen from Figure 9 that the fluctuation of the conventional strategy is obvious in a short period of time in the sections with frequent traffic flow, the fluctuation of the Q-learning strategy is significantly reduced, and the D3QN strategy is the lowest, indicating that the energy-saving driving strategy can reduce the fluctuation of traffic flow and improve the stability of traffic flow, and the D3QN strategy is superior to Q-learning.
According to the probability density distribution diagram of vehicle speed in Figure 10, the vehicle speed of conventional driving strategy is relatively dispersed, the energy-saving driving strategy is relatively centralized, and q3dn strategy is more centralized than Q-learning strategy. D3QN strategy can control the vehicle speed between 30~50 km/h, which greatly ensures the stability of traffic flow.

6. Conclusions

(1)
Based on the in-depth analysis of traffic flow frequency changing sections, combined with the Internet of vehicles technology, the application framework of energy-saving driving strategy for traffic flow frequency changing sections is constructed. Its core is that the network connected autonomous vehicles operate in the vehicle road collaborative environment. Relying on the high-precision and low delay real-time traffic information transmitted by the vehicle network platform, the collaborative decision-making module of Cavs can generate the global or local optimal lane change strategy, so as to realize the collaborative optimization of path planning and lane maintenance/transformation.
(2)
A car following and lane changing model of mixed traffic flow is constructed, and the behavior difference between manual driving and intelligent driving vehicles is accurately described by introducing the driver risk coefficient, the influence factor of the acceleration of the vehicle in front and the random slowing probability model; The unified energy consumption model integrates the energy consumption calculation of fuel vehicles, electric vehicles and hybrid vehicles through the power type coefficient, which solves the complexity of superposition calculation by vehicle type, and provides a unified benchmark for collaborative optimization. Combined with the real-time information interaction (roadside detection, platform decision-making, vehicle collaboration) in the Internet of vehicles environment, the model can dynamically adjust the driving strategy (such as stable car following or safe lane changing) according to the traffic state, taking into account the safety and energy conservation.
(3)
D3QN deep reinforcement learning algorithm is introduced as the learning engine of energy-saving driving strategy, which not only solves the limitations of traditional reinforcement learning, but also optimizes the balance between energy saving and safety. By separating the state value function and the action advantage function, the algorithm gives priority to long-term energy-saving actions, and introduces the safety distance penalty term and random moderation probability constraint, so that CAV can take into account the minimum fuel consumption and collision risk avoidance when guiding the hybrid fleet.
(4)
The simulation scenario is from the intersection of Haigang road and Binhai Avenue in Qingdao to the intersection of dawangang road and Binhai Avenue. According to the simulation results, compared with the conventional strategy, D3QN strategy can save 8.41%~6.67% of fuel consumption under different proportions of intelligent connected vehicles (CAV); Compared with Q-learning strategy, it can still save 1.94%~1.5% of fuel consumption. And with the increase of traffic density, this energy-saving advantage is more obvious. When the traffic flow density increases from 0 to 30 pcu·km−1, the increase of fuel consumption with D3QN strategy is more gentle, indicating that it has stronger energy-saving adaptability in congestion scenarios. In terms of traffic flow stability, D3QN strategy can effectively reduce the speed fluctuation. Through the quantitative analysis of the standard deviation, the vehicle speed of the conventional strategy fluctuates violently in a short time. Although the Q-learning strategy has improved, the speed fluctuation of the D3QN strategy is the lowest. From the probability density distribution of vehicle speed, the D3QN strategy can centrally control the vehicle speed in the range of 30~50 km/h, significantly improve the stability of traffic flow, and reduce the energy consumption waste and traffic efficiency loss caused by frequent acceleration and deceleration.
(5)
The simulation verification of this study was only conducted on specific road sections in Qingdao, and the universality of the conclusion under different road topologies needs further verification. The current micro models have not fully considered the interdependence between link costs and traffic flow. Therefore, future core work will focus on the following three points: (a) establishing a macro micro integrated model: coupling this micro simulation with the macro transportation system model, and evaluating the system level net effect of energy-saving strategies through iterative feedback (such as whether it will trigger traffic volume transfer). (b) Implement data-driven model calibration: Focus on utilizing floating car data and probe car data, following advanced data fusion methods, to empirically calibrate the traffic flow basic map, car following model parameters, and energy consumption model parameters, in order to improve the accuracy of model characterization in real environments. (c) Validation at an intelligent connected experimental field: Referring to the model of advanced experimental fields such as Santander, high-quality trajectory, flow, and energy consumption data are obtained in real urban environments to ultimately validate and enhance the practicality and robustness of this framework.

Author Contributions

Conceptualization, M.G. and D.Q.; methodology, K.W. and Y.C.; simulation and validation, M.G.; writing—original draft preparation, M.G. and D.Q.; writing—review and editing, D.Q. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant no. 52272311).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dai, S.; Qu, D.; Meng, Y.; Yang, Y.; Duan, Q. Evolutionary Game Mechanisms of Lane Changing for Intelligent Connected Vehicles on Traffic Flow Frequently Changing Sections. Complex Syst. Complex. Sci. 2024, 21, 128–135+153. [Google Scholar]
  2. Qu, D.; Liu, H.; Yang, Z.; Dai, S. Dynamic allocation mechanism and model of traffic flow in bottleneck section based on vehicle infrastructure cooperation. J. Jilin Univ. (Eng. Technol. Ed.) 2024, 54, 2187–2196. [Google Scholar]
  3. Wang, Z.; Wu, G.; Barth Matthew, J. Cooperative eco-driving at signalized intersections in a partially connected and automated vehicle environment. IEEE Trans. Intell. Transp. Syst. 2019, 21, 2029–2038. [Google Scholar] [CrossRef]
  4. Xu, S.; Li, S.; Deng, K.; Li, S.; Cheng, B. A unified pseudospectral computational framework for optimal control of road vehicles. IEEE/ASME Trans. Mechatron. 2014, 20, 1499–1510. [Google Scholar] [CrossRef]
  5. Sun, C.; Li, X.; Hu, H.; Yu, W. Eco-driving strategy for intelligent connected vehicles considering secondary queuing. J. Transp. Eng. Inf. 2023, 21, 92–102. [Google Scholar]
  6. Li, T.; Xie, B.; Liu, T.; Chen, H.; Wang, Z. A Rule-based Energy-saving Driving Strategy for Battery Electric Bus at Signalized Intersections. J. Transp. Syst. Eng. Inf. Technol. 2024, 24, 139–150. [Google Scholar]
  7. Dong, H.; Zhuang, W.; Chen, B. Enhanced eco-approach control of connected electric vehicles at signalized intersection with queue discharge prediction. IEEE Trans. Veh. Technol. 2021, 70, 5457–5469. [Google Scholar] [CrossRef]
  8. Lakshmanan, V.K.; Sciarretta, A.; El Ganaoui-Mourlan, O. Cooperative eco-driving of electric vehicle platoons for energy efficiency and string stability. IFAC-Pap. 2021, 54, 133–139. [Google Scholar] [CrossRef]
  9. Kim, Y.; Guanetti, J.; Borrelli, F. Compact cooperative adaptive cruise control for energy saving: Air drag modelling and simulation. IEEE Trans. Veh. Technol. 2021, 70, 9838–9848. [Google Scholar] [CrossRef]
  10. Chen, J.; Li, S.E.; Tomizuka, M. Interpretable end-to-end urban autonomous driving with latent deep reinforcement learning. IEEE Trans. Intell. Transp. Syst. 2021, 23, 5068–5078. [Google Scholar] [CrossRef]
  11. Saxena, D.M.; Bae, S.; Nakhaei, A.; Fujimura, K.; Likhachev, M. Driving in dense traffic with model-free reinforcement learning. In Proceedings of the IEEE International Conference on Robotics and Automation, Paris, France, 31 May–31 August 2020; pp. 5385–5392. [Google Scholar]
  12. Feng, Y.; Jing, S.; Hui, F.; Zhao, X.; Liu, J. Deep reinforcement learning-based lane-changing trajectory planning method of intelligent and connected vehicles. J. Automot. Saf. Energy 2022, 13, 705–717. [Google Scholar]
  13. Jiang, H.; Zhang, J.; Zhang, H.; Hao, W.; Ma, C. A Multi-objective Traffic Control Method for Connected and Automated Vehicle at Signalized Intersection Based on Reinforcement Learning. J. Transp. Inf. Saf. 2024, 42, 84–93. [Google Scholar]
  14. Guo, Q.; Angah, O.; Liu, Z.; Ban, X. Hybrid deep reinforcement learning based eco-driving for low-level connected and automated vehicles along signalized corridor. Transp. Res. Part C Emerg. Technol. 2021, 124, 102980. [Google Scholar] [CrossRef]
  15. Zeng, X.; Zhu, M.; Guo, K.; Wang, Y.; Feng, D. Optimization of Energy-Saving Driving Strategy on Urban Ecological Road with Mixed Traffic Flows. J. Tongji Univ. (Nat. Sci.) 2024, 52, 1909–1918. [Google Scholar]
  16. Alonso, B.; Musolino, G.; Rindone, C.; Vitetta, A. Estimation of a fundamental diagram with heterogeneous data sources: Experimentation in the city of santander. ISPRS Int. J. Geo-Inf. 2023, 12, 418. [Google Scholar] [CrossRef]
  17. Ying, P.; Zeng, X.; Song, H.; Shen, T.; Yuan, T. Energy-efficient train operation with steep track and speed limits: A novel Pontryagin’s maximum principle-based approach for adjoint variable discontinuity cases. IET Intell. Transp. Syst. 2021, 15, 1183–1202. [Google Scholar] [CrossRef]
Figure 1. Decision making flow chart of networked autonomous vehicles.
Figure 1. Decision making flow chart of networked autonomous vehicles.
Vehicles 07 00118 g001
Figure 2. Framework diagram of energy-saving driving application for connected vehicles.
Figure 2. Framework diagram of energy-saving driving application for connected vehicles.
Vehicles 07 00118 g002
Figure 3. Energy saving Strategy Control Flow Chart.
Figure 3. Energy saving Strategy Control Flow Chart.
Vehicles 07 00118 g003
Figure 4. Flow Chart of Deep Reinforcement Learning Algorithm.
Figure 4. Flow Chart of Deep Reinforcement Learning Algorithm.
Vehicles 07 00118 g004
Figure 5. D3QN algorithm model diagram.
Figure 5. D3QN algorithm model diagram.
Vehicles 07 00118 g005
Figure 6. D3QN Dual Branch Network Architecture Diagram.
Figure 6. D3QN Dual Branch Network Architecture Diagram.
Vehicles 07 00118 g006
Figure 7. Convergence Curve of D3QN Algorithm.
Figure 7. Convergence Curve of D3QN Algorithm.
Vehicles 07 00118 g007
Figure 8. Schematic diagram of the relationship between fuel consumption and traffic density under different proportions of CAV.
Figure 8. Schematic diagram of the relationship between fuel consumption and traffic density under different proportions of CAV.
Vehicles 07 00118 g008aVehicles 07 00118 g008b
Figure 9. Comparison of Vehicle Speed Time Series.
Figure 9. Comparison of Vehicle Speed Time Series.
Vehicles 07 00118 g009
Figure 10. Probability density distribution of vehicle speed.
Figure 10. Probability density distribution of vehicle speed.
Vehicles 07 00118 g010
Table 1. Simulation Parameters of Mixed Traffic Flow Model.
Table 1. Simulation Parameters of Mixed Traffic Flow Model.
ParameterSymbolNumerical Value
road length/m L 1610
number of lanes m 3
CAV conductor/m l c a v 5
car conductor/m l h v 5
Speed limit of road section (km·h−1) v 0 30
deceleration/(m·s−2) a 1 0.6
maximum deceleration/(m·s−2) a 2 1
starting position of road section/m x 1 415
end position of road section/m x 2 125
traffic density/(pcu·km−1) ρ 0~30
stochastic moderation probability P s l o w 0.2~0.5
phase interval/m l g a p 10
Table 2. D3QN Energy saving Algorithm Parameter Settings.
Table 2. D3QN Energy saving Algorithm Parameter Settings.
ParameterNumerical Value
number of training rounds500
Pre-experimental steps500
minimum batch size100
γ 0.90
λ 0.001
ε 0.5
parameter update cycle50 steps
network layers3
experience playback pool size500
state matrix size1 × 12 × 1 + 1
action space size3 × 3
Table 3. Simulation Results Data Table.
Table 3. Simulation Results Data Table.
Traffic Flow
Density
CAV Ratio/%
Strategy020406080100
Fuel UseEffectFuel UseEffectFuel UseEffectFuel UseEffectFuel UseEffectFuel UseEffect
0Q-learning109.804.69%109.504.95%109.305.12%109.105.30%108.905.47%108.705.64%
D3QN108.206.08%108.006.25%107.806.42%107.606.60%107.406.77%107.206.94
5Q-learning120.404.82%119.904.69%119.504.63%119.104.57%118.704.50118.304.44
D3QN118.606.25%118.206.04%117.805.99%117.405.93%117.005.87116.605.82
10Q-learning129.105.21%128.405.10%127.905.05%127.405.00%126.904.94132.904.89
D3QN127.006.76%126.406.58%125.906.53%125.406.49%124.906.44126.406.40
15Q-learning135.605.57%134.805.40%134.205.29%133.605.18%133.005.07132.404.95
D3QN133.207.24%132.507.02%131.906.91%131.306.81%130.706.71130.106.60
20Q-learning145.805.81%144.905.67%144.105.63%143.305.60%142.505.57144.705.53
D3QN143.107.56%142.307.36%141.607.27%140.907.18%140.207.09139.507.00
25Q-learning150.806.22%149.806.08%148.906.06%148.006.03%147.106.01146.205.98
D3QN147.808.08%146.907.90%146.107.82%145.307.75%144.507.67143.707.59
30Q-learning157.506.53%156.406.40%155.506.33%154.606.25%153.706.17152.806.09
D3QN154.208.52%153.208.32%152.308.25%151.408.18%150.508.11149.608.05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, M.; Qu, D.; Wang, K.; Chen, Y.; Zhan, J. Optimization Method of Energy Saving Strategy for Networked Driving in Road Sections with Frequent Traffic Flow Changes. Vehicles 2025, 7, 118. https://doi.org/10.3390/vehicles7040118

AMA Style

Gao M, Qu D, Wang K, Chen Y, Zhan J. Optimization Method of Energy Saving Strategy for Networked Driving in Road Sections with Frequent Traffic Flow Changes. Vehicles. 2025; 7(4):118. https://doi.org/10.3390/vehicles7040118

Chicago/Turabian Style

Gao, Minghao, Dayi Qu, Kedong Wang, Yicheng Chen, and Jintao Zhan. 2025. "Optimization Method of Energy Saving Strategy for Networked Driving in Road Sections with Frequent Traffic Flow Changes" Vehicles 7, no. 4: 118. https://doi.org/10.3390/vehicles7040118

APA Style

Gao, M., Qu, D., Wang, K., Chen, Y., & Zhan, J. (2025). Optimization Method of Energy Saving Strategy for Networked Driving in Road Sections with Frequent Traffic Flow Changes. Vehicles, 7(4), 118. https://doi.org/10.3390/vehicles7040118

Article Metrics

Back to TopTop