Next Article in Journal
Development of a Bamboo Toothbrush Handle Machine with a Human–Machine Interactive Interface for Optimizing Process Conditions
Previous Article in Journal
Estimation of the Extent of the Vulnerability of Agriculture to Climate Change Using Analytical and Deep-Learning Methods: A Case Study in Jammu, Kashmir, and Ladakh
Previous Article in Special Issue
Emission Quantification for Sustainable Heavy-Duty Transportation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Variable Speed Limit Control for the Motorway–Urban Merging Bottlenecks Using Multi-Agent Reinforcement Learning

Department of Control for Transportation and Vehicle Systems, Faculty of Transportation Engineering and Vehicle Engineering, Budapest University of Technology and Economics, Műegyetem rkp. 3, H-1111 Budapest, Hungary
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(14), 11464; https://doi.org/10.3390/su151411464
Submission received: 31 May 2023 / Revised: 17 July 2023 / Accepted: 19 July 2023 / Published: 24 July 2023
(This article belongs to the Special Issue Control System for Sustainable Urban Mobility)

Abstract

:
Traffic congestion is a typical phenomenon when motorways meet urban road networks. At this special location, the weaving area is a recurrent traffic bottleneck. Numerous research activities have been conducted to improve traffic efficiency and sustainability at bottleneck areas. Variable speed limit control (VSL) is one of the effective control strategies. The primary objective of this paper is twofold. On the one hand, turbulent traffic flow is to be smoothed on the special weaving area of motorways and urban roads using VSL control. On the other hand, another control method is provided to tackle the carbon dioxide emission problem over the network. For both control methods, a multi-agent reinforcement learning algorithm is used (MAPPO: multi-agent proximal policy optimization). The VSL control framework utilizes the real-time traffic state and the speed limit value in the last control step as the input of the optimization algorithm. Two reward functions are constructed to guide the algorithm to output the value of the dynamic speed limit enforced within the VSL control area. The effectiveness of the proposed control framework is verified via microscopic traffic simulation using simulation of urban mobility (SUMO). The results show that the proposed control method could shape a more homogeneous traffic flow, and reduces the total waiting time over the network by 15.8%. In the case of the carbon dioxide minimization strategy, the carbon dioxide emission can be reduced by 10.79% in the recurrent bottleneck area caused by the transition from motorways to urban roads.

1. Introduction

As an important transportation infrastructure, motorways represent the overall level of a country’s transportation system to a large extent and play a vital role in the development of the national economy. However, motorways are also facing increasingly frequent traffic congestion [1]. As a road section that is prone to generating and spreading congestion, the traffic bottleneck area is an important research object to improve motorway traffic management capabilities. According to the specific time and location of occurrence, traffic bottleneck areas can be classified into recurring and nonrecurring bottleneck areas [2]. Recurring traffic bottleneck areas are the confluence area, diversion area, and weaving area formed when the main line of the motorway merges with the entrance ramp and exit ramp. Congestion frequently occurs in recurring traffic bottleneck areas [3]. The closure of lanes due to road construction, bad weather, or traffic accidents has stochastic characteristics, which are the causes of nonrecurring bottleneck areas [4,5,6].
Compared with motorways, the distribution of entrance and exit ramps at the merging area of motorways and urban roads is more concentrated, and adjacent entrances and exits are more closely connected with urban roads, resulting in more traffic conflicts in the merging area of motorways and urban roads [7,8]. Compared with urban roads, the merging area of motorways and urban roads often service more traffic demand due to the closed traffic environment and connectivity of settlements and cities, resulting in a broader range of traffic bottlenecks and faster spread speed [9,10].
As a part of the intelligent transportation system (ITS), variable speed limit (VSL) control has been widely used as one of the traffic control measures to improve traffic efficiency [11,12,13], benefit the environment [14,15], and enhance traffic safety [16,17] in bottleneck areas. By adjusting the speed limit of the main line upstream of the bottlenecks, the number of vehicles entering the motorway bottleneck area during the congestion period is controlled within a certain range to keep the traffic state more uniform and stable. VSL control methods in traffic management can be divided into different categories according to their basic approaches and techniques. The most used VSL control strategies are rule-based approaches [18,19] and model-based approaches. Rule-based VSL control approaches use a set of predefined rules and thresholds to determine the appropriate speed limits. These rules take into account factors such as traffic density, volume, occupancy, and historical data to set speed limits. Rule-based VSL approaches may cause traffic flow fluctuations when the traffic state exceeds or falls below the set thresholds. Model-based VSL control approaches utilize mathematical models to represent and predict traffic behavior. Model-based approaches using optimization algorithms to determine the optimal speed limits. Model-based approaches are further classified into open-loop optimization approaches [20,21,22,23,24,25] and feedback control approaches [26,27,28]. Model-based approaches require accurate models to describe the state of traffic flow. Due to the need for a large number of numerical calculations and the fact that the model contains many parameters to be calibrated, the model-based VSL control approaches have limitations in large-scale traffic control applications. In addition, the portability of these VSL control approaches needs further work. This means when applying the control algorithm to a new scenario, the traffic flow model needs to be re-calibrated, and the controller parameters need to be adjusted manually.
With the development of artificial intelligence technology, VSL control based on reinforcement learning (RL) can automatically adapt to various traffic environments and achieve optimal control effects without specific traffic flow models [29,30,31]. Existing RL-based VSL control approaches use the matured deep Q-network (DQN) developed in 2015 [32]. The continuous development of reinforcement learning research has brought new solutions to RL-based VSL control problems.
This study proposes a fully cooperative multi-agent reinforcement learning framework to solve the multi-section VSL control optimization problem at the merging area of motorways and urban roads. This framework is based on the multi-agent proximal policy optimization (MAPPO) algorithm proposed in 2021 [33], which has been proven powerful for target localization [34], production scheduling [35], and trajectory planning [36] but has not been utilized to solve the VSL optimization problem. Under the constraints of rational speed, traffic flow state information, action representing discrete speed limit values, and reward composed of occupancy data and carbon dioxide emissions are designed to smooth traffic and reduce emissions. The proposed multi-agent reinforcement learning framework is tested on the microscopic traffic simulator simulation of urban mobility (SUMO) [37].

2. Methodology

This study presents a VSL control framework, applying multi-agent reinforcement learning (MARL) with the proximal policy optimization (PPO) algorithm. Section 2.1 introduces the literature background of RL. Section 2.2 presents the basis of the PPO algorithm. Section 2.3 extends the basic PPO algorithm to MARL.

2.1. Reinforcement Learning

Reinforcement learning (RL) is a machine learning (ML) algorithm. During the learning process, the agent obtains the optimal strategy by trying different action choices and adjusting the evaluation value of the action according to the feedback of the environment. During each interaction between the agent and the environment, the input to the agent is the environment state s. The agent chooses action a as the output to transit the state of the environment to state s , and at the same time, the agent receives reward r. RL aims to find a series of optimal action sets [38]. Deep learning (DL) is a method based on representational learning of data in ML, which abstracts the features contained in the data based on the value of the original data for representation [39]. The core of DL is the deep neural network (DNN), inspired by the principles of neural networks (NNs) in biology. NN abstracts the structure of the human brain and the response mechanism to external stimuli in a mathematical model.
Deep reinforcement learning (DRL) is an algorithm that combines RL with the ability to learn from the environment and DL with powerful representation capabilities. DRL has two branches: Q-learning and policy gradient (PG) algorithms [40]. Q-learning is mainly based on the iteration of the action–value function to model the value of the state space to find the optimal strategy. The Q-learning algorithm has excellent interpretability and debuggability but relies on extensive experience sample storage to establish an accurate model. PG directly outputs the probability of the action. Since no additional state–action value is introduced, PG requires fewer hyperparameters to be tuned, and due to the gradient method utilized, PG will update toward the direction of the optimization strategy, which means it has good convergence. But the disadvantage is that PG can easily converge to a local optimal value.
The actor–critic (AC) algorithm [41] combines the advantages of Q-learning and PG. AC uses the single-step update advantage of Q-learning to allow the Q-network as a critic to learn offline (i.e., an agent that learns through experience samples that are not acquired by itself). The critic provides an evaluation of the actor’s action or the direction of the gradient descent in a single-step policy update, which greatly improves the update frequency of the actor and accelerates the learning rate. In this way, the problem of the low efficiency of the PG in episode updates can be solved, but there is a problem of convergence difficulty. To solve the difficulty of convergence of the AC algorithm and speed up training, the asynchronous advantage actor–critic (A3C) algorithm [42] pushes the AC algorithm into multi-threads for synchronous training. The emerging trust region policy optimization (TRPO) algorithm [43] solves the problem of fluctuations in the A3C algorithm when balancing the variance and bias of the model.

2.2. Proximal Policy Optimization

Based on PG and TRPO, the proximal policy optimization (PPO) algorithm [44] is proposed as an improved method. The PPO algorithm solves the shortcomings of previous RL algorithms, such as low data utilization efficiency, poor robustness of the PG algorithm, and the complexity of the TRPO algorithm. The PG algorithm is based on calculating the estimator of the policy gradient and substituting it into the stochastic gradient ascent algorithm. The unbiased estimate of the gradient is
δ J ^ ( θ ) = E t ^ [ δ θ l o g π θ ( a t | s t ) A ^ t ]
where π θ is a random strategy for actor network parameters θ ; A ^ t = Q t V t is the advantage function estimated by critic; Q t is the state–action value function; and V t is the state–value function. a t is the action space, and s t is the state space. However, PG needs to re-interact with the environment after each update of the parameters, calculate the advantage function of the new strategy, and then update the parameters, making the update speed slow. The TRPO algorithm uses the importance sampling method. TRPO adopts the new and old strategy networks θ and θ o l d to allow the network to interact with the environment and use the collected data to train the network θ so that the agent can perform multiple parameter updates in one interaction with the environment, which improves the update speed.
In addition, to prevent the gap between the new network θ and the old network θ o l d from being too large, the relative entropy is used to measure the difference between the two, then the objective function is
m a x θ J ( θ ) = E t ^ π θ ( a t | s t ) π θ o l d ( a t | s t ) A ^ t ,
s u b j e c t t o E t ^ [ K L [ π θ o l d ( a t | s t ) , π θ ( a t | s t ) ] ] ζ
where ζ is the radius of the trust domain, and K L is the boundary conditions. The PPO adds the KL of TRPO as a penalty item to the objective function and applies an adaptive penalty of the β parameter to the KL divergence. The actor’s objective function is
m a x θ J K L ( θ ) = E t ^ π θ ( a t | s t ) π θ o l d ( a t | s t ) A ^ t β D K L [ π θ o l d ( a t | s t ) | | π θ ( a t | s t ) ]
where D K L ( π θ o l d | | π θ ) represents the divergence of K L between the π θ o l d and the π θ controls the difference in each episode of policy updates. β is adaptively adjusted according to the preset K L divergence threshold. In practice, the clip function is used to limit the probability ratio r t ( θ ) to ( 1 ε , 1 + ε ) , where ε is a hyperparameter, also taking π θ π θ o l d A ^ t as the optimization object. When A ^ t > 0 , increase the probability of π θ ( a t | s t ) and vice versa. The actor’s objective function is
m a x θ J c l i p ( θ ) = E t ^ c l i p π θ ( a t | s t ) π θ o l d ( a t | s t ) , 1 ε , 1 + ε A ^ t , π θ ( a t | s t ) π θ o l d ( a t | s t ) A ^ t
The optimization framework of the PPO algorithm for traffic control is shown in Figure 1.

2.3. Multi-Agent Proximal Policy Optimization

An individual agent is inadequate to handle multi-section VSL control. For this reason, the multi-agent reinforcement learning (MARL) framework is developed. Based on the centralized training with decentralized execution (CTDE) framework, the PPO algorithm is extended to the MAPPO algorithm. That is, each individual PPO agent is trained using a global value function. After the training is completed, each individual PPO agent obtains a decentralized strategy, and action is taken locally based on this strategy. Then, centralized training is applied to make the decentralized strategies learned by individual PPO agents work cooperatively. In the fully cooperative MAPPO algorithm, all individual PPO agents share common reward signals. During centralized training, a global critic network is used to evaluate the state value and calculate A ^ t i using the generalized advantage estimation (GAE) method. During decentralized execution, each individual PPO agent relies on its local observation state to realize a distributed decision-making interactive environment. The global critic network uses GAE to estimate a common A ^ t based on the reward R t i obtained by each agent’s action trajectory during centralized training. The objective function for agent i is
m a x θ J i ( θ ) = E t ^ c l i p π θ ( a t i | s t i ) π θ o l d ( a t i | s t i ) , 1 ε , 1 + ε A ^ t i , π θ ( a t i | s t i ) π θ o l d ( a t i | s t i ) A ^ t i

3. Simulation Environment

Unlike other control algorithms, the DRL algorithm does not directly learn the model from a given data set but learns by interacting with the environment continuously to generate data. Implementing the training of the algorithm directly in the real traffic network is not feasible, which will bring huge costs and safety hazards. The usual feasible method is to train the RL algorithm with the traffic simulation platform and then migrate the trained model to the real traffic network. In addition, the multi-agent RL framework requires an environment to interact. To implement the proposed multi-agent reinforcement learning framework, an accurate and efficient traffic simulation platform is essential. The simulation of urban mobility (SUMO) is an open-source, multi-modal traffic simulator that can be extended dynamically and is highly customizable through the embedded traffic control interface (TraCI).
To test the control performance of the proposed multi-agent reinforcement learning framework, a dynamic traffic network representing the merging area when the motorway transited to the urban roads i sgenerated based on the OpenStreetMap (OSM) data. Figure 2 shows the schematic diagram of the proposed MAPPO-based VSL control at the typical recurrent bottleneck area located on the border of Budapest in Hungary. The intersection with the ramps and the existence of traffic lights in the urban road network leads to frequent lane-changing behavior of vehicles in the merging area, which further leads to congestion.
The geometric structure of this bottleneck area in SUMO is shown in Figure 3. The control section is divided into seven homogeneous sections, named “section_a”, “section_b”, “section_c”, “section_d”, “section_e”, “section_f”, and “section_g”, each of which is one kilometer long. There are four vehicle routes in the road network. Two of them start from the motorways and end on the urban roads Alkotás Street and Hegyalja Street, while the other two are the on-ramp flow starting from Egér Street. According to on-site historical measurement data, traffic flows are assigned to the road network at a rate of 2/3, with more vehicles traveling to Alkotás Street. The total demand on the main road is 4657 veh/h and 1039 veh/h on the ramp. Two traffic light control systems exist at the end of this regional network. The control schema is the fixed cycle with 90 s cycle time. The green phase is 60 s, and the yellow phase is 3 s.
To connect the dynamic traffic environment in SUMO with the MAPPO algorithm, a custom environment “SUMO_Env()” is created in PyCharm, which contains the main subset functions needed for the training, initialization simulation, state representation, action space, reward, termination, reset, and step functions.
The initialization of the simulation function initializes the settings of the simulation environment. The simulation time is set to 7200 s. The original speed limit of the road network is set to 27.78 m/s (100 km/h) for motorways and 13.89 m/s (50 km/h) for urban roads according to the actual traffic rules.
The states obtained by RL agents in different environments are various. For example, in the robot simulation environment, the state representation is mainly composed of joint dynamics data in the physical sense [45]. In the game task, the state information agents obtained from the environment are almost exclusively image data [46]. In financial decision tasks, the state information includes stock market bonds, various K-line indicators, transactions, and financial data [47]. Detailed state representation helps RL algorithms extract critical information. The state-space representation in this paper is the vehicle occupancy data in each section, including main roads and ramps.
The action space represents the actions of the PPO agents applied to the traffic environment. Six discrete speed limit values make up the action space in this study for the dynamic traffic environments, which is [ 36.11 , 30.55 , 25 , 19.44 , 13.89 , 8.3 ] , the unit is m/s. The goal of agents is to cooperate with each other to find the optimal speed limits set for different road sections.
The reward function setting is very critical in RL problems. The reward function represents the optimization goal of the training. It is the only feedback from the environment to the actions that the agent takes. Commonly used reward functions in RL-based VSL control are total time spent (TTS) [30,48], density distribution [12,49], mean speed in the bottleneck [50] that is utilized to improve traffic efficiency and the negative sum of the values of tail gas emissions [50,51,52] that take the environmental sustainability into account. In this paper, two reward functions are introduced to smooth the traffic flow and reduce traffic emissions. The first reward function is the negative of the standard deviation of the vehicle occupancy data on different road sections. The second function is the negative sum of the scaled values of carbon dioxide emissions.
The termination of training depends on the training iteration and simulation. When the training iteration reaches the preset value and the simulation time reaches the predefined value, the connection between the MAPPO framework and SUMO simulation is closed.
The reset function is responsible for resetting the environment to its initial state before the start of a new episode. It sets the state of each agent in the environment to its initial state.
The step function executes a single step in the environment for each PPO agent and collects the resulting observations, rewards, and speed limits. The step function receives the actions chosen by each individual PPO agent and applies them to the environment. After applying the actions, the step function collects the current state of the environment, which is used as input for the agent’s policy update. The step function calculates the rewards obtained by each agent based on the predefined reward function. The step function determines if the current state of the environment is a terminal state, indicating the end of the episode. The step function also collects the speed limits chosen by the agents, which can be considered the control input and can be used for further analysis.

4. Results and Discussion

In the sequel, the simulation results are introduced in detail, and relevant discussion is provided as well.

4.1. Training Setting of MAPPO Algorithm

To select the appropriate variables for the training process, hyperparameters need to be properly tuned. The set of MAPPO hyperparameters utilized in this paper is shown in Table 1.
The training performance is verified by the reward curve through 600 iterations. In each iteration, a random seed is applied for the proposed control strategies to keep the diversity. Seven PPO agents are used in this paper, and each agent controls a road section; they work cooperatively to minimize the reward function. The GAE λ parameter represents the parameter used in GAE methods that controls the trade-off between bias and variance in the advantage estimation. The number of PPO epochs per update determines how many times the network parameters are updated based on the collected data during a single training iteration. The number of hidden layers in the neural network is 3, the number of neurons in the hidden layer is 64, and the activation function of the hidden layer is the ReLu function, which helps with the vanishing gradient problem, allows the network to learn non-linear relationships, and provides sparse activation.
Traffic performance and sustainability measurements are taken in a dynamic traffic environment based on the control output obtained by different reward settings. The traffic performance measurement, including the occupancy data of each road section and its distribution, shows the effect of the proposed control framework in smoothing the traffic. The sustainability measurements are presented by the reduction in carbon dioxide emissions.

4.2. Traffic Performance Measurements

Figure 4 shows the learning curve over 600 iterations, utilizing the negative standard deviation of the vehicle occupancy of different road sections as the reward function. Min-Max normalization is performed on the obtained rewards. The reward value changes significantly when the number of training iteration reaches about 76. Although there are some oscillations, the reward value is generally stable in the subsequent training process.
As Figure 5 shows, the MAPPO-based VSL controller starts to reduce the speed limit of all controlled sections at step 25. The speed limit is stable between 8.3 m/s and 13.89 m/s after 104 steps. When traffic congestion is about to be formed at the bottleneck area, the MAPPO-based VSL controller reduces the speed limit of the controlled road sections quickly to 13.89 m/s to prevent capacity drop at the bottleneck area. In this paper, we assume that drivers’ acceptance of the speed limit is 100%. The proposed speed limit can be spread to the road network in a traditional way, i.e., variable speed limit signs. With the connected and automated vehicles (CAVs) and intelligent transportation infrastructure on the road network, the proposed speed limit can be synchronously received by CAV with vehicle-to-vehicle (V2V) communication technology.
The changes in the occupancy data of different road sections during the simulation period without VSL control are shown in Figure 6. The first 300 s is the warming-up time. The occupancy change shows that from 2400 s to the end of the simulation, the occupancy of the motorway sections before merging with the ramp varies between 36% and 68%. With the proposed MAPPO-based VSL control, as shown in Figure 7, the occupancy of the motorway sections before merging with the ramp and “section_f” reaches almost the same value of 46% at the end of the simulation. The occupancy of “section_e” and “section_g” reaches the same value of 30%. From this, a conclusion can be made that the controller truly achieves the predefined goal, which is that by controlling the speed limit of each road section, the vehicles are evenly distributed on each road section.
Figure 8 intuitively demonstrates the distribution of vehicles on the road network without VSL control. It shows that vehicles queued at “section_a”, “section_b”, “section_c”, and “section_e”. This was caused by merging with the ramp and the backpropagation of congestion occurring in the urban area. After implementing MARL-based VSL control, as shown in Figure 9, vehicles began distributing uniformly and concentratedly on all motorway sections. This is because the PPO agents started to reduce the upstream speed limit to obtain the optimal reward defined by the agent policy. The interquartile range of the box plot shows how the data fluctuate. Compared with the non-implementation of the VSL control case, occupancy fluctuations in the upstream section of the motorway are significantly smaller when applying the MARL-based VSL control.

4.3. Sustainability Measurements

Figure 10 shows the learning curve over 600 iterations utilizing the negative sum of the scaled carbon dioxide emissions as the reward function. As defined in Section 3, the MAPPO-based controller trends to minimize carbon dioxide emissions over the road network.
As Figure 11 shows, the MAPPO-based VSL controller started to reduce the speed limit of all controlled sections a bit later than the controller dedicated to smoothing the traffic flow. The final speed limit chosen by the agents is 8.3 m/s for all sections. This can be explained by the innate character of the SUMO emission model. SUMO utilizes the third version of the Handbook Emission Factors for Road Transport (HBEFA 3) [53] model to calculate emissions. The homogeneous low-speed traffic flow keeps low pollutant emissions. So the agents choose the same relatively low-speed limit for all road sections.
Figure 12 shows the comparison of the total carbon dioxide emissions in kilograms without VSL control, with only single-section VSL control, and with multi-section VSL control over the simulation time with the first 300 s of warming up time. The figure verifies the performance of the multi-agent RL-based VSL controller, showing that the proposed controller performs well in terms of the sustainability metrics. Specifically, compared with the no-control case, the proposed VSL control strategy reduces the total carbon dioxide emissions of the road network by 11.2%. By only controlling the upstream section close to the bottleneck, the total carbon dioxide emissions are just reduced 0.41%. This means that compared with only adjusting the speed limit of the upstream section close to the bottleneck, our MARL-based VSL controller reduces the total carbon dioxide emissions by 10.79%.
Figure 13 shows the comparison of the total waiting time in seconds without VSL control, with only single-section VSL control, and with multi-section VSL control. Compared with the single-section control case, the proposed MARL-based VSL control strategy further reduces the total waiting time through the network by 15.8%. The oscillations in the total waiting time are significantly smaller than in the uncontrolled and single-section control cases. This shows that the waiting times for vehicles tend to be similar throughout the simulation time, which demonstrate a reduction in the stop-and-go phenomenon of the vehicles.
Figure 12 and Figure 13 demonstrate the superiority of our control strategy compared to the single-section VSL control. Due to the high synergism of the traffic system, single-section VSL control will lead vehicles to decelerate suddenly before the bottleneck area. Meanwhile, vehicles on the upstream section continuously drive into the bottleneck area at high speed. High-speed cruising and low-speed congestion contribute to high road network carbon dioxide emissions. The multi-section VSL control divides the upstream of the bottleneck area into multiple deceleration sections. The control idea is to induce and control the upstream traffic flow to prevent the upstream traffic flow from accumulating quickly in the bottleneck area so that the vehicles can pass through the bottleneck area in an orderly manner. The reduction in waiting time confirms the feasibility of this control idea.

5. Conclusions

Different from the previous research, which only performs reinforcement learning-based variable speed limit control on the upstream section close to the bottleneck area, this paper proposes a multi-agent proximal policy optimization strategy for a multi-section variable speed limit control, which implements individual agents on all motorway sections to smooth traffic flow and reduce traffic emissions. A real-world traffic network is generated in the SUMO simulator based on the on-site historical measurement data, representing a recurrent bottleneck area when motorways transit to urban roads. The results unequivocally show improvements in the overall vehicle distribution, the total waiting time reduction, the carbon dioxide emissions reduction over the road network and how the agents apply speed limits that lead to this goal. A more homogeneous traffic flow is formed after applying the proposed MARL-based VSL control. Compared to the existing RL-based control method, which only controls the speed limit of the upstream section close to the bottleneck, our control idea outperforms. In the case of traffic performance, the total waiting time through the network is reduced by 15.8%. Regarding sustainability measurement, the total carbon dioxide emissions over the network are reduced by 10.79%.
In conclusion, the paper investigates the effect of two different controllers independently in a credible way. Limited by the iteration speed of MAPPO algorithms and the running speed of the microscopic traffic simulation, the proposed MARL-based VSL control framework is unable to realize real-time on-site traffic management. As a future work, the proposed MAPPO-based VSL control framework will be extended to multi-objective optimization to balance traffic performance and emissions by adjusting the weights of multiple reward designs on the designated road network.

Author Contributions

Conceptualization, X.F. and T.T.; methodology, X.F. and T.T.; case study, X.F.; validation, X.F.; writing—original draft preparation, X.F.; writing—review and editing, X.F., T.P. and T.T.; supervision, T.T.; project administration, T.T. and T.P. All authors have read and agreed to the published version of the manuscript.

Funding

Project no. TKP2021-NVA-02 was implemented with the support provided by the Ministry of Culture and Innovation of Hungary from the National Research, Development and Innovation Fund, financed under the TKP2021-NVA funding scheme. The research was supported by the European Union within the framework of the National Laboratory for Autonomous Systems (RRF-2.3.1-21-2022-00002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ITSIntelligent Transportation System
VSLVariable Speed Limit
DRLDeep Reinforcement Learning
DQNDeep Q-Network
MAPPOMulti-Agent Proximal Policy Optimization
SUMOSimulation of Urban Mobility
RLReinforcement Learning
MLMachine Learning
DLDeep Learning
DNNDeep Neural Networks
NNNeural Networks
PGPolicy Gradient
ACActor–Critic
A3CAdvantage Actor–Critic
TRPOTrust Region Policy Optimization
PPOProximal Policy Optimization
MARLMulti-Agent Reinforcement Learning
CTDECentralized Training with Decentralized Execution
GAEGeneralized Advantage Estimation
TraCITraffic Control Interface
OSMOpenStreetMap
TTTTotal Time Spent
CAVConnected and Automated Vehicles
V2VVehicle to Vehicle
HBEFA 3Third Version of Handbook Emission Factors for Road Transport

References

  1. Wang, C.; Zhang, J.; Xu, L.; Li, L.; Ran, B. A new solution for freeway congestion: Cooperative speed limit control using distributed reinforcement learning. IEEE Access 2019, 7, 41947–41957. [Google Scholar] [CrossRef]
  2. Szele, A.; Kisgyörgy, L. Traffic operation on a road network with recurrent congestion. In Proceedings of the WIT Transactions on The Built Environment, Rome, Italy, 10–12 October 2022; WIT Press: Southampton, UK, 2018. [Google Scholar] [CrossRef] [Green Version]
  3. Tang, L.; Wang, Y.; Zhang, X. Identifying recurring bottlenecks on urban expressway using a fusion method based on loop detector data. Math. Probl. Eng. 2019, 2019, 5861414. [Google Scholar] [CrossRef] [Green Version]
  4. Den Hertog, D.; Van Zante-De Fokkert, J.; Sjamaar, S.; Beusmans, R. Optimal working zone division for safe track maintenance in The Netherlands. Accid. Anal. Prev. 2005, 37, 890–893. [Google Scholar] [CrossRef] [PubMed]
  5. Kerner, B.S. Features of Traffic Congestion caused by bad Weather Conditions or Accident. arXiv 2007, arXiv:0712.1728. [Google Scholar]
  6. Zeng, J.; Qian, Y.; Lv, Z.; Yin, F.; Zhu, L.; Zhang, Y.; Xu, D. Expressway traffic flow under the combined bottleneck of accident and on-ramp in framework of Kerner’s three-phase traffic theory. Phys. Stat. Mech. Its Appl. 2021, 574, 125918. [Google Scholar] [CrossRef]
  7. Li, H.; Zhang, J.; Zhang, Z.; Huang, Z. Active lane management for intelligent connected vehicles in weaving areas of urban expressway. J. Intell. Connect. Veh. 2021, 4, 52–67. [Google Scholar] [CrossRef]
  8. Jithender, J.; Mehar, A. Estimation of Queue Length at Signalized Intersection Under Non-Lane Based Traffic Conditions. Periodica Polytechnica Transp. Eng. 2023, 51, 31–39. [Google Scholar] [CrossRef]
  9. Péter, T.; Lakatos, I.; Pup, D.; Szauter, F.; Szabó, K. An integrated analysis of processes concerning traffic and vehicle dynamics, development of laboratory applying real traffic scenarios. In Proceedings of the 2017 ASME/IEEE International Conference on Mechatronic and Embedded Systems and Applications (MESA2017), Cleveland, OH, USA, 6–9 August 2017. [Google Scholar]
  10. Lakatos, I. Diagnostic measurement for the effective performance of motor vehicles. Acta Polytech. Hung. 2013, 10, 239–249. [Google Scholar] [CrossRef]
  11. Coppola, A.; Di Costanzo, L.; Pariota, L.; Bifulco, G.N. Fuzzy-based Variable Speed Limits System under Connected Vehicle environment: A Simulation-Based Case Study in the city of Naples. IEEE Open J. Intell. Transp. Syst. 2023, 4, 267–278. [Google Scholar] [CrossRef]
  12. Li, Z.; Liu, P.; Xu, C.; Duan, H.; Wang, W. Reinforcement learning-based variable speed limit control strategy to reduce traffic congestion at freeway recurrent bottlenecks. IEEE Trans. Intell. Transp. Syst. 2017, 18, 3204–3217. [Google Scholar] [CrossRef]
  13. Wang, X.; Zhang, R.; Gou, Y.; Liu, J.; Zhao, L.; Li, Y. Variable speed limit control method of freeway mainline in intelligent connected environment. J. Adv. Transp. 2021, 2021, 8863487. [Google Scholar] [CrossRef]
  14. Zegeye, S.K.; De Schutter, B.; Hellendoorn, J.; Breunesse, E.A. Variable speed limits for area-wide reduction of emissions. In Proceedings of the 13th International IEEE Conference on Intelligent Transportation Systems, Funchal, Madeira, Portugal, 19–22 September 2010; pp. 507–512. [Google Scholar]
  15. Chen, R.; Zhang, T.; Levin, M.W. Effects of variable speed limit on energy consumption with autonomous vehicles on urban roads using modified cell-transmission model. J. Transp. Eng. Part A Syst. 2020, 146, 04020049. [Google Scholar] [CrossRef]
  16. Li, Z.; Xu, C.; Guo, Y.; Liu, P.; Pu, Z. Reinforcement learning-based variable speed limits control to reduce crash risks near traffic oscillations on freeways. IEEE Intell. Transp. Syst. Mag. 2020, 13, 64–70. [Google Scholar] [CrossRef]
  17. Allaby, P.; Hellinga, B.; Bullock, M. Variable speed limits: Safety and operational impacts of a candidate control strategy for freeway applications. IEEE Trans. Intell. Transp. Syst. 2007, 8, 671–680. [Google Scholar] [CrossRef]
  18. Li, D.; Ranjitkar, P. A fuzzy logic-based variable speed limit controller. J. Adv. Transp. 2015, 49, 913–927. [Google Scholar] [CrossRef]
  19. Vrbanić, F.; Ivanjko, E.; Kušić, K.; Čakija, D. Variable speed limit and ramp metering for mixed traffic flows: A review and open questions. Appl. Sci. 2021, 11, 2574. [Google Scholar] [CrossRef]
  20. Liu, H.; Zhang, L.; Sun, D.; Wang, D. Optimize the settings of variable speed limit system to improve the performance of freeway traffic. IEEE Trans. Intell. Transp. Syst. 2015, 16, 3249–3257. [Google Scholar] [CrossRef]
  21. Li, Y.; Chow, A.H.; Cassel, D.L. Optimal control of motorways by ramp metering, variable speed limits, and hard-shoulder running. Transp. Res. Rec. 2014, 2470, 122–130. [Google Scholar] [CrossRef]
  22. Yuan, N.; Ma, M.; Liang, S.; Wang, W.; Zhang, H. Optimal control method of freeway based on tollbooths lane configuration and variable speed limit control. Phys. Stat. Mech. Its Appl. 2022, 603, 127801. [Google Scholar] [CrossRef]
  23. Carlson, R.C.; Papamichail, I.; Papageorgiou, M.; Messmer, A. Optimal mainstream traffic flow control of large-scale motorway networks. Transp. Res. Part Emerg. Technol. 2010, 18, 193–212. [Google Scholar] [CrossRef]
  24. Yang, X.; Lu, Y.; Chang, G. Proactive optimal variable speed limit control for recurrently congested freeway bottlenecks. In Proceedings of the Transportation Research Board Meeting, Washington, DC, USA, 13–17 January 2013. [Google Scholar]
  25. Wang, S.; Stern, R.; Levin, M.W. Optimal control of autonomous vehicles for traffic smoothing. IEEE Trans. Intell. Transp. Syst. 2021, 23, 3842–3852. [Google Scholar] [CrossRef]
  26. Frejo, J.R.D.; Núnez, A.; De Schutter, B.; Camacho, E.F. Hybrid model predictive control for freeway traffic using discrete speed limit signals. Transp. Res. Part Emerg. Technol. 2014, 46, 309–325. [Google Scholar] [CrossRef]
  27. Carlson, R.C.; Papamichail, I.; Papageorgiou, M. Local feedback-based mainstream traffic flow control on motorways using variable speed limits. IEEE Trans. Intell. Transp. Syst. 2011, 12, 1261–1276. [Google Scholar] [CrossRef]
  28. Lu, X.Y.; Shladover, S. MPC-based variable speed limit and its impact on traffic with V2I type ACC. In Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 3923–3928. [Google Scholar]
  29. Kušić, K.; Dusparic, I.; Guériau, M.; Gregurić, M.; Ivanjko, E. Extended variable speed limit control using multi-agent reinforcement learning. In Proceedings of the 2020 IEEE 23rd International Conference on Intelligent Transportation Systems (ITSC), Rhodes, Greece, 20–23 September 2020; pp. 1–8. [Google Scholar]
  30. Zhu, F.; Ukkusuri, S.V. Accounting for dynamic speed limit control in a stochastic traffic environment: A reinforcement learning approach. Transp. Res. Part C Emerg. Technol. 2014, 41, 30–47. [Google Scholar] [CrossRef]
  31. Cheng, M.; Zhang, C.; Jin, H.; Wang, Z.; Yang, X. Adaptive Coordinated Variable Speed Limit between Highway Mainline and On-Ramp with Deep Reinforcement Learning. J. Adv. Transp. 2022. [Google Scholar] [CrossRef]
  32. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef]
  33. Yu, C.; Velu, A.; Vinitsky, E.; Wang, Y.; Bayen, A.; Wu, Y. The Surprising Effectiveness of PPO in Cooperative, Multi-Agent Games. arXiv 2021, arXiv:2103.01955. [Google Scholar]
  34. Alagha, A.; Singh, S.; Mizouni, R.; Bentahar, J.; Otrok, H. Target localization using multi-agent deep reinforcement learning with proximal policy optimization. Future Gener. Comput. Syst. 2022, 136, 342–357. [Google Scholar] [CrossRef]
  35. Lohse, O.; Pütz, N.; Hörmann, K. Implementing an Online Scheduling Approach for Production with Multi Agent Proximal Policy Optimization (MAPPO). In Proceedings of the Advances in Production Management Systems, Artificial Intelligence for Sustainable and Resilient Production Systems: IFIP WG 5.7 International Conference, APMS 2021, Nantes, France, 5–9 September 2021; Proceedings Part V.. Springer: Berlin/Heidelberg, Germany, 2021; pp. 586–595. [Google Scholar]
  36. Park, Y.M.; Hassan, S.S.; Tun, Y.K.; Han, Z.; Hong, C.S. Joint Trajectory and Resource Optimization of MEC-Assisted UAVs in Sub-THz Networks: A Resources-based Multi-Agent Proximal Policy Optimization DRL with Attention Mechanism. arXiv 2022, arXiv:2209.07228. [Google Scholar]
  37. Lopez, P.A.; Behrisch, M.; Bieker-Walz, L.; Erdmann, J.; Flötteröd, Y.P.; Hilbrich, R.; Lücken, L.; Rummel, J.; Wagner, P.; Wießner, E. Microscopic Traffic Simulation using SUMO. In Proceedings of the 21st IEEE International Conference on Intelligent Transportation Systems, Maui, HI, USA, 4–7 November 2018. [Google Scholar]
  38. Wiering, M.A.; Van Otterlo, M. Reinforcement learning. Adapt. Learn. Optim. 2012, 12, 729. [Google Scholar]
  39. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  40. Sutton, R.S.; McAllester, D.; Singh, S.; Mansour, Y. Policy gradient methods for reinforcement learning with function approximation. Adv. Neural Inf. Process. Syst. 1999, 12, 1057–1063. [Google Scholar]
  41. Konda, V.; Tsitsiklis, J. Actor-critic algorithms. Adv. Neural Inf. Process. Syst. 1999, 12, 1008–1014. [Google Scholar]
  42. Mnih, V.; Badia, A.P.; Mirza, M.; Graves, A.; Lillicrap, T.; Harley, T.; Silver, D.; Kavukcuoglu, K. Asynchronous methods for deep reinforcement learning. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 19–24 June 2016; pp. 1928–1937. [Google Scholar]
  43. Schulman, J.; Levine, S.; Abbeel, P.; Jordan, M.; Moritz, P. Trust region policy optimization. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 1889–1897. [Google Scholar]
  44. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal policy optimization algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
  45. De Witt, C.S.; Peng, B.; Kamienny, P.A.; Torr, P.; Böhmer, W.; Whiteson, S. Deep multi-agent reinforcement learning for decentralized continuous cooperative control. arXiv 2020, arXiv:2003.06709. [Google Scholar]
  46. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M. Playing atari with deep reinforcement learning. arXiv 2013, arXiv:1312.5602. [Google Scholar]
  47. Li, Y.; Ni, P.; Chang, V. Application of deep reinforcement learning in stock trading strategies and stock forecasting. Computing 2020, 102, 1305–1322. [Google Scholar] [CrossRef]
  48. Kušić, K.; Ivanjko, E.; Gregurić, M. A comparison of different state representations for reinforcement learning based variable speed limit control. In Proceedings of the 26th Mediterranean Conference on Control and Automation (MED), Zadar, Croatia, 19–22 June 2018; pp. 1–6. [Google Scholar]
  49. Wu, Y.; Tan, H.; Jiang, Z.; Ran, B. ES-CTC: A deep neuroevolution model for cooperative intelligent freeway traffic control. arXiv 2019, arXiv:1905.04083. [Google Scholar]
  50. Schaul, T.; Quan, J.; Antonoglou, I.; Silver, D. Prioritized experience replay. arXiv 2015, arXiv:1511.05952. [Google Scholar]
  51. Xu, Z.; Cao, Y.; Kang, Y.; Zhao, Z. Vehicle emission control on road with temporal traffic information using deep reinforcement learning. IFAC-PapersOnLine 2020, 53, 14960–14965. [Google Scholar] [CrossRef]
  52. Wu, Y.; Tan, H.; Qin, L.; Ran, B. Differential variable speed limits control for freeway recurrent bottlenecks via deep actor-critic algorithm. Transp. Res. Part Emerg. Technol. 2020, 117, 102649. [Google Scholar] [CrossRef]
  53. Keller, M.; Hausberger, S.; Matzer, C.; Wüthrich, P.; Notter, B. HBEFA Version 3.3. Backgr. Doc. Berne 2017, 12. Available online: https://www.google.com.hk/search?q=Keller%2C+M.%3B+Hausberger%2C+S.%3B+Matzer%2C+C.%3B+Wuthrich%2C+P.%3B+Notter%2C+B.+HBEFA+Version+3.3&ei=YzC6ZI30DOfM2roP3f6piAs&ved=0ahUKEwjNsdemoJ-AAxVnplYBHV1_CrEQ4dUDCA4&uact=5&oq=Keller%2C+M.%3B+Hausberger%2C+S.%3B+Matzer%2C+C.%3B+Wuthrich%2C+P.%3B+Notter%2C+B.+HBEFA+Version+3.3&gs_lp=Egxnd3Mtd2l6LXNlcnAiUktlbGxlciwgTS47IEhhdXNiZXJnZXIsIFMuOyBNYXR6ZXIsIEMuOyBXdXRocmljaCwgUC47IE5vdHRlciwgQi4gSEJFRkEgVmVyc2lvbiAzLjMyBxAAGB4YsANIv0VQkQJYxztwAXgAkAEAmAEAoAEAqgEAuAEDyAEA-AEC-AEB4gMEGAEgQYgGAZAGAQ&sclient=gws-wiz-serp (accessed on 30 May 2023).
Figure 1. PPO agent optimization framework.
Figure 1. PPO agent optimization framework.
Sustainability 15 11464 g001
Figure 2. Schematic diagram of variable speed limit control at the merging area, where the motorway transitions to the urban road network.
Figure 2. Schematic diagram of variable speed limit control at the merging area, where the motorway transitions to the urban road network.
Sustainability 15 11464 g002
Figure 3. Geometric structure of simulated recurrent bottleneck area.
Figure 3. Geometric structure of simulated recurrent bottleneck area.
Sustainability 15 11464 g003
Figure 4. The learning curve dedicated to smooth traffic flow over the 600 iterations.
Figure 4. The learning curve dedicated to smooth traffic flow over the 600 iterations.
Sustainability 15 11464 g004
Figure 5. The speed limit of different road sections with smooth traffic as the control goal.
Figure 5. The speed limit of different road sections with smooth traffic as the control goal.
Sustainability 15 11464 g005
Figure 6. Occupancy data of different road sections during the simulation period without VSL control.
Figure 6. Occupancy data of different road sections during the simulation period without VSL control.
Sustainability 15 11464 g006
Figure 7. Occupancy data of different road sections during the simulation period with VSL control.
Figure 7. Occupancy data of different road sections during the simulation period with VSL control.
Sustainability 15 11464 g007
Figure 8. Distribution of vehicles occupancy data on each road section without VSL control.
Figure 8. Distribution of vehicles occupancy data on each road section without VSL control.
Sustainability 15 11464 g008
Figure 9. Distribution of vehicles occupancy data on each road section with VSL control.
Figure 9. Distribution of vehicles occupancy data on each road section with VSL control.
Sustainability 15 11464 g009
Figure 10. The learning curve dedicated to reducing carbon dioxide emissions over the 600 iterations.
Figure 10. The learning curve dedicated to reducing carbon dioxide emissions over the 600 iterations.
Sustainability 15 11464 g010
Figure 11. The speed limit of different road sections with reducing emissions as a control goal.
Figure 11. The speed limit of different road sections with reducing emissions as a control goal.
Sustainability 15 11464 g011
Figure 12. Total carbon dioxide emissions over the network with and without VSL control.
Figure 12. Total carbon dioxide emissions over the network with and without VSL control.
Sustainability 15 11464 g012
Figure 13. Total waiting time with and without VSL control.
Figure 13. Total waiting time with and without VSL control.
Sustainability 15 11464 g013
Table 1. Multi-agent proximal policy optimization hyperparameters for training.
Table 1. Multi-agent proximal policy optimization hyperparameters for training.
HyperparametersValue
Number of training iterations600
Learning rate0.0005
Number of agent7
PPO clip parameter θ 0.2
Discount factor γ 0.99
GAE λ parameter0.95
Time step per update120
Number of PPO epochs per update15
Hidden layers64 × 64 × 64
Hidden layers activation functionRELU
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fang, X.; Péter, T.; Tettamanti, T. Variable Speed Limit Control for the Motorway–Urban Merging Bottlenecks Using Multi-Agent Reinforcement Learning. Sustainability 2023, 15, 11464. https://doi.org/10.3390/su151411464

AMA Style

Fang X, Péter T, Tettamanti T. Variable Speed Limit Control for the Motorway–Urban Merging Bottlenecks Using Multi-Agent Reinforcement Learning. Sustainability. 2023; 15(14):11464. https://doi.org/10.3390/su151411464

Chicago/Turabian Style

Fang, Xuan, Tamás Péter, and Tamás Tettamanti. 2023. "Variable Speed Limit Control for the Motorway–Urban Merging Bottlenecks Using Multi-Agent Reinforcement Learning" Sustainability 15, no. 14: 11464. https://doi.org/10.3390/su151411464

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop