Next Article in Journal
Some Considerations about the Anodic Limit of Ionic Liquids Obtained by Means of DFT Calculations
Previous Article in Journal
Decentralized Stochastic Control with Finite-Dimensional Memories: A Memory Limitation Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Agent Deep Reinforcement Learning-Based Popular Content Distribution Scheme in Vehicular Networks

1
College of Information Science and Technology, Jinan University, Guangzhou 510632, China
2
Guangdong Key Laboratory of Data Security and Privacy Preserving, Guangzhou 511443, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(5), 792; https://doi.org/10.3390/e25050792
Submission received: 30 March 2023 / Revised: 8 May 2023 / Accepted: 11 May 2023 / Published: 12 May 2023
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
The Internet of Vehicles (IoV) enables vehicular data services and applications through vehicle-to-everything (V2X) communications. One of the key services provided by IoV is popular content distribution (PCD), which aims to quickly deliver popular content that most vehicles request. However, it is challenging for vehicles to receive the complete popular content from roadside units (RSUs) due to their mobility and the RSUs’ constrained coverage. The collaboration of vehicles via vehicle-to-vehicle (V2V) communications is an effective solution to assist more vehicles to obtain the entire popular content at a lower time cost. To this end, we propose a multi-agent deep reinforcement learning (MADRL)-based popular content distribution scheme in vehicular networks, where each vehicle deploys an MADRL agent that learns to choose the appropriate data transmission policy. To reduce the complexity of the MADRL-based algorithm, a vehicle clustering algorithm based on spectral clustering is provided to divide all vehicles in the V2V phase into groups, so that only vehicles within the same group exchange data. Then the multi-agent proximal policy optimization (MAPPO) algorithm is used to train the agent. We introduce the self-attention mechanism when constructing the neural network for the MADRL to help the agent accurately represent the environment and make decisions. Furthermore, the invalid action masking technique is utilized to prevent the agent from taking invalid actions, accelerating the training process of the agent. Finally, experimental results are shown and a comprehensive comparison is provided, which demonstrates that our MADRL-PCD scheme outperforms both the coalition game-based scheme and the greedy strategy-based scheme, achieving a higher PCD efficiency and a lower transmission delay.

1. Introduction

The Internet of Vehicles (IoV) enables data exchange among moving vehicles on the road and fixed RSUs [1,2]. In the past two decades, data dissemination in IoV has become an important research topic because it can improve traffic safety as well as travel comfort [3,4]. There are mainly two types of data for dissemination in IoV: safety-related data and infotainment data. Safety-related data is generally real-time traffic information that should be processed quickly and reliably among vehicles. All vehicles on the road are interested in safety-related data [5]. Infotainment data usually refers to the data that helps to enhance the travel experience and improve travel comfort for passengers [6]. Applications of infotainment data can be classified into two categories: the personalized service that provides tailored content to different vehicles based on specific needs, and the popular content service that delivers data that most vehicles are interested in when located in a specific area or at a specific time. The popular content service requires both low latency and high distribution to ensure quick and efficient delivery to as many passing vehicles as possible. Since safety-related data also requires low latency and high distribution, it can be seen as a special case of popular content. In general, popular content refers to data that is of interest or mandatory for most vehicles, such as popular videos, traffic emergency notifications, and local travel advertisements [7].
In PCD, RSUs distribute popular content to vehicles within their coverage areas, also known as the area of interest (AoI). However, due to the high mobility of vehicles and the limited coverage of RSUs, vehicles may not receive complete content from a single RSU. Moreover, the deployment cost of RSUs is too high to cover all areas, so the number of RSUs available to vehicles is limited. Therefore, collaboration among vehicles via V2V communication is necessary to exchange partial content when they are out of the RSU coverage. However, V2V communication also poses some challenges for PCD. First, wireless communication between vehicles is prone to interference, signal attenuation, and even conflicts. Second, the topology of the vehicular network changes constantly, resulting in unstable wireless links between vehicles.

1.1. Related Work

Push-based information dissemination, which can be seen as a special case of PCD, has been investigated. Schwartz et al. [8] proposed the push-based sample and robust dissemination (SRD) protocol by an efficient periodic hello message mechanism that provided limited local topology information for the protocol. The SRD protocol can prevent the problem of broadcast storms in dense networks by an optimized broadcast suppression technique. Di Felice et al. [9] proposed the dynamic backbone-assisted MAC protocol that constructed a virtual backbone of vehicles in IoV and prevented contention using a new forwarding method. Kumar and Lee [10] proposed the P2P cooperative caching scheme, where vehicles request content from neighbours by P2P, while only contacting the RSU if no neighbours could respond. d’Orey et al. [11] proposed a method called neighbour-aware virtual infrastructure, which extended the RSU coverage virtually by selecting some vehicles as a virtual infrastructure based on network load or application requirements. Messages were broadcasted based on these virtual infrastructures. Nawaz Ali et al. [12] proposed the enhanced cooperative load-balancing system to achieve data dissemination through multiple relaying RSUs. However, these works assume small data sizes and do not address the challenges of efficiently delivering large structured content. In addition, the scheme proposed by Nawaz Ali et al. [12] entails a high deployment cost of RSUs.
Delay-tolerant data dissemination, which is an important application in IoV but different from the PCD, has also been usually investigated by adopting clustering and broadcasting approaches. In 2019, Shafi and Venkata Ratnam [13] proposed a cross layer cluster-based routing approach for efficient multimedia data dissemination with improved reliability in IoV. The approach used information from the physical and network layers to dynamically select suitable relay nodes and routing paths to reduce packet loss and delay. In 2022, Shafi and Venkata Ratnam [14] further proposed a different cross layer cluster by the ant colony optimization called ESBCA, for multimedia data dissemination in IoV. The ESBCA approach used different parameters for cluster head selection, such as relative velocity, propagation range, travel time, node density, etc., to improve route stability and data dissemination. Sun et al. [15] proposed a collaborative multicast scheme that combined multicast with D2D-assisted relay for transmitting multimedia data in vehicular networks, and designed a communication quality metric and a relay selection algorithm to optimize data transmission. Jeevitha and Bhuvaneswari [16] proposed a cluster-based cooperative multicast scheme for efficient multimedia data dissemination in hybrid cellular–D2D vehicular networks. The scheme first clustered the multicast group members using cellular signalling, then selected the cooperative forwarding set using graph theory algorithms, and finally used D2D communication to achieve both intra-cluster and inter-cluster multicast. Hasson and Abbas [17] proposed a clustering algorithm, which used a variable range of cluster heads to find cluster heads for each new vehicle. The clustering algorithm helped to make the network more stable with a high delivery rate, high coverage, and low information loss rate. However, these approaches are not suitable for scenarios with long transmission durations due to the dynamic topology of vehicular networks.
Popular content distribution has attracted many researchers due to the fact that popular contents are often not small-scale and are usually of interest to most of the vehicles on the road. One main-stream method for PCD is on the basis of the coalition game theory. In 2013, Wang et al. [18] proposed a PCD scheme based on coalition formation games, in which vehicles were divided into subnetworks, then a greedy algorithm was used to select the best broadcaster in each subnetwork. In 2018, Hu et al. [19] proposed a cooperative transmission scheme based on coalition games, which dynamically clustered vehicles into different groups according to their locations, and then used game theory to find the optimal transmission network structure within each group, so that vehicles could share data blocks and achieve efficient dissemination of popular content. When the vehicle movement caused location changes, the scheme could adaptively update the clusters and transmission network structure. In 2019, Chen et al. [20] proposed a cooperative V2V transmission scheme based on coalition graph games, called CVCG, where a unique utility function was well designed and a dynamic location-based partitioning scheme of vehicles was provided to reduce algorithm complexity and improve the scalability. In 2020, Wang et al. [21] proposed a full-duplex communication scheme based on coalition games, which improved the content transmission efficiency and fairness through cooperation and competition among vehicles, and designed a distributed algorithm that enabled vehicles to autonomously choose suitable coalitions according to their own cache, channel, and demand. In 2022, Zhang et al. [22] proposed an edge vehicle computing collaborative content dissemination strategy based on fuzzy logic and coalition graph games, which used V2V communication to assist RSUs in distributing content, adjusted the relay vehicle ratio according to vehicle density, and improved the content dissemination range and efficiency. However, coalition formation game-based schemes require multiple games between vehicles, inevitably spending a lot of time on reaching some equilibrium.
Different from coalition game-based methods, researchers have used other schemes to improve the efficiency of PCD. Li et al. [7] proposed a cooperative popular content distribution scheme based on the symbol level network coding, called CodeOn, which could effectively improve the content downloading speed and reliability in vehicular networks. In this scheme, the server encoded popular content into symbols and transmitted them to vehicles. Vehicles obtained symbols from the server or other vehicles and stored them in their caches. Vehicles also broadcasted symbols according to their own and their neighbours’ cache statuses for cooperative distribution. Once vehicles had collected enough symbols, they could decode them back into content. Huang and Wang [23] proposed a collaborative downloading scheme for popular content dissemination in urban vehicular networks, which had three innovations: a cell-based clustering strategy, a topology pre-creation and update scheme, and a cross-cluster relay selection and generation selection strategy. Chen et al. [24] proposed a popular content distribution scheme based on the group acknowledgment strategy. The scheme divided popular content into multiple segments and selected appropriate forwarding nodes and retransmission strategies according to the dynamic information of vehicles, to improve data transmission efficiency and reliability in vehicular networks.
With the development of AI technology, machine learning, especially deep reinforcement learning (DRL), has been applied in various fields, which can provide autonomous and adaptive control in vehicular applications [25]. Different reinforcement learning-based data dissemination schemes have been designed for vehicular networks (e.g., see references [26,27,28,29,30]). In 2018, Wu et al. [26] proposed a multi-tier multi-access edge clustering architecture that utilized some mobile vehicles as edges and generated different levels of clusters for integrating multiple types of wireless communication technologies, such as cellular communications, mmWave, and IEEE 802.11p. This clustering was conducted by a fuzzy-logic-based approach, where cluster head nodes were selected (by taking into account the vehicle velocity, network topology, and antenna height) and responsible for providing gateway capabilities between different types of communications. Finally, combined with the Q-learning algorithm, spatial intelligence could be achieved. Similar to [26], in [27], the multi-access vehicular environment with different types of wireless communication technologies was considered, and the idea of choosing some mobile vehicles as edges was also employed to propose a reinforcement learning-based end-edge-cloud collaboration routing scheme to find routes in a proactive manner with a low communication overhead. Mchergui et al. [28] proposed a vehicular network broadcasting scheme based on motion parameters, fuzzy system, and deep learning, which could select the optimal relay nodes according to the relay quality index and disseminate data through a multicast mode. Lolai et al. [29] considered data routing in IoV and used a Q-learning reinforcement algorithm to achieve data packet routing between the source and destination by selecting effective road segments and intermediate vehicles based on various parameters. Lou and Hou [30] used DRL to optimize the Markov decision process model and propose a high-definition map dissemination problem that considers the age of information and degree of interest, then solved it with the proximal policy optimization method. Although there are many achievements of DRL-based data dissemination in vehicular networks, DRL schemes for PCD are still very scarce.

1.2. Contributions

In this work, we concentrate on DRL-based vehicular collaborations to assist PCD. However, for multi-vehicle environments, if the traditional DRL is used to control data exchanging behaviours among multiple vehicles, the dimension of the agent’s action will increase exponentially with the number of vehicles. This would cause a large number of invalid actions in the agent’s action space, resulting in a sparse reward problem that affects the training of the agent. To solve the sparse reward problem, we adopt a multi-agent deep reinforcement learning (MADRL) framework. This framework allows each vehicle to correspond to an agent that makes decisions independently, effectively reducing the dimension of the agent’s action. The main contributions of our work are summarized as follows.
  • A multi-agent deep reinforcement learning-based popular content distribution (MADRL-PCD) scheme in vehicular networks is proposed. We divide PCD into two phases. First, vehicles obtain part of the popular content from a few RSUs in the V2R phase. Then, based on the different parts of the content they own, vehicles form a vehicular ad hoc network (VANET) for data sharing to acquire the whole popular content in the V2V phase. We train the agent to coordinate data transmissions in the VANET by multi-agent reinforcement learning. Simulation results show that our scheme outperforms the greedy-based scheme and the game theory-based CVCG scheme with a lower delivery delay and a higher delivery efficiency.
  • To reduce the complexity of the learning algorithm, we propose a vehicle clustering algorithm based on spectral clustering to divide all vehicles in the V2V stage into multiple non-interfering groups, so that data exchange occurs only among vehicles within the same group. Moreover, by clustering, the number of vehicles within each group does not exceed the number of agents, which is helpful to apply the MADRL efficiently. Based on the vehicle partition, we abstract a multi-vehicle environment training and train a fixed group of intelligent agents using the MAPPO algorithm.
  • When applying MADRL, we define the agent’s observation and the environment’s state as matrix forms. Considering that the agent’s decision is based on the correlation of any two vehicles’ information, we use the self-attention mechanism to construct the agent’s neural network. Moreover, we use the invalid action masking technique to avoid unreasonable actions by the agent and accelerate its training process.
The remainder of this article is structured as follows. Section 2 describes the system model for the PCD process in vehicular networks. The channel model of vehicular communications and the algorithm of vehicle clustering are also introduced, and then the PCD problem is proposed in Section 2. Section 3 introduces the proposed MADRL-PCD scheme in vehicular networks in detail, where the training approach and the MADRL network architecture are presented. Section 4 shows the simulation results of the MADRL-PCD scheme, and provides an exhaustive comparison from different viewpoints. Section 5 concludes our work.

2. System Model and Problem Formulation

The system model considered in this work is illustrated in Figure 1. It consists of a straight two-way road equipped with an RSU at each end of the road, where new vehicles enter from both ends of the road and immediately fall within the coverage range of the RSUs. Popular content will be delivered to vehicles passing through this road. Suppose the popular content is divided into M small chunks of equal size. In each time slot, the two RSUs randomly broadcast a chunk of the popular content, which may be received by vehicles within their respective coverage ranges. Each vehicle has a memory that caches the chunks of popular content received from the RSU. As the vehicle moves, it eventually leaves the coverage of the RSU. At this moment, the vehicle may have already cached a certain number of content chunks. We denote the cache of vehicle i by  C i = c i , 0 , c i , 1 , , c i , M 1  , where  c i , j = 1   indicates that vehicle i has cached the j-th chunk of the popular content; otherwise  c i , j = 0  .
Once vehicles travel beyond the coverage range of the RSUs, they enter the V2V phase. We assume that vehicles can work in full-duplex mode, allowing them to send and receive data simultaneously. In order to more effectively control and optimize the data exchange process between vehicles in the V2V phase, we use a spectral clustering-based method to classify vehicles before the V2V phase. This classification method divides the vehicles into B groups, with each group potentially containing a different number of vehicles. The different groups remain independent of each other and do not interfere with each other. That is, vehicles within each group can exchange data to improve the popular content acquisition rate, but vehicles across groups do not exchange data to avoid communication interference by using different communication channels with different spectra. We denote the set of groups as  B = b 0 , b 1 , , b B 1  .
After clustering, each group of vehicles forms a multi-vehicle environment. In this environment, the vehicles in each group share a wireless channel spectrum. We assign each vehicle a DRL agent and use MADRL to train these agents to collaborate efficiently. Each vehicle deploys a pre-trained MADRL agent and uses the information of all vehicles in the same group to determine its optimal target for receiving data. The vehicles then work together to compose a data exchange strategy for the group, which they follow to send a chunk of popular content to any requesting vehicle. After a fixed transmission time, all vehicles enter the next round of V2V collaboration, which involves re-clustering the vehicles, obtaining a new collaboration strategy, and communicating.

2.1. Channel Model

We assume that vehicles and RSUs use dedicated short-range communication in both the V2R and V2V phases. Wireless communication suffers from various fading effects, including large- and small-scale fading. The path loss of large-scale fading is considered and can be characterized by the Friis model [31,32]. The receiving signal power is given as
P r ^ = P t G t G r λ 2 4 π 2 d 2 L s y s
where  P r ^   is the receiving signal power with path loss only,  P t   is the transmitted power,  G t   and  G r   are the transmitter and receiver antenna gain coefficients, respectively,  λ   is the electromagnetic wavelength in free space,  L s y s   is the system loss factor, and d is the distance between the transmitting and receiving antennas. By introducing the notation  k f G t G r λ 2 4 π 2 d 2 L  , Equation (1) can be represented as  P r ^ k f P t  . The small-scale fading effects can be characterized by the Nakagami-m fading model [33,34]. Due to Nakagami-m fading, the receiving signal power follows a Gamma distribution given by
P r G a m m a m , P r ^ m
where  P r ^   is the power of the receiving signal affected by both pass loss fading and Nakagami fading. The parameter m depends on the distance d between the sender and receiver. When the sender and receiver are close, m is large, and the small-scale fading effect is small. Conversely, when they are far apart, m is small and the small-scale fading effect is large, causing the received signal strength to fluctuate more severely. To account for this distance-dependent behaviour, we can express m as a piece-wise function [35], given by
m = 3 , d 50 1.5 , 50 < d 150 1 , d > 150 .
In the V2R stage, the roadside unit sends data to the vehicles within its coverage area. The signal  Y j   received by vehicle j can be represented as
Y j = h r s u , j X r s u + N
where  X r s u   is the transmission signal with power  P r s u   of the RSU,  h r s u , j   is the channel gain caused by path loss and multipath effects. N is the additive Gaussian white noise with an expectation of 0 and a variance of  σ 2  , i.e.,  N N ( 0 , σ 2 )  . Therefore, the SNR of the vehicle j in the V2R phase can be represented as
S N R r s u , j = P r r s u , j σ 2
where  P r ( r s u , j ) G a m m a m , k f P r s u m  . The transmission rate in the V2R phase is
R r s u , j = W ^ 2 log 2 1 + S N R r s u , j
where  W ^   is the total channel bandwidth.
In the V2V stage, vehicles send data to each other. During this stage, the transmission signals between different vehicles will interfere with each other. The received signal of the vehicle j can be represented as
Y j = h i , j X i + k ψ i ( h k , j X k ) + N
where  h i , j   and  h k , j   are the channel gains of the target link from vehicle i to vehicle j and interfering link from vehicle k to vehicle j, respectively.  X i   and  X k   are the transmission signals of vehicle i and vehicle k, respectively.  ψ i   is the set of parallel transmission vehicles interfering with the transmission of vehicle i. Therefore, the signal to interference plus noise ratio (SINR) at vehicle j in the V2V stage can be represented as
S I N R i , j = P r i , j α k ψ i P r k , j + σ 2
where  P r ( l , j ) G a m m a m , k f P l m  ,  α   is the interference rate of other transmitting vehicles to vehicle j. The transmission rate in the V2V phase is
R i , j = W ^ 2 log 2 1 + S I N R i , j .
To ensure communication link quality, we set an SINR threshold  S I N R t h r e  . The receiver can only decode the data correctly if the SINR is higher than this threshold. In other words, the communication link from an RSU or vehicle i to vehicle j is only valid if  S I N R j S I N R t h r e  .
To avoid interference from the vehicles’ transmitted signals to other vehicles’ receiving signals, we assume that the total wireless channel band is divided into two equal parts. Two neighbouring groups use different bands for communication.

2.2. Vehicle Clustering

We apply reinforcement learning to coordinate data exchange among vehicles in the V2V phase. However, this phase may involve a large number of vehicles, which complicates the algorithm’s design. Vehicle clustering is an effective way to enhance IoV communication performance in a large network setting with complicated road conditions and a huge number of vehicles. In most existing studies, most vehicle clustering methods select cluster head nodes first, and then form vehicle groups around these cluster heads. The cluster head node acts as the gateway of the group, and other vehicles in the group primarily communicate with it, thereby realizing communication between different groups. However, our goal of clustering is to limit the communication behaviour between vehicles of different groups by mitigating interferences in V2V collaborations, so there is no need for a complex process of selecting cluster heads. Therefore, we propose a simple and effective vehicle clustering algorithm based on the spectral clustering method [36] to divide vehicles into multiple groups. The vehicle clustering algorithm is a two-stage algorithm. In the first stage, vehicles are partitioned into different groups by spectral clustering. In the second stage, each group is optimized to ensure that the number of vehicles and the inter-vehicle distances are within predefined thresholds.
When applying spectral clustering, we need to determine two important parameters: similarity and the number of clusters. According to Equation (1), we define the similarity between two vehicles as  1 d 2  , where d is the Euclidean distance between them. We denote the position of a vehicle by  ( x , y )  , where y corresponds to the lane and x is the distance from one of the standard endpoints of the road. This definition reflects the wireless channel quality and the formula is much easier to compute.
The number of clusters, which is also the number of vehicle groups after dividing all vehicles, depends on both vehicle density and inter-vehicle similarity. It can vary under different road environments. Both factors increase the algorithm’s complexity, so we approach it from the perspective of expected results. In the V2V phase, we use MADRL to train agents to control the data exchange behaviour of vehicles within a vehicle group, and the MADRL training process requires a fixed number of agents. This leads to an upper limit on the number of vehicles within a vehicle group, denoted by  N b  . Therefore, to keep the number of vehicles within each vehicle group no higher than  N b  , we need to divide the vehicles into at least  B ^ = N / N b   groups. We use  B ^   as the number of vehicle clusters. According to the definition of similarity, we can construct the similarity matrix W and the degree matrix D. The j-th element  w i j   in the i-th row of the similarity matrix W is the similarity  1 / d i j 2   based on the distance  d i j   between vehicle i and vehicle j. The degree matrix D is a diagonal matrix where the element  d i   in the i-th row is the sum of the similarities between vehicle i and the other vehicles, i.e.,  d i = j i w i j  . Since spectral clustering converts the clustering problem into a graph optimal partitioning problem through graph Laplacian, we need to construct the graph Laplacian. The unnormalized graph Laplacian can be expressed as
L = D W .
Then we compute the eigenvectors corresponding to the top  B ^   smallest eigenvalues of the graph Laplacian L, denoted as  u 0 , u 1 , , u B ^ 1  . These  B ^   eigenvectors can form a matrix U by columns, denoted as
U = u 0 u 1 u B ^ 1 .
Then a new low-dimensional representation U of the data points is produced, where each row corresponds to a data point. The data points exhibit more salient clustering features in this representation, which enables us to apply the k-means algorithm for partitioning. By the k-means algorithm,  B ^   cluster centroids (of vehicles) are randomly initialized, then each vehicle is assigned to the closest cluster based on the Euclidean distance to the centroid. The centroid of each cluster is updated by taking the centre of all the vehicles assigned to that cluster. Then the assignment is repeated and updated until convergence or a maximum number of iterations is reached [37].
However, using only the spectral clustering algorithm may cause two problems as shown in Figure 2. The first problem is that some groups have more than  N b   vehicles, which exceeds the limit of the trained agents. The second problem is that some groups have vehicles that are too far apart to communicate effectively. To solve these problems, in the second stage, we further process the spectral clustering results. Specifically, we denote a vehicle communication range  r v e h   and assume that vehicles cannot communicate with each other if their distance exceeds  r v e h  . We introduce a queue Q and a set  B  , then put all vehicle groups partitioned by spectral clustering into queue Q. Each time we take out the first vehicle group b from the queue Q, all vehicles in the group are sorted according to their positions and the position of the first vehicle is recorded. Then, we introduce an empty vehicle group  b  . For each vehicle in group b, we first add it to group  b  . If the number of vehicles in group  b   exceeds the upper limit  N b   after adding this vehicle, or if the Eulerian distance between this vehicle and the first vehicle exceeds the communication range  r v e h  , then all other vehicles except it in group  b   are saved as a legal vehicle group and put into set  B  , and it is updated as a one-point set of group  b  , and its position is used as the position of the new first vehicle. Finally, when queue Q is empty, set  B   contains the final clustering results of the vehicles. The overall procedure for the vehicle clustering algorithm is presented in Algorithm 1.
Algorithm 1 Vehicle clustering algorithm
Input: All vehicle positions  P = { p 0 , p 1 , , p N 1 }  , the maximum number of vehicles in a group  N b  , the communication range  r v e h  
Output: The partition result  B  
1:
Calculate the Eulerian distance between all vehicles based on their positions
2:
Construct a similarity matrix W and a degree matrix D
3:
Compute the unnormalized graph Laplacian L by Equation (10)
4:
Calculate the expected number of vehicle groups  B ^ = N / N b  
5:
Compute the first  B ^   eigenvectors  u 0 , , u B ^ 1   of L and form them into a matrix  U R N × B ^   by columns
6:
Treat each row of U as a point in  R B ^   and cluster them into  B ^   clusters via k-means clustering algorithm
7:
Assign each vehicle to the cluster corresponding to its row in U, then get the initial partition result  B  
8:
Put all the groups in  B   into a queue Q
9:
Set  B =  
10:
while   Q   do
11:
     Take out a group b from the head of the queue
12:
     Sort the vehicles in b according to a certain coordinate axis
13:
     Set  b =  
14:
     Record the coordinate of the first vehicle  x 0   in b
15:
     for    i = 1 , , b   do
16:
          Add the i-th vehicle in b to the sub-group  b  
17:
          if  | b | > N b   or  x i x 0 > r r s u   then
18:
               Save all the vehicles in  b   except the last one as a new group, and add the group to  B  
19:
               Update the sub-group  b   using the one-point set of the i-th vehicle
20:
               Update the coordinate of the first vehicle  x 0 = x i  
21:
          end if
22:
     end for
23:
     Add the sub-group  b   to  B  
24:
end while

2.3. Problem Formulation

The main goal of popular content distribution is to reduce the download latency of popular content, i.e., the time for a vehicle to acquire the complete content. We divide the system time T into equal-length, non-overlapping time slots. At each time slot t, the vehicles are first clustered by the vehicle clustering algorithm proposed in Section 2.2, then the agent forms a collaborative communication policy among the vehicles based on the clustering results, and finally the vehicles transmit data according to the policy. We use  a i , j ( t ) { 0 , 1 }   to indicate whether vehicle i sends a chunk of content data to vehicle j. We denote the download latency of vehicle i to obtain the popular content by  T i  , which is the sum of the V2R and V2V phases. We aim to minimize the average time to obtain the complete content for all vehicles, expressed by
min 1 N i = 0 N T i
s . t . j N a i , j t = 1 , i N
i N a i , j t = 1 , j N
S I N R i , j S I N R t h r e , i , j N
where the constraint (12a) ensures that each vehicle only sends data to one target at a time, constraint (12b) ensures that each vehicle only receives data from one neighbour at a time, and constraint (12c) is a necessary condition for the communication link.

3. Multi-Agent Deep Reinforcement Learning

We model the collaborative data exchange process as a Markov decision process and use MADRL to control the data transmission of vehicles. MADRL has multiple agents trained in the same environment. These interact with the environment several times, in order to learn changes in the environment as well as each other’s strategies. However, each time the clustering result is different the same agent may interact with different agents in the next time slot. This prevents the effective training of the agents. Therefore, we construct an abstract training environment where each vehicle’s position is fixed. In each training, a random number of vehicles is generated. The number  N ^   of the vehicles satisfies  3 N ^ N b  . The positions of these vehicles are randomly distributed over a section of the road. Each vehicle caches a random number of popular content chunks. Each vehicle is an agent that independently decides whether and from whom to obtain a chunk of content at each slot.

3.1. Environmental Modelling

We use decentralized partially observable Markov decision processes (DEC-POMDP) [38] to model our environment. DEC-POMDP consists of the following elements: the agent’s observation of the environment, the agent’s action, the state of the environment, and the reward function. The observation  o i ( t )   of the agent i is the environmental information observed by vehicle i at slot t, which is used to decide which neighbouring vehicle to request chunks of popular content from. To make a decision, two conditions are required: first, the agent needs to know which neighbouring vehicle has the content chunks that the vehicle i needs, i.e., agent needs to know the  C i   of all neighbouring vehicles; second, the agent needs to know the position  ( x , y )   of the neighbouring vehicle to check the communication feasibility. Ultimately, we represent the observation  o i ( t )   in the form of a two-dimensional matrix. The action  a i t   of agent i is the choice of vehicle i to request content chunks from at slot t. The action  a i t   is an integer ranging from 0 to  N b 1  . Since the number of vehicles in the environment may be less than  N b 1  , and some target vehicles may not provide the required content chunks or be incommunicable, we use the invalid action masking technique [39] to prevent the agent from selecting these invalid targets. It is worth noting that when multiple vehicles choose to request chunks of popular content from the same target vehicle, we adopt a greedy strategy to resolve this conflict. That is, the target vehicle selects the closest vehicle among all requesting vehicles to send content chunks to. The state  s i t   of the environment consists of both observable and unobservable information of all agents. For simplicity, we use the observable information of all vehicles plus the action of vehicle i at the previous slot as the environmental state. The state  s i t   is also a two-dimensional matrix. For the setting of the reward function, it is obvious that the total time for all vehicles to obtain the whole popular content is reduced when the number of vehicles that receive data successfully at each slot is increased. Therefore, we use the number of vehicles that receive data successfully at slot t as the reward signal. To encourage the agents to send popular content chunks as early as possible and actively explore unsampled actions, we introduce the technique of reward shaping. The final reward function is defined as
r i t = β r N d o n e t + β p N s u c c t N n e e d t
where  N d o n e ( t )  ,  N s u c c ( t )  , and  N n e e d ( t )   are the numbers of vehicles that have acquired the complete popular content, received content chunks, and need to acquire content chunks at time slot t, respectively.  β r   and  β p   are the reward factor and the penalty factor, respectively.
Each agent learns an optimal policy to maximize the expected cumulative discounted reward. The cumulative discounted reward of agent i taking action  a i ( t )   at state  o i ( t )   at a slot t is denoted as  Q i o i ( t ) , a i ( t )  , and can be described by the Bellman equation as
Q i o i ( t ) , a i ( t ) = E r i ( t ) + γ E Q i o i ( t + 1 ) , a i ( t + 1 )
where  γ   is the discount factor,  o i ( t + 1 )   is the next state of the environment, and  a i ( t + 1 )   is next action of agent i.

3.2. MADRL-PCD Network Architecture

To train the agents, we adopt the multi-agent proximal policy optimization (MAPPO) algorithm. In MAPPO, each agent consists of two neural networks, called the actor network and the critic network. It is worth noting that the critic network is only used in the training phase to help the actor network update and optimize its policy, and in practical use, only the actor network needs to be deployed on the vehicles to make decisions.
The structure of the actor network consists of an input layer, multiple hidden layers, and an output layer as shown in Figure 3a. The input layer is responsible for receiving the observation  o i ( t )   of the vehicle. The hidden layer mainly consists of the encoder layer of the transformer architecture and the fully connected layer. We use the encoder layer in the hidden layer because we consider that the observation  o i ( t )   defined in Section 3.1 is a two-dimensional matrix that is suitable as an input to the encoder layer. The most important is that the observation  o i ( t )   can be considered as a sequence, and each row in  o i ( t )   is information about a vehicle that can correspond to a data column in the sequence. The relationship between each vehicle’s information and other vehicle information can be captured by the self-attention mechanism in the encoder layer. The fully connected layer is responsible for linearly transforming the output values of the encoder layer to highlight the relationship between the vehicle corresponding to the agent and other vehicles. The output layer takes the output of the hidden layers through a softmax function to obtain a vector of  N b   elements, represented as the probability distribution of the selection of each neighbouring vehicle by the vehicle corresponding to the agent. To avoid selecting invalid actions, such as selecting vehicles that do not exist or for which no popular content data blocks can be provided, we use an invalid action masking technique before the softmax function by setting the probability value corresponding to the invalid action to zero. After obtaining the probability distribution, if the actor network is actually used, the vehicle corresponding to the subscript with the largest probability value is selected as the best action to target; if it is the training phase, we use a polynomial distribution to randomly sample the actions in order to meet the agent’s exploration of unknown actions.
The structure of the critic network is shown in Figure 3b. Since the environmental state  s i ( t )   as the input of the critic network is defined similarly to the observation  o i ( t )  , the structure of the critic network is similar to that of the actor network. However, the output of the final critic network is a predicted value of the cumulative reward for the discount, so the hidden layer of the critic network has one more fully connected layer than the hidden layer of the actor network, which maps the output vector to a value.

4. Performance Evaluation

We utilize SUMO (version 1.12.0) [40], a well-established traffic simulation software, to build the simulation environment. The road in the environment is 10,000 m long and each RSU has a coverage of 450 m. The traffic flow follows a Poisson distribution with either 1800 or 3600 vehicles per hour. All vehicles enter from both ends of the road with an average speed of 20 m/s. The speed of each vehicle is perturbed by a normal random variable with a mean of 1 and a variance of  0.1   per time slot. The popular content is divided into multiple chunks, each of 8 MB in size.
The training environment for MADRL consists of a 250 m long road with two lanes in each direction. We set the number of popular content chunks to 16. The actor network consists of 3 transformer encoder layers and 3 fully connected layers, each with 32 neurons. The critic network has a similar structure to the actor network, but includes an additional fully connected layer at the end to convert the output vector into a scalar value. Both networks utilize the leaky ReLU function as their activation function. We use the adaptive moment estimation as the optimization algorithm for both networks, employing learning rates of  0.001   and  0.0005  , respectively. The discount factor  γ   is set to  0.9  . The simulation parameters are listed in Table 1.
We compare our proposed MADRL-PCD scheme with the CVCG scheme [20] and a greedy policy-based scheme. The CVCG scheme obtains the vehicle collaboration strategy by establishing a coalition graph game in each group, which is represented by a directed graph. Specifically, starting from an empty graph, the scheme randomly selects a vehicle from the group at each time, and updates the directed graph, which updates the target of sending and receiving data from this vehicle, to improve the vehicle collaboration strategy.This process continues until no choice of vehicle in the group can update the graph. At this point, the directed graph corresponds to the optimal vehicle collaboration strategy for the group. The greedy scheme selects the nearest neighbouring vehicle j that can provide the required content chunk to the vehicle i among all neighbouring vehicles.
Figure 4 shows the learning process of the MADRL algorithm, where each data point represents the average of 6400 reward values. It can be seen that the average reward obtained by the MADRL algorithm is low at the beginning of the training process. However, as the epoch increases, the average reward gradually increases until it reaches a relatively stable value. The curves indicate that the MADRL algorithm is convergent.
Figure 5 shows a function between the number of time slots and the number of vehicles that receive the entire popular content. The figure also considers popular contents with different sizes of 128 MB and 256 MB, shown in Figure 5a and Figure 5b, respectively. From Figure 5, we can see that the number of time slots increases as the number of vehicles receiving the whole popular content increases. Further, as the size of the popular content grows larger, more V2V collaborations are required, resulting in a larger number of required time slots (i.e., a higher delay). We can also see from Figure 5 that the proposed scheme outperforms both the CVCG scheme and the greedy scheme by taking fewer time slots to achieve a similar delivery efficiency (i.e., the same number of vehicles receiving the whole popular content). The reason is that although the traffic flow remains constant at 1800 vehicles per hour, the vehicle movements and initial content chunks of vehicles entering the V2V phase are identical for all three schemes, more vehicles in the CVCG and greedy schemes cannot complete the popular content acquisition as required. This implies that within the same time (i.e., after experiencing the same number of time slots), the proposed scheme achieves a higher PCD efficiency as more vehicles receive the whole popular content, than the CVCG and greedy schemes, which can also be seen in Figure 5.
For each vehicle, we also count the number of time slots it takes to receive the entire popular content, and then compute the average for the first 100 vehicles. Figure 6 shows the average number of time slots required for each vehicle to obtain the complete popular content at different traffic flows. Figure 6a shows the results for popular content with a size of 128 MB. Our proposed scheme outperforms both the CVCG and greedy schemes. Figure 6b shows the results for popular content with a size of 256 MB. The proposed scheme still outperforms the CVCG scheme, which in turn outperforms the greedy scheme. The advantage of the MADRL-PCD scheme is that fewer vehicles fail to receive the whole popular content within the same time period and under the same conditions as in the CVCG and greedy schemes. Therefore, it requires fewer V2V collaborations to achieve the same content delivery rate as the CVCG and greedy schemes do.
In addition to studying the impact of traffic flow, we investigated how different vehicle speeds affect the time required for vehicles to receive the complete popular content. Figure 7 shows that there is no significant difference in the number of time slots required for each vehicle to obtain popular content with three average vehicle speeds of 10, 20 and  40   m/s. As the vehicle speed decreases, the variation in channel gain between vehicles decreases, resulting in a consistent popular content distribution efficiency at average vehicle speeds of 10 and  20   m/s. On the other hand, as the vehicle speed increases, the channel environment between vehicles changes more frequently, resulting in a higher probability of errors during data transmission and a greater possibility of popular context chunk delivery failure (more time slots are required to obtain the complete popular content). However, since we set a relatively narrow communication distance for vehicles, the popular content distribution efficiency does not decrease significantly even when the average vehicle speed is  40   m/s.
Both the proposed MADRL-PCD scheme and the CVCG scheme rely on a control centre to facilitate vehicle partitioning and environment data collection, as well as communication between vehicles and the centre. Figure 8 shows that the average number of communications for each vehicle when it communicates with the control centre during the target selection phase is a function of the number of time slots. From this figure, we can see that the proposed MADRL-PCD scheme has a lower communication burden than the CVCG scheme. The reason for this is that the proposed MADRL-PCD scheme only involves four steps per time slot: uploading vehicles’ state information, downloading all vehicles’ state information, uploading vehicles’ selected targets, and downloading the final cooperation plan. However, the CVCG scheme, besides the same steps as the proposed scheme, also needs to communicate with the control centre several times to update the directed graph, which can only be performed sequentially.

5. Conclusions and Discussion

In this paper, we proposed a multi-agent deep reinforcement learning-based popular content distribution scheme (MADRL-PCD) for delivering the popular content in the IoV. Firstly, we assume that RSUs randomly broadcast popular content chunks, and vehicles acquire parts of the popular content from an RSU when passing through its communication range. Once they leave the RSU’s communication range, vehicles collaborate with each other by exchanging their received popular content chunks. Secondly, we proposed a vehicle clustering algorithm based on spectral clustering to group all vehicles in the V2V phase. This clustering algorithm helps to not only reduce the complexity of the PCD scheme but also improve the overall efficiency of the IoV. Thirdly, we use the MAPPO algorithm to train the neural network. To accelerate the training process of agents, we employed a self-attention mechanism to build the neural network. We also used the invalid action masking technique to prevent the agent from taking invalid actions. Finally, experimental results show that the MADRL-PCD scheme outperforms both the greedy scheme of selecting the nearest vehicle and the game theoretic-based CVCG scheme in terms of delivery efficiency and transmission delay.
In this work, there are some limitations in the system model. One limitation is that each vehicle is assumed to be deployed with a single antenna. Further work may employ the multi-input-multi-output (MIMO) antennas technique to improve PCD by taking into account the MIMO diversity. The other limitation is that each vehicle is assumed to send data to only one target at a time and receivedata from only one neighbour at a time. This restricts the establishment of valid communication links for V2V collaboration. This problem becomes more severe when the number of vehicles in a group is small and there are multiple concurrent transmissions that cause mutual interference. A possible solution is to allow vehicles to broadcast data to multiple targets simultaneously, so that other vehicles in the same group can receive the same data. By choosing appropriate broadcasting vehicles, the intra-group interference can be reduced and the number of valid communication links can be increased, thus enhancing V2V collaboration. Then the efficiency and time cost of PCD could be improved. Therefore, we will focus on designing PCD schemes for broadcast-based V2V communication.

Author Contributions

Conceptualization, X.H. and W.C.; methodology, W.C. and X.H.; software, W.C. and Q.G.; validation, W.C., X.H., Q.G. and S.Z.; formal analysis, W.C. and X.H.; investigation, W.C.; resources, S.Z.; data curation, W.C.; writing—original draft preparation, W.C.; writing—review and editing, X.H., Q.G. and S.Z.; visualization, Q.G. and S.Z.; supervision, X.H., Q.G. and S.Z.; project administration, X.H.; funding acquisition, X.H., Q.G. and S.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Science and Technology Planning Project of Guangdong (2020B0909030005, 2020B1212030003, 2020ZDZX3013, 2023ZZ03), Guangdong Provincial NSF (2021A1515011906), NSFC (61871201, 62077028, 61877029), the Science and Technology Planning Project of Guangzhou (202102020872, 202206030007), Key Laboratory of Smart Education of Guangdong Higher Education Institutes, Jinan University (2022LSYS003), the teaching reform research projects of Jinan University (JG2021112). the Opening Project of Key Laboratory of Safety of Intelligent Robots for State Market Regulation (GQI-KFKT202205) and the High Performance Public Computing Service Platform of Jinan University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yousefi, S.; Mousavi, M.S.; Fathy, M. Vehicular Ad Hoc Networks (VANETs): Challenges and Perspectives. In Proceedings of the 2006 6th International Conference on ITS Telecommunications, Chengdu, China, 21–23 June 2006; pp. 761–766. [Google Scholar] [CrossRef]
  2. Kaiwartya, O.; Abdullah, A.H.; Cao, Y.; Altameem, A.; Prasad, M.; Lin, C.T.; Liu, X. Internet of Vehicles: Motivation, Layered Architecture, Network Model, Challenges, and Future Aspects. IEEE Access 2016, 4, 5356–5373. [Google Scholar] [CrossRef]
  3. Yin, J.; ElBatt, T.; Yeung, G.; Ryu, B.; Habermas, S.; Krishnan, H.; Talty, T. Performance Evaluation of Safety Applications over DSRC Vehicular Ad Hoc Networks. In Proceedings of the 1st ACM International Workshop on Vehicular Ad Hoc Networks, VANET’04, Philadelphia, PA, USA, 1 October 2004; Association for Computing Machinery: New York, NY, USA, 2004; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  4. Cheng, H.T.; Shan, H.; Zhuang, W. Infotainment and Road Safety Service Support in Vehicular Networking: From a Communication Perspective. Mech. Syst. Signal Process. 2011, 25, 2020–2038. [Google Scholar] [CrossRef]
  5. El-Rewini, Z.; Sadatsharan, K.; Selvaraj, D.F.; Plathottam, S.J.; Ranganathan, P. Cybersecurity Challenges in Vehicular Communications. Veh. Commun. 2020, 23, 100214. [Google Scholar] [CrossRef]
  6. Shahwani, H.; Attique Shah, S.; Ashraf, M.; Akram, M.; Jeong, J.P.; Shin, J. A Comprehensive Survey on Data Dissemination in Vehicular Ad Hoc Networks. Veh. Commun. 2022, 34, 100420. [Google Scholar] [CrossRef]
  7. Li, M.; Yang, Z.; Lou, W. CodeOn: Cooperative Popular Content Distribution for Vehicular Networks using Symbol Level Network Coding. IEEE J. Sel. Areas Commun. 2011, 29, 223–235. [Google Scholar] [CrossRef]
  8. Schwartz, R.S.; R. Barbosa, R.R.; Meratnia, N.; Heijenk, G.; Scholten, H. A Directional Data Dissemination Protocol for Vehicular Environments. Comput. Commun. 2011, 34, 2057–2071. [Google Scholar] [CrossRef]
  9. Di Felice, M.; Bedogni, L.; Bononi, L. Group Communication on Highways: An Evaluation Study of Geocast Protocols and Applications. Ad Hoc Netw. 2013, 11, 818–832. [Google Scholar] [CrossRef]
  10. Kumar, N.; Lee, J.H. Peer-to-Peer Cooperative Caching for Data Dissemination in Urban Vehicular Communications. IEEE Syst. J. 2014, 8, 1136–1144. [Google Scholar] [CrossRef]
  11. d’Orey, P.M.; Maslekar, N.; de la Iglesia, I.; Zahariev, N.K. NAVI: Neighbor-Aware Virtual Infrastructure for Information Collection and Dissemination in Vehicular Networks. In Proceedings of the 2015 IEEE 81st Vehicular Technology Conference (VTC Spring), Glasgow, UK, 11–14 May 2015; pp. 1–6. [Google Scholar] [CrossRef]
  12. Nawaz Ali, G.; Chong, P.H.J.; Samantha, S.K.; Chan, E. Efficient Data Dissemination in Cooperative Multi-RSU Vehicular Ad Hoc Networks (VANETs). J. Syst. Softw. 2016, 117, 508–527. [Google Scholar] [CrossRef]
  13. Shafi, S.; Venkata Ratnam, D. A Cross Layer Cluster Based Routing Approach for Efficient Multimedia Data Dissemination with Improved Reliability in VANETs. Wirel. Pers. Commun. 2019, 107, 2173–2190. [Google Scholar] [CrossRef]
  14. Shafi, S.; Venkata Ratnam, D. An Efficient Cross Layer Design of Stability Based Clustering Scheme Using Ant Colony Optimization in VANETs. Wirel. Pers. Commun. 2022, 126, 3001–3019. [Google Scholar] [CrossRef]
  15. Sun, J.; Dong, P.; Du, X.; Zheng, T.; Qin, Y.; Guizani, M. Cluster-based Cooperative Multicast for Multimedia Data Dissemination in Vehicular Networks. In Proceedings of the 2020 IEEE Wireless Communications and Networking Conference (WCNC), Seoul, Republic of Korea, 25–28 May 2020; pp. 1–6. [Google Scholar] [CrossRef]
  16. Jeevitha, R.; Bhuvaneswari, N. CPDD: Clustering and Probabilistic based Data Dissemination in Vehicular Adhoc Networks. Indian J. Sci. Technol. 2022, 15, 2303–2316. [Google Scholar] [CrossRef]
  17. Hasson, S.T.; Abbas, A.T. A Clustering Approach to Model the Data Dissemination in VANETs. In Proceedings of the 2021 1st Babylon International Conference on Information Technology and Science (BICITS), Babil, Iraq, 28–29 April 2021; pp. 337–342. [Google Scholar] [CrossRef]
  18. Wang, T.; Song, L.; Han, Z.; Jiao, B. Dynamic Popular Content Distribution in Vehicular Networks using Coalition Formation Games. IEEE J. Sel. Areas Commun. 2013, 31, 538–547. [Google Scholar] [CrossRef] [Green Version]
  19. Hu, J.; Chen, C.; Liu, L. Popular Content Distribution Scheme with Cooperative Transmission Based on Coalitional Game in VANETs. In Proceedings of the 2018 21st International Symposium on Wireless Personal Multimedia Communications (WPMC), Chiang Rai, Thailand, 25–8 November 2018; pp. 69–74. [Google Scholar] [CrossRef]
  20. Chen, C.; Hu, J.; Qiu, T.; Atiquzzaman, M.; Ren, Z. CVCG: Cooperative V2V-Aided Transmission Scheme Based on Coalitional Game for Popular Content Distribution in Vehicular Ad-Hoc Networks. IEEE Trans. Mob. Comput. 2019, 18, 2811–2828. [Google Scholar] [CrossRef]
  21. Wang, Y.; Wu, H.; Niu, Y.; Han, Z.; Ai, B.; Zhong, Z. Coalition Game Based Full-Duplex Popular Content Distribution in mmWave Vehicular Networks. IEEE Trans. Veh. Technol. 2020, 69, 13836–13848. [Google Scholar] [CrossRef]
  22. Zhang, D.G.; Zhu, H.L.; Zhang, T.; Zhang, J.; Du, J.Y.; Mao, G.Q. A New Method of Content Distribution Based on Fuzzy Logic and Coalition Graph Games for VEC. Clust. Comput. 2023, 26, 701–717. [Google Scholar] [CrossRef]
  23. Huang, W.; Wang, L. ECDS: Efficient Collaborative Downloading Scheme for Popular Content Distribution in Urban Vehicular Networks. Comput. Netw. 2016, 101, 90–103. [Google Scholar] [CrossRef] [Green Version]
  24. Chen, C.; Xiao, T.; Zhao, H.; Liu, L.; Pei, Q. GAS: A Group Acknowledgment Strategy for Popular Content Distribution in Internet of Vehicle. Veh. Commun. 2019, 17, 35–49. [Google Scholar] [CrossRef]
  25. Li, Y. Deep Reinforcement Learning: An Overview. arXiv 2017, arXiv:1701.07274. [Google Scholar] [CrossRef]
  26. Wu, C.; Liu, Z.; Zhang, D.; Yoshinaga, T.; Ji, Y. Spatial Intelligence toward Trustworthy Vehicular IoT. IEEE Commun. Mag. 2018, 56, 22–27. [Google Scholar] [CrossRef]
  27. Wu, C.; Liu, Z.; Liu, F.; Yoshinaga, T.; Ji, Y.; Li, J. Collaborative Learning of Communication Routes in Edge-Enabled Multi-Access Vehicular Environment. IEEE Trans. Cogn. Commun. Netw. 2020, 6, 1155–1165. [Google Scholar] [CrossRef]
  28. Mchergui, A.; Moulahi, T.; Nasri, S. Relay Selection Based on Deep Learning for Broadcasting in VANET. In Proceedings of the 2019 15th International Wireless Communications & Mobile Computing Conference (IWCMC), Tangier, Morocco, 24–28 June 2019; pp. 865–870. [Google Scholar] [CrossRef]
  29. Lolai, A.; Wang, X.; Hawbani, A.; Dharejo, F.A.; Qureshi, T.; Farooq, M.U.; Mujahid, M.; Babar, A.H. Reinforcement Learning Based on Routing with Infrastructure Nodes for Data Dissemination in Vehicular Networks (RRIN). Wirel. Netw. 2022, 28, 2169–2184. [Google Scholar] [CrossRef]
  30. Lou, C.; Hou, F. Efficient DRL-based HD map Dissemination in V2I Communications. In Proceedings of the ICC 2022—IEEE International Conference on Communications, Dubrovnik, Croatia, 30 May–3 June 2022; pp. 5041–5046. [Google Scholar] [CrossRef]
  31. Friis, H. A Note on a Simple Transmission Formula. Proc. IRE 1946, 34, 254–256. [Google Scholar] [CrossRef]
  32. Rappaport, T.S. Wireless Communications–Principles and Practice, (The Book End). Microw. J. 2002, 45, 128–129. [Google Scholar]
  33. He, R.; Renaudin, O.; Kolmonen, V.M.; Haneda, K.; Zhong, Z.; Ai, B.; Hubert, S.; Oestges, C. Vehicle-to-Vehicle Radio Channel Characterization in Crossroad Scenarios. IEEE Trans. Veh. Technol. 2016, 65, 5850–5861. [Google Scholar] [CrossRef]
  34. Turgut, E.; Gursoy, M.C. Coverage in Heterogeneous Downlink Millimeter Wave Cellular Networks. IEEE Trans. Commun. 2017, 65, 4463–4477. [Google Scholar] [CrossRef] [Green Version]
  35. Liu, L.; Chen, C.; Qiu, T.; Zhang, M.; Li, S.; Zhou, B. A Data Dissemination Scheme Based on Clustering and Probabilistic Broadcasting in VANETs. Veh. Commun. 2018, 13, 78–88. [Google Scholar] [CrossRef]
  36. Von Luxburg, U. A Tutorial on Spectral Clustering. Stat. Comput. 2007, 17, 395–416. [Google Scholar] [CrossRef]
  37. MacKay, D. Information Theory, Inference and Learning Algorithms; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  38. Oliehoek, F.A.; Amato, C. A Concise Introduction to Decentralized POMDPs; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  39. Huang, S.; Ontañón, S. A Closer Look at Invalid Action Masking in Policy Gradient Algorithms. arXiv 2020, arXiv:2006.14171. [Google Scholar] [CrossRef]
  40. Lopez, P.A.; Behrisch, M.; Bieker-Walz, L.; Erdmann, J.; Flötteröd, Y.P.; Hilbrich, R.; Lücken, L.; Rummel, J.; Wagner, P.; Wießner, E. Microscopic Traffic Simulation using SUMO. In Proceedings of the 21st IEEE International Conference on Intelligent Transportation Systems, Maui, HI, USA, 4–7 November 2018; IEEE: Piscataway, NJ, USA, 2018. [Google Scholar]
Figure 1. System model.
Figure 1. System model.
Entropy 25 00792 g001
Figure 2. Two cases where the spectral clustering results need to be further segmented. (a) A group with too many vehicles, (b) A group with no direct communication links between vehicles.
Figure 2. Two cases where the spectral clustering results need to be further segmented. (a) A group with too many vehicles, (b) A group with no direct communication links between vehicles.
Entropy 25 00792 g002
Figure 3. The structure of the actor and critic networks.
Figure 3. The structure of the actor and critic networks.
Entropy 25 00792 g003
Figure 4. Learning process of the MADRL algorithm.
Figure 4. Learning process of the MADRL algorithm.
Entropy 25 00792 g004
Figure 5. The number of time slots vs. the number of vehicles with the whole popular content.
Figure 5. The number of time slots vs. the number of vehicles with the whole popular content.
Entropy 25 00792 g005
Figure 6. Average number of time slots needed to acquire the whole popular content.
Figure 6. Average number of time slots needed to acquire the whole popular content.
Entropy 25 00792 g006
Figure 7. The number of time slots required for vehicles to obtain popular content at different speeds.
Figure 7. The number of time slots required for vehicles to obtain popular content at different speeds.
Entropy 25 00792 g007
Figure 8. Average number of communications.
Figure 8. Average number of communications.
Entropy 25 00792 g008
Table 1. Parameters of simulation.
Table 1. Parameters of simulation.
ParameterValue
Length of simulation environment10,000 m
Length of training environment250 m
Coverage of RSU450 m
The size of popular content chunk8 MB
Transmit power of RSU P r s u = 33   dBm
Transmit power of vehicles P v e h = 23   dBm
Antenna gain   G t = 1 , G r = 1  
System loss   L s y s = 1  
Additive Gaussian white noise σ 2 = 114   dBm
Interference rate   α = 0.1  
SINR threshold S I N R t h r e = 10   dB
The learning rate of actor network   0.001  
The learning rate of critic network   0.0005  
Discount factor   γ = 0.9  
The reward factor   β r = 100  
The penalty factor   β p = 10  
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, W.; Huang, X.; Guan, Q.; Zhao, S. A Multi-Agent Deep Reinforcement Learning-Based Popular Content Distribution Scheme in Vehicular Networks. Entropy 2023, 25, 792. https://doi.org/10.3390/e25050792

AMA Style

Chen W, Huang X, Guan Q, Zhao S. A Multi-Agent Deep Reinforcement Learning-Based Popular Content Distribution Scheme in Vehicular Networks. Entropy. 2023; 25(5):792. https://doi.org/10.3390/e25050792

Chicago/Turabian Style

Chen, Wenwei, Xiujie Huang, Quanlong Guan, and Shancheng Zhao. 2023. "A Multi-Agent Deep Reinforcement Learning-Based Popular Content Distribution Scheme in Vehicular Networks" Entropy 25, no. 5: 792. https://doi.org/10.3390/e25050792

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop