Next Article in Journal
Information Conversion in Measuring Channels with Optoelectronic Sensors
Next Article in Special Issue
Modifying Hata-Davidson Propagation Model for Remote Sensing in Complex Environments Using a Multifactional Drone
Previous Article in Journal
Laser-Induced Graphene Electrodes Modified with a Molecularly Imprinted Polymer for Detection of Tetracycline in Milk and Meat
Previous Article in Special Issue
Deep Reinforcement Learning for UAV Trajectory Design Considering Mobile Ground Users
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Power Allocation and Energy Cooperation for UAV-Enabled MmWave Networks: A Multi-Agent Deep Reinforcement Learning Approach

by
Mari Carmen Domingo
Department of Network Engineering, BarcelonaTech (UPC) University, 08860 Castelldefels, Spain
Sensors 2022, 22(1), 270; https://doi.org/10.3390/s22010270
Submission received: 17 November 2021 / Revised: 22 December 2021 / Accepted: 29 December 2021 / Published: 30 December 2021
(This article belongs to the Special Issue UAV Control and Communications in 5G and beyond Networks)

Abstract

:
Unmanned Aerial Vehicle (UAV)-assisted cellular networks over the millimeter-wave (mmWave) frequency band can meet the requirements of a high data rate and flexible coverage in next-generation communication networks. However, higher propagation loss and the use of a large number of antennas in mmWave networks give rise to high energy consumption and UAVs are constrained by their low-capacity onboard battery. Energy harvesting (EH) is a viable solution to reduce the energy cost of UAV-enabled mmWave networks. However, the random nature of renewable energy makes it challenging to maintain robust connectivity in UAV-assisted terrestrial cellular networks. Energy cooperation allows UAVs to send their excessive energy to other UAVs with reduced energy. In this paper, we propose a power allocation algorithm based on energy harvesting and energy cooperation to maximize the throughput of a UAV-assisted mmWave cellular network. Since there is channel-state uncertainty and the amount of harvested energy can be treated as a stochastic process, we propose an optimal multi-agent deep reinforcement learning algorithm (DRL) named Multi-Agent Deep Deterministic Policy Gradient (MADDPG) to solve the renewable energy resource allocation problem for throughput maximization. The simulation results show that the proposed algorithm outperforms the Random Power (RP), Maximal Power (MP) and value-based Deep Q-Learning (DQL) algorithms in terms of network throughput.

1. Introduction

Unmanned Aerial Vehicles (UAVs) are aircrafts without a human pilot on board. UAVs are able to establish on-demand wireless connectivity faster than terrestrial communications and they can adjust their height and position to provide robust channels with short-range line-of-sight links [1]. Therefore, UAV-aided wireless communication is a promising solution to provide temporary connections to devices without infrastructure coverage (e.g., due to severe shadowing in urban areas) or after telecommunication infrastructure has been damaged in natural disasters [1].
UAVs are identified as an important component of future-generation (5G/B5G) wireless networks due to their salient attributes (dynamic deployment ability, strong line-of-sight connection links and additional design degrees of freedom with the controlled mobility) [2].
In UAV-assisted wireless communications, UAVs are employed to provide wireless access for terrestrial users.
We distinguish between three use cases [1]:
(1)
UAV-aided ubiquitous coverage: UAVs act as aerial base stations to achieve seamless coverage for a given geographical area. Some related applications are a fast communication service recovery in disaster scenarios and temporary traffic offloading in cellular hotspots.
(2)
UAV-aided relaying: UAVs are employed as aerial relays between far-apart terrestrial users or user groups. Some examples of applications include UAV-enabled cellular coverage extension and emergency response.
(3)
UAV-aided information dissemination and data collection: UAVs are used as aerial access points (APs) to disseminate (or collect) information to (from) ground nodes. Some related applications are UAV-aided wireless sensor networks and IoT communications.
The use of UAVs as aerial nodes to provide wireless sensing support has several advantages compared to ground sensing [3]. UAV-based sensing has a wider field of view due to the elevated height and reduced signal blockage of UAVs. In addition, UAV mobility enables to sense hard-to-reach poisonous or hazardous areas. Furthermore, the mobility of UAVs enables to perform sensing performance optimization by dynamically adjusting the trajectory of the UAVs.
UAV-based sensing has a wide range of potential applications, such as precision agriculture, smart logistics, 3D environment map construction, search and rescue, and military operations. There is a growing interest in the development of UAV-based sensing applications.
In Ref. [4], a complete framework for the data acquisition from wireless sensor nodes using a swarm of UAVs is introduced. It covers all the steps from the sensor clustering to the collision-avoidance strategy. In addition, a hybrid UAV-WSN system that improves the acquisition of environmental data in large areas has been proposed [5].
UAVs are a popular and cost-effective technology to capture high spatial and temporal resolution remote sensing (RS) images for a wide range of precision agriculture applications [6]. UAVs equipped with dual-band crop-growth sensors can achieve high-throughput acquisition of crop-growth information. IoT and UAV can monitor the incidence of crop diseases and pests from the ground micro and air macro perspectives, respectively [7]. In these applications, UAVs collect data from sensor nodes distributed over a large area. It is required to synchronize the UAV route with the activation period of each sensor node. The UAV path through all sensor nodes is optimized in Ref. [8] to reduce the flight time of the UAV and maximize the sensor nodes’ lifetime. In addition, an aerial-based data collection system based on the integration of IoT, LoRaWAN, and UAVs has been developed [9]. It consists of three main parts: (a) sensor nodes distributed throughout a farm; (b) a LoRaWAN-based communication network, which collects data from sensors and conveys them to the cloud; and (c) a path planning optimization technique for the UAV to collect data from all sensors.
In IoT communications [10], UAVs have also been proposed to assist the localisation of terrestrial Internet of Things (IoT) sensors and provide relay services in 6G networks. A mobile IoT device [11], located at a distant unknown location, has been traced using a group of UAVs equipped with received signal strength indicator (RSSI) sensors. In smart logistics, a UAV-based system aimed at automating inventory tasks has been designed and evaluated [12]. It is able to keep the traceability of industrial items attached to Radio-Frequency IDentification (RFID) tags.
In search and rescue operations, a real-time human detection and gesture recognition system based on a UAV with a camera is proposed [13]. The system is able to detect, track, and count people; it also recognizes human rescue gestures. Yolo3-tiny is used for human detection and a deep neural network is used for gesture recognition. UAV-assisted wireless networks can benefit from gigabit data transmissions by using 5G millimeter wave (mmWave) communications. The millimeter wave frequency band ranges from around 30 GHz to 300 GHz, corresponding to wavelengths from 10 to 1 mm. This key technology delivers higher data rates due to a higher bandwidth [14].
UAV-enabled mmWave networks offer a lot of potential advantages. On the one hand, the large available spectrum resources of mmWave communication and flexible beamforming can meet the requirements of high data rate and flexible coverage for UAVs serving as base stations (UAV-BSs) in UAV-assisted cellular networks [3]. These networks consist of a base station (BS) mounted on a flying UAV in the air, and mobile stations (MSs) distributed on the ground or at low altitude. High data rate communication links between the MSs and UAV BS are desirable in typical applications (e.g., to send control commands and large video monitoring traffic data from many camera sensors) [15]. On the other hand, the existence of a line-of-sight (LOS) path in the link from a UAV to the ground favors that mmWave communication obtains a high beamforming gain. However, higher propagation loss and the use of a large number of antennas in mmWave networks give rise to high energy consumption and UAVs are constrained by their low-capacity onboard battery.
Energy harvesting (EH) is a viable solution to reduce the energy cost of UAV-enabled mmWave networks; green energy can be harvested from renewable energy sources (e.g., solar, wind, electromagnetic radiations) to power UAVs. Energy-harvesting powered UAVs can prolong longer their operational duration as well as the wireless connectivity services they offer [16,17,18]. However, the random nature of renewable energy makes it challenging to maintain robust connectivity in UAV-assisted terrestrial cellular networks. Energy cooperation (also known as energy sharing or energy transfer) has been introduced in Ref. [19] to alleviate the harvested energy imbalance problem, where a source assists a relay by transferring a portion of its remaining energy to the relay.
We consider a UAV-assisted mmWave cellular network. Some UAVs will have plenty of energy because their flight duration is shorter or because they have harvested abundant energy due to better environmental conditions (e.g., sunshine without clouds). Energy cooperation allows that these UAVs can send their excessive energy to other UAVs with reduced energy.
In the literature, several works have investigated energy cooperation in renewable energy harvesting-enabled cellular networks. In Ref. [20], an adaptive traffic management and energy cooperation algorithm has been developed to jointly determine the amount of energy shared between BSs, the user association to BSs, and the sub-channel and power allocation in BSs. In Ref. [21], an energy-aware power allocation algorithm was developed in energy cooperation-enabled mmWave cellular networks. It uses renewable energy harvesting to maximize the network utility while keeping the data and energy queue lengths at a low level. In Ref. [22], a power allocation strategy is proposed that uses energy cooperation to maximize the throughput in ultra-dense Internet of Things (IoT) networks. However, these contributions do not analyze energy cooperation in UAV-assisted cellular networks.
Several authors proposed using energy transfer in UAV-enabled wireless communication systems. In Ref. [23], the total energy consumption of a UAV is minimized while accomplishing the minimal data transmission requests of the users; in the downlink the UAV transfers wireless energy to charge the users, while in the uplink the users utilize the harvested energy to transmit data to the UAV. Similarly, in Ref. [24], a downlink wireless power transfer and an uplink information transfer is proposed for mmWave UAV-to-ground networks. However, these contributions do not analyze energy cooperation between UAV-BSs.
In this paper, we propose a power allocation algorithm based on energy harvesting and energy cooperation to maximize the throughput of a UAV-assisted mmWave cellular network. This optimal power allocation and energy transfer problem can be regarded as a discrete-time Markov Decision Process (MDP) [25] with continuous state and action space. As statistical or complete knowledge about the environment, the real channel state, and the energy harvesting arrival is not easily observable, traditional model-based methods cannot be leveraged to tackle with this MDP. Therefore, we adopt multi-agent deep reinforcement learning (DRL) [26] to solve this problem and propose a multi-UAV power allocation and energy cooperation algorithm based on the Multi-Agent Deep Deterministic Policy Gradient (MADDPG) [27] method to optimize the policies for UAVs. To the best of our knowledge this is the first paper that analyses energy cooperation between UAV-BSs in UAV-assisted mmWave cellular networks and develops a DRL algorithm to maximize the network throughput. The proposed DRL algorithm can be applied in an emergency communication system for disaster scenarios. In these scenarios user devices that are out of the coverage range from UAVs cannot obtain wireless access. Therefore, it is important that UAVs increase their wireless coverage and reduce the channel access delay. Since UAVs are limited by their battery power, energy harvesting and energy cooperation are promising solutions to satisfy the requirements of an emergency communication system.
The contributions of this paper are summarized as follows:
  • We study optimal power allocation strategies for UAV-assisted mmWave cellular networks when there is channel-state uncertainty and the amount of harvested energy can be treated as a stochastic process.
  • We formulate the renewable energy resource allocation problem for throughput maximization using multi-agent DRL and propose an MADDPG-based multi-UAV power allocation algorithm based on energy cooperation to solve this problem.
  • Simulation results show that our proposed algorithm outperforms the Random Power (RP), Maximal Power (MP) and value-based Deep Q-Learning (DQL) algorithms and achieves a higher average network throughput.
The paper is structured as follows. In Section 2, we analyze our system model. In Section 3, we state the renewable energy resource allocation problem and formulate it as an MDP with the objective to maximize the throughput. In Section 4, we introduce the MADDPG-based multi-UAV power allocation algorithm based on energy cooperation for solving the MDP. Simulation results are presented in Section 5. Finally, the paper is concluded in Section 6.

2. System Model

Our network architecture is shown in Figure 1. For clarity, we summarize all the following notations and their definitions in Table 1. We consider a multi-antenna mmWave UAV-enabled wireless communication system, where multiple UAV-mounted aerial base stations (BSs) fly over the region and serve a group of users on the ground. We consider that each UAV is dedicated to serve a cluster of users with the same requirements. The locations of the UAV-enabled BSs are modelled as a Poisson point process (PPP) Φ j with density λ j . We consider that users are static. We assume that the location of users is modelled as a Poisson cluster process (PCP) Φ i   with density λ i   [28]. We also assume that all UAVs are elevated at the same altitude H j   0 .
UAVs are powered by hybrid energy sources. Onboard energy, along with a part of the harvested energy, is used to maintain the flight while the rest harvested energy is used to support the communication modules of the UAVs. The imbalance of energy harvesting between UAVs is compensated through energy cooperation.
The UAVs and user sets are denoted as M = 1 , , M and L = 1 , , L . The total number of users served by UAVj, j 1 , 2 , , M , can be represented by Lj only associated with UAVj. For simplicity, the typical user set is associated with the closest UAV-BS; that is, the UAV that maximizes the average received SNR.
This paper focuses on the design of an optimal power allocation strategy to maximize the throughput for multi-UAV networks over N time slots. It is assumed that all UAVs communicate without the assistance of a central controller and have no global knowledge of wireless channel communication environments. This means that the channel state information (CSI) between a UAV and the mobile devices of the users is known locally.

2.1. Blockage Model

A major challenge in mmWave communications is the blockage effect [14], namely, mmWave signals are blocked by physical obstacles in their propagation. We adopt the building blockage model introduced in Ref. [29], which defines an urban area as a set of buildings in a square grid. The mean number of buildings per square kilometer is ρ . The fraction of area covered by buildings to the total area is α . Each building has a height which is a Rayleigh-distributed random variable with scale parameter κ . The probability of a UAV having a line-of-sight (LOS) connection to the user i when the horizontal transmission distance is r   is given by
P L h t , h r , r = n = 0 max 0 , d 1 ( 1 exp ( max h t , h r n + 0.5 h t h r d 2 2 κ 2 )
where   h t is the transmitter height, h r   is the receiver height, d = r ρ α and . is the floor function. Furthermore, the probability for a non-line-of-sight (NLOS) transmission is P N . = 1 P L . .

2.2. UAV-to-Ground Channel Model

The path loss law in the UAV network is given by [30]:
L h t , h r , r B P L h t , h r ,   r C L r 2 + h t h r 2 α L + B P N h t , h r ,   r C N r 2 + h t h r 2 α N
where B x   is a Bernoulli random variable with parameter x . The parameters α L and α N are the LOS and NLOS path loss exponents, and C L and C N are the intercepts of the LOS and NLOS links.
The amplitude of the received UAV-to-ground mmWave signal can be modelled as a Nakagami- m fading distribution for both the LOS and NLOS propagation conditions at mmWave frequency bands. Let h ^ S . be the small-scale fading term on the l -th link. h ^ S 2 is a normalized Gamma random variable. h ^ L ~ Γ N L , 1 N L for LOS and h ^ N ~ Γ N N , 1 N N for NLOS, where N L and N N represent the Nakagami fading parameters for the LOS and NLOS links.
At time slot t , the LOS channel gain from the j -th UAV BS located at x t   2   to a i -th ground user located at x r   2   can be expressed as
H i , j h t , h r ,   x r   = L h t , h r ,   x r G x r h ^ x r 2
where G x r is the directional antenna gain.
When the transmission has the maximum antenna gain, the channel gain is H i , j 0 h t , h r ,   x r   = L ( h t , h r ,   x r ) G 0 h ^ x r 2 .

2.3. Directional Beamforming

Beamforming, also known as spatial filtering, concentrates the signal energy over a narrow beam by means of highly directional signal transmission or reception; this way, the spectral efficiency is improved [14]. The narrow beams of the mmWave signals allow to achieve highly directional signals along the desired directions. We assume that N B and N U   antenna arrays are deployed at both the UAV-BSs and the mobile user sets, respectively. It is required to use efficient alignment policies (beam tracking, beam training, hierarchical beam codebook design, accurate estimation of the channel, etc.) to align the beams between transmitter and receiver [14].
We consider that the UAV-BS and user set antennas adopt a sectorized model that is shown in Figure 2. The antenna array pattern is characterized by four parameters: the half-power beamwidth in the azimuth plane θ c a , the half-power beamwidth in the elevation plane θ c e , the mean lobe gain M c   and the side lobe gain m c , where c t , r   refers to the transmitter (UAV-BS) and the receiver (user set), respectively.
The directivity gain at one receiver located at l from the j -th UAV-BS can be expressed as follows:
G l = G M t , m t , θ t a , θ t e G M r , m r , θ r a , θ r e
where G ( θ c a , θ c e ,   M c , m c ) denotes the directional antenna gain.
The transmitter and receiver should adjust their antenna directions towards each other to achieve the maximum beamforming gain G 0 = M t M r .

2.4. Signal Model

The UAV-to-ground user pair communication is affected by the interference signals from the remaining UAVs. Therefore, the received signal-to-interference-plus-noise ratio (SINR) from UAV j located at x t 2   to user i at time slot t can be expressed as
γ i j t = H i , j t P i , j t I i , j t + σ 2
where H i , j t   denotes the channel gain between UAVj and user i at time slot t , P i , j t   is the transmit power selected by UAVj at time slot t , I i , j is the interference to UAVj that satisfies I i , j t = m M ,   m j H i , m t P i , m t and σ 2 is the noise power level.

3. Problem Formulation

In this section, we investigate the optimal power allocation and energy transfer problem for throughput maximization in UAV-enabled mmWave networks, which can be regarded as MDP. Since the real channel state and energy harvesting arrival are not easily observable, traditional model-based methods are infeasible to tackle with this MDP. Therefore, we reformulate this problem using multi-agent DRL to make it solvable.

3.1. Throughput Maximization Problem

We introduce the following notation:
  • t 1 , 2 , , N is one time slot of a finite horizon of N time slots.
  • τ = t n t n 1 ,   t 1 , 2 , , N is the time slot duration.
  • E j       is the amount of harvested energy for UAVj at time slot t .
  • C is the battery capacity of each UAVj.
  • B j 0 , , C is the battery state for UAVj at time slot t .
  • P j is the transmission power allocated by UAVj to serve all its users.
The theoretical downlink rate of user i ,   i L j   connected to UAVj at time slot t is given by
R i , j t = W l o g 2 1 + γ i , j t
where W is the mm Wave transmission bandwidth.
The total throughput at timeslot t is
U t = j = 1 M i = 1 L j R i , j t
The current battery capacity of each UAVj stores mainly the renewable energy harvested during the current time slot, the energy transferred by other UAVs during the energy cooperation process, and the remaining energy from the last time slot.
The charging rate of the energy storage is usually less than the energy arrival rate, because of the limited energy conversion efficiency of the circuits. We consider that the charging rate and energy arrival rate are expressed as E j and E j ^   ,   respectively. Therefore, E j = η E j ^ ≥ 0, where 0 <   η 1 is the imperfect conversion efficiency. In the rest of the paper, the energy arrival rate refers to the effective energy arrival rate that is assimilated by the system, i.e., the charging rate of the storage E j .
After E j is harvested at time slot t , it is stored in the battery and is available for transmission in time slot t + 1 . The rechargeable battery is assumed to be ideal, which means that no energy is lost with energy storing or retrieving. Once the battery is full, the additional harvested energy is removed.
The battery energy level of UAVj at the time t + 1 is
B j t + 1 = min { C , B j t + E j t τ P j t j = 1 , j j M ε j j t + j = 1 , j j M β ε j j t }
where ε j j denotes the energy transferred from UAVj to UAVj, ε j j denotes the energy transferred from UAVj’ to UAVj and β 0 ,   1 is the energy transfer efficiency between two UAVs.
The problem can be formulated as follows:
P1. Throughput optimization problem
Find:   P i , j t
Max:
t N U t
Subject to:
0 E j t   E m a x t
0 τ P j t   B j t + 1
P j t = i = 1 L j P i , j t
0 j = 1 , j j M ε j j t   min { C ,   B j t + E j t }
  0 j = 1 , j j M ε j j t   min { C , B j t + E j t }  
The objective function of problem P1 aims at finding the best values of P i , j t that maximizes the throughput. We observe that P1 is a non-linear optimization problem.
Constraint (10) limits the harvested energy. Constraint (11) determines that the total energy consumed by each UAV should not exceed its battery level. Constraint (12) refers to the total allocated power by UAV j to serve all its users at time t . Since the energy storage at each UAV is limited, Constraint (13) expresses that the total energy transferred from UAVj to other UAVs should not exceed the current battery energy level. The same applies to Constraint (14) for the total transferred energy transferred from UAVj’ to other UAVs.
The optimization problem may be solved only if the complete information about energy harvesting arrival and channel state information (CSI) is known. RL algorithms can achieve near-optimal performance even without prior knowledge about the CSI, the user arrival, the energy arrival, etc. [31]. In what follows, we analyze this problem under the MDP framework and reformulate it by adopting multi-agent RL.

3.2. Multi-Agent RL Formulation

The proposed problem can be considered an MDP. Therefore, multi-agent RL can be adopted to solve this problem efficiently.
Each UAV can be regarded as an agent in the proposed system and all the network settings can be regarded as the environment. UAVs can be characterized by a tuple S ,   A j j M , P , R j j   M as follows:
  • S denotes the state space including all possible states of UAVs in the system at each time slot. The state of the j -th UAV, denoted by s j = ( γ j t , R j t 1 ,   E j t ) ,   is described by the current SINR of the users served by UAVj, the link’s corresponding downlink rate R j t 1 at the last time slot and the current harvested energy of UAVj, respectively.
  • γ j t = { γ 1 j t , γ 2 j t , γ L j j t } , j 1 , 2 ,   M refers to the SINR of the current serving users of UAV   j . R j t 1 = { R 1 j t 1 , R 2 j t 1 , R L j j t 1 }, j 1 , 2 ,   M refers to the downlink rate of the current serving users of UAV   j .
  • A j , j M denotes the action space consisting on all the available actions of j -th UAV at each time slot. The action of the j -th UAV, denoted by a j , is defined as a j = P j . This means that each UAV selects the power allocated to serve its users. At state s j , the available action set of the j -th UAV is expressed as A j = A s j .
  • P : S M × j = 1 M A j S is the state transition function, which maps the state spaces and the action spaces of all UAVs in the current time slot to their state spaces in the next time slot.
  • R j , j   M   is the reward function of the j -th UAV, which maps the state spaces and the action spaces of the UAV in the current time slot to its expected reward. The reward of the j -th UAV, denoted by r j , is defined as r j = U t .
Each UAV is motivated to maximize the throughput by making decisions on power allocation. In our system, the policy of a UAV is defined as a mapping from its state space to its action space, denoted by π . At the beginning of each time slot the j -th UAV observes the state of all UAVs ,     s = s j , from state space S M , and takes an action a j based on its policy π j . Actually, a UAV cannot know the states of other UAVs by itself. However, before making decision in a cycle, the UAV can observe the states of other UAVs by sending a beacon. After making the decision, the UAV will keep its decision unchanged till the end of the current cycle. The policy of the j -th UAV can be defined as a j = π j s , where s is the state of all UAVs in the system and a j   is the action of the j -th UAV. After that, the UAV receives a reward r j and then observes the next state s , namely, the states of all UAVs at the beginning of the next time slot. Therefore, the throughput maximization problem can be transformed into maximizing the total accumulated rewards of all UAVs in the system by optimizing their policies; that is,
P2. Maximization of the total accumulated rewards
Max:
t N r j

4. Proposed Multi-Agent Reinforcement Learning Algorithm

Our target is to design an efficient algorithm for power allocation and energy transfer for UAVs that maximizes the throughput. Existing approaches such as dynamic programming is not suitable for such challenging tasks. Therefore, we adopt an RL algorithm to cope with the problem (P2). DRL has a better performance on tasks that have a sophisticated state space and time-varying environment than traditional reinforcement learning. There are different kinds of DRL that could deal with different situations, e.g., Deep Q-Learning (DQL) could work well with a limited action space and deep deterministic policy gradient (DDPG [32]) has a remarkable performance with continued action space. There are two ways to apply the DDPG algorithm in our proposed scenario. The first solution would be to have a global DDPG agent that outputs all UAVs’ actions and there is only one reward function in this centralized fully observable case. However, if we consider one action (transmission power) with an infinite action space for each UAV, a global agent would have to cope with an exponential number of actions, which would become a problem. Another solution is to apply DDPG on each UAV; in this case, we would have multiple DDPG agents that output actions for each UAV. However, this solution is inefficient compared to MADDPG [27] because at every time slot each UAV agent will be trying to learn to predict the actions of the other UAVs while also taking its own actions. On the other hand, MADDPG is the state-of-the art solution for multi-agent DRL. It employs a centralized critic and decentralized actors. Actors can use the estimated policies of other agents for learning. This way, agents are supplied with information about the other UAVs’ observations and potential actions, transforming an unpredictable environment into a predictable one. This additional information is used to simplify the training, as long as it is not used at the test time (centralized training with decentralized execution). In this paper, we propose a MADDPG-based multi-UAV design algorithm for power allocation and energy transfer to optimize multiple UAVs’ policies. Afterwards, we introduce the training process of the proposed algorithm.

4.1. Algorithm Design

MADDPG is an actor-critic algorithm [33] designed for multi-agent environments. Actors are responsible for learning policies and critics evaluate the actors’ action choices.
MADDPG adopts a strategy based on centralized training and distributed execution. Each UAV works as an agent and has an actor network θ j μ θ 1 μ ,   , θ M μ , which means that each agent j takes continuous policies μ j with regard to parameters θ j μ . Each agent also has a critic network θ j Q θ 1 Q ,   , θ M Q . The critics are fed with information about the global state S and actions A of all agents. They are aware of the actions of all agents and output a Q   value that describes how good joint action A is on state S . Target networks serve as stable targets for learning. Each agent has an actor target network θ j μ θ 1 μ ,   , θ M μ and a critic target network θ j Q θ 1 Q ,   , θ M Q .
The loss function of the critic network is calculated by
L θ j Q = E s t ~ D [ ( Q j ( s t , a t | θ j Q ) y j t ) 2 ]
where D is the replay buffer that stores historical experience. y j t is defined as
y j t = r j t + λ Q j s t + 1 , a j t + 1 | θ j Q | a j t + 1 = μ j s j t + 1
The gradient of the expected reward for agent j with deterministic policies μ j   is given by
Θ j μ   J ( Θ j μ ) = E s t ~ D [ a j t   Q j s t , a j t θ j μ a j t | a j t = μ j s j t ]
The MADDPG algorithm is shown as Algorithm 1.
Algorithm 1: MADDPG-based multi-UAV power allocation algorithm based on energy cooperation
Input: The structures of the actor network, critic network, and their target networks; Number of episodes
Output: Policy π j ;
1: Initialize the replay memory R M with size X.
2: Initialize critics Q j ( s t , a t | θ j Q ) and actors μ j ( s j |   θ j μ ) with random weights θ j Q   and   θ j μ
Initialize target networks with random weights θ j Q and θ j μ
3: Receive initial state s j t ;
4: for   n e p i = 1 : N e p i  do
5:  for   t = 1 :   N do
6:   Each agent selects an action a j t = μ j ( s j t |   θ j μ ) +   v j ,   where   v j is the exploration of the action.
7:   Receive reward r j t and observe next state s j t + 1
8:   Store ( s j t ,   a j t ,   r j t , s j t + 1 ) in the replay memory R M
9:   if   R M is full do
10:    Sample a batch of random samples ( ( s j t ,   a j t ,   r j t , s j t + 1 ) from R M .
11:    Set with (17).
12:    Update the actor of the estimated network θ j μ   with (18).
13:    Update the critic of the estimated network θ j Q   with (16).
Update the target network parameters with θ j Q ξ θ j Q + 1 ξ θ j Q ,   θ j μ ξ θ j μ + 1 ξ θ j μ
14:  end for
15: end for
16: Return θ j μ .
17: Choose optimal action a j t * = μ j ( s j t |   θ j μ ) at time t .

4.2. Complexity Analysis

The computation complexity and the space complexity for the proposed MADDPG algorithm can be estimated by the replay memory and the neural networks’ architecture. In MADDPG, the training network of each agent consists of two sets of actor networks and two sets of critic networks. The time complexity (computations) is given with regard to the floating-point operations per second (FLOPS). The neural networks are fully connected layer networks. For dot products of a P   vector and a P × Q matrix, the FLOPS is 2 P 1 Q because for every column in matrix we need to multiply P   times and add P 1 times. It is also necessary to derive the computations of the activation layers. In this case addition, subtraction, multiplication, division, exponentiation, square root, etc., count as a single FLOP. Therefore, the computations are Q with Q inputs for ReLU layers, 4 × Q for sigmoid layers and 6 × Q for tanh layers.
We consider that u a , j is the unit number in the j -th layer of the actor, and u c , j the number of neurons in the k -th layer of the critic. The number of layers for the actor and critic networks are J and K , respectively.
Therefore, the time complexity of the training is:
v a c t u i + 2 j = 0 J 1 u a , j u a , j + 1 + 2 k = 0 K 1 u c , k u c , k + 1            = O j = 0 J 1 u a , j u a , j + 1 + k = 0 K 1 u c , k u c , k + 1  
where u i means the unit number in the i -th layer and v a c t means the corresponding parameters determined by the type of the activation layer.
Space is needed to store the learning transition. The memory replay in MADDPG occupies some space to store the state sets; therefore, the space complexity is N . For a fully connected layer in both the actor and the critic network, there is a P × Q matrix and a Q bias vector. The memory for a fully connected layer is P + 1 Q .   The space complexity of the neural networks is given by
2 j = 0 J 1 u a , j u a , j + 1 + 2 k = 0 K 1 u c , k u c , k + 1 + v a c t u i + N            = O j = 0 J 1 u a , j u a , j + 1 + k = 0 K 1 u c , k u c , k + 1 + O N  

5. Simulation Results

In this section, we present the simulation results of the multi-UAV power allocation and energy transfer algorithm. The simulation parameters are given in Table 2. We assume that the mmWave network is operated at 28 GHz. We set each actor or critic network as a four-layer neural network with two hidden layers, in which the number of neurons in the two hidden layers are 64 and 128, respectively. The activation function for the hidden layers is rectified linear unit (ReLU) f x = max o , x . The Adam algorithm is adopted as the optimizer and the learning rate is set as exponentially decayed to improve the performance of the training.
Next, we compare the proposed algorithm MADDPG with the Random Power (RP), Maximal Power (MP), Multi-Armed Bandit (MAB) and value-based Deep Q-Learning (DQL) algorithms. Upper confidence bound (UCB) is used to solve the MAB problem. RP and MP are two classical algorithms, whereas MADDPG, DQL and MAB are RL-based algorithms. Maximal Power (MP) consumes as much energy as possible in each time slot to improve its immediate throughput regardless of the future and the performance of the other UAVs. Random Power (RP) only consumes a part of the energy chosen randomly in each time slot. The average throughput for the five power allocation schemes is shown in Figure 3 as a function of the number of UAVs. We observe that the average throughput is increased with the number of UAVs. MADDPG achieves the highest average throughput in all testing scenarios and outperforms the other algorithms. The gap between the RP/MP allocation schemes and the rest of the algorithms is decreased when the number of UAVs increases.
The average throughput for the different policies is shown in Figure 4 as a function of the number of time slots. We can observe that MADDPG always outperforms the other algorithms. Since MADDPG does not divide the action space into discrete values like DQL, it can select a better action in each time slot without quantization errors. We notice that when the number of time slots increases, the average throughput is much larger for the RL-based algorithms, because they can adjust the transmission power in a smart way. MAB shows similar but worse behavior than DQL.
The average throughput as a function of the number of users is illustrated in Figure 5 with four UAVs. We observe that the average throughput is increased because more users are served. MADDPG improves the average throughput compared to the other approaches. It is 13.6% higher than DQL, 22.53% higher than MAB, 46.24% higher than RP and 49.56% higher than RP for 24 users, which proves the effectiveness of the proposed approach.
The energy transfer efficiency between two UAVs is shown in Figure 6 as a function of the number of users with four UAVs for MADDPG. We observe that the energy transfer efficiency has a high value for a different number of users.
The average throughput for the RL-algorithms is shown in Figure 7 as a function of the energy arrival E m a x . We observe that the average throughput is increased with the maximum energy harvested E m a x . The average throughput for MADDPG is higher than for DQL and for MAB. Since the amount of collected energy is lower than the size of the battery capacity the throughput is increased for larger values of E m a x .
The average throughput for the RL-algorithms is shown in Figure 8 as a function of the battery capacity C . We observe that the average throughput is increased with the battery capacity. The average throughput is higher for MADDPG than for DQL and for MAB. We observe that when the battery capacity is increased the throughput values for the policies tend to stabilize since the value of E m a x limits the system throughput increase.
The average throughput for the RL-algorithms is shown in Figure 9 as a function of the energy transfer efficiency between two UAVs β . We observe that the average throughput is increased with the energy transfer efficiency. The average throughput is higher for MADDPG than for DQL and for MAB for all the values of β .
The convergence behavior of the reinforcement-based algorithms in terms of average reward is shown in Figure 10 for a network of three UAVs and 14 users. The convergence behavior is around 1400 iterations for MADDPG. For DQL it is shorter (around 600 iterations). Finally, for MAB the convergence time is larger (around 1100 iterations).

6. Conclusions

In this paper, the optimal power allocation strategies for UAV-assisted mmWave cellular networks were analyzed. A power allocation algorithm based on energy harvesting and energy cooperation is proposed to maximize the throughput of a UAV-assisted mmWave cellular network. Since there is channel-state uncertainty and the amount of harvested energy can be treated as a stochastic process, we propose an optimal multi-agent deep reinforcement learning algorithm (DRL) named Multi-Agent Deep Deterministic Policy Gradient (MADDPG) to solve the renewable energy resource allocation problem for throughput maximization. The simulation results show that the proposed algorithm outperforms the Random Power (RP), Maximal Power (MP), Multi-Armed Bandit (MAB) and value-based Deep Q-Learning (DQL) algorithms in terms of network throughput. Besides, the RL-based algorithms outperform the traditional RP and MP algorithms and show improved generalization performance, since they can adjust the transmission power in a smart way. The average throughput is increased with the number of UAVs, the energy arrival, the battery capacity and the energy transfer efficiency between two UAVs. When the battery capacity is increased the throughput values for the RL policies tend to stabilize since the value of E m a x limits the system throughput increase.
MADDPG can be applied to many tasks with discrete or continuous state/action space and joint optimization problems of multiple variables. It can successfully solve user scheduling, channel management and power allocation problems in different types of communication networks. The optimization of the locations of UAVs and their trajectories [34] is an important topic. Therefore, we will further investigate the development of a joint power allocation and UAV trajectory approach as future work.

Funding

This work was supported by the Agencia Estatal de Investigación of Ministerio de Ciencia e Innovación of Spain under project PID2019-108713RB-C51 MCIN/AEI /10.13039/501100011033.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zeng, Y.; Zhang, R.; Lim, T.J. Wireless communications with unmanned aerial vehicles: Opportunities and challenges. IEEE Commun. Mag. 2016, 54, 36–42. [Google Scholar] [CrossRef] [Green Version]
  2. Li, B.; Fei, Z.; Zhang, Y. UAV Communications for 5G and Beyond: Recent Advances and Future Trends. IEEE Internet Things J. 2019, 6, 2241–2263. [Google Scholar] [CrossRef] [Green Version]
  3. Wu, Q.; Xu, J.; Zeng, Y.; Ng, D.W.K.; Al-Dhahir, N.; Schober, R.; Swindlehurst, A.L. A Comprehensive Overview on 5G-and-Beyond Networks with UAVs: From Communications to Sensing and Intelligence. IEEE J. Sel. Areas Commun. 2021, 39, 2912–2945. [Google Scholar] [CrossRef]
  4. Roberge, V.; Tarbouchi, M. Parallel Algorithm on GPU for Wireless Sensor Data Acquisition Using a Team of Unmanned Aerial Vehicles. Sensors 2021, 21, 6851. [Google Scholar] [CrossRef]
  5. Popescu, D.; Dragana, C.; Stoican, F.; Ichim, L.; Stamatescu, G. A Collaborative UAV-WSN Network for Monitoring Large Areas. Sensors 2018, 18, 4202. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Yao, L.; Wang, Q.; Yang, J.; Zhang, Y.; Zhu, Y.; Cao, W.; Ni, J. UAV-Borne Dual-Band Sensor Method for Monitoring Physiological Crop Status. Sensors 2019, 19, 816. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Gao, D.; Sun, Q.; Hu, B.; Zhang, S. A Framework for Agricultural Pest and Disease Monitoring Based on Internet-of-Things and Unmanned Aerial Vehicles. Sensors 2020, 20, 1487. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Just, G.E.; Pellenz, M.E.; Lima, L.A.; Chang, B.S.; Souza, R.D.; Montejo-Sánchez, S. UAV Path Optimization for Precision Agriculture Wireless Sensor Networks. Sensors 2020, 20, 6098. [Google Scholar] [CrossRef] [PubMed]
  9. Behjati, M.; Noh, A.B.M.; Alobaidy, H.A.H.; Zulkifley, M.A.; Nordin, R.; Abdullah, N.F. LoRa Communications as an Enabler for Internet of Drones towards Large-Scale Livestock Monitoring in Rural Farms. Sensors 2021, 21, 5044. [Google Scholar] [CrossRef]
  10. Khisa, S.; Moh, S. Medium Access Control Protocols for the Internet of Things Based on Unmanned Aerial Vehicles: A Comparative Survey. Sensors 2020, 20, 5586. [Google Scholar] [CrossRef]
  11. Spyridis, Y.; Lagkas, T.; Sarigiannidis, P.; Argyriou, V.; Sarigiannidis, A.; Eleftherakis, G.; Zhang, J. Towards 6G IoT: Tracing Mobile Sensor Nodes with Deep Learning Clustering in UAV Networks. Sensors 2021, 21, 3936. [Google Scholar] [CrossRef]
  12. Fernández-Caramés, T.M.; Blanco-Novoa, O.; Froiz-Míguez, I.; Fraga-Lamas, P. Towards an Autonomous Industry 4.0 Warehouse: A UAV and Blockchain-Based System for Inventory and Traceability Applications in Big Data-Driven Supply Chain Management. Sensors 2019, 19, 2394. [Google Scholar] [CrossRef] [Green Version]
  13. Liu, C.; Szirányi, T. Real-Time Human Detection and Gesture Recognition for On-Board UAV Rescue. Sensors 2021, 21, 2180. [Google Scholar] [CrossRef]
  14. Zhang, L.; Zhao, H.; Hou, S.; Zhao, Z.; Xu, H.; Wu, X.; Wu, Q.; Zhang, R. A Survey on 5G Millimeter Wave Communications for UAV-Assisted Wireless Networks. IEEE Access 2019, 7, 117460–117504. [Google Scholar] [CrossRef]
  15. Xiao, Z.; Xia, P.; Xia, X.-G. Enabling UAV cellular with millimeter-wave communication: Potentials and approaches. IEEE Commun. Mag. 2016, 54, 66–73. [Google Scholar] [CrossRef] [Green Version]
  16. Kingry, N.; Towers, L.; Liu, Y.-C.; Zu, Y.; Wang, Y.; Staheli, B.; Katagiri, Y.; Cook, S.; Dai, R. Design, Modeling and Control of a Solar-Powered Quadcopter. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–26 May 2018; pp. 1251–1258. [Google Scholar]
  17. Balraj, S.; Ganesan, A. Indirect Rotational Energy Harvesting System to Enhance the Power Supply of the Quadcopter. Def. Sci. J. 2020, 70, 145–152. [Google Scholar] [CrossRef]
  18. Zhang, J.; Lou, M.; Xiang, L.; Hu, L. Power cognition: Enabling intelligent energy harvesting and resource allocation for solar-powered UAVs. Future Gener. Comput. Syst. 2020, 110, 658–664. [Google Scholar] [CrossRef]
  19. Gurakan, B.; Ozel, O.; Yang, J.; Ulukus, S. Energy Cooperation in Energy Harvesting Communications. IEEE Trans. Commun. 2013, 61, 4884–4898. [Google Scholar] [CrossRef] [Green Version]
  20. Lee, H.-S.; Lee, J.-W. Adaptive Traffic Management and Energy Cooperation in Renewable-Energy-Powered Cellular Networks. IEEE Syst. J. 2019, 14, 132–143. [Google Scholar] [CrossRef]
  21. Xu, B.; Chen, Y.; Carrion, J.R.; Loo, J.; Vinel, A. Energy-Aware Power Control in Energy Cooperation Aided Millimeter Wave Cellular Networks with Renewable Energy Resources. IEEE Access 2016, 5, 432–442. [Google Scholar] [CrossRef]
  22. Li, Y.; Zhao, X.; Liang, H. Throughput Maximization by Deep Reinforcement Learning with Energy Cooperation for Renewable Ultradense IoT Networks. IEEE Internet Things J. 2020, 7, 9091–9102. [Google Scholar] [CrossRef]
  23. Yang, Z.; Xu, W.; Shikh-Bahaei, M. Energy Efficient UAV Communication with Energy Harvesting. IEEE Trans. Veh. Technol. 2019, 69, 1913–1927. [Google Scholar] [CrossRef] [Green Version]
  24. Zhu, Y.; Zheng, G.; Wong, K.-K.; Dagiuklas, T. Spectrum and Energy Efficiency in Dynamic UAV-Powered Millimeter Wave Networks. IEEE Commun. Lett. 2020, 24, 2290–2294. [Google Scholar] [CrossRef]
  25. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
  26. Arulkumaran, K.; Deisenroth, M.P.; Brundage, M.; Bharath, A.A. Deep Reinforcement Learning: A Brief Survey. IEEE Signal Process. Mag. 2017, 34, 26–38. [Google Scholar] [CrossRef] [Green Version]
  27. Lowe, R.; Wu, Y.; Tamar, A.; Harb, J.; Abbeel, O.P.; Mordatch, I. Multi-agent Actor-critic for Mixed Cooperative-competitive Environments. In Proceedings of the Thirty-First Annual Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 6379–6390. [Google Scholar]
  28. Yi, W.; Liu, Y.; Nallanathan, A.; Karagiannidis, G.K. A Unified Spatial Framework for Clustered UAV Networks Based on Stochastic Geometry. In Proceedings of the 2018 IEEE Global Communications Conference (GLOBECOM), Abu Dhabi, UAE, 9–13 December 2018; pp. 1–6. [Google Scholar]
  29. ITU-R. Recommendation p.1410-5: Propagation Data and Prediction Methods Required for the Design of Terrestrial Broadband Radio Access Systems Operating in a Frequency Range from 3 to 60 Ghz; ITU: Geneva, Switzerland, 2012. [Google Scholar]
  30. Bai, T.; Heath, R.W. Coverage and Rate Analysis for Millimeter-Wave Cellular Networks. IEEE Trans. Wirel. Commun. 2015, 14, 1100–1114. [Google Scholar] [CrossRef]
  31. Li, H.; Lv, T.; Zhang, X. Deep Deterministic Policy Gradient Based Dynamic Power Control for Self-Powered Ultra-Dense Networks. In Proceedings of the 2018 IEEE Globecom Workshops (GC Wkshps), Abu Dhabi, United Arab Emirates, 9–13 December 2018; pp. 1–6. [Google Scholar]
  32. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
  33. Grondman, I.; Busoniu, L.; Lopes, G.A.; Babuska, R. A Survey of Actor-Critic Reinforcement Learning: Standard and Natural Policy Gradients. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2012, 42, 1291–1307. [Google Scholar] [CrossRef] [Green Version]
  34. Arani, A.H.; Hu, P.; Zhu, Y. Re-envisioning Space-Air-Ground Integrated Networks: Reinforcement Learning for Link Optimization. In Proceedings of the ICC 2021—IEEE International Conference on Communications, Montreal, QC, Canada, 14–23 June 2021; pp. 1–7. [Google Scholar]
Figure 1. Network architecture.
Figure 1. Network architecture.
Sensors 22 00270 g001
Figure 2. Sectorized antenna pattern.
Figure 2. Sectorized antenna pattern.
Sensors 22 00270 g002
Figure 3. Average throughput as a function of the number of UAVs.
Figure 3. Average throughput as a function of the number of UAVs.
Sensors 22 00270 g003
Figure 4. Average throughput as a function of the number of time slots.
Figure 4. Average throughput as a function of the number of time slots.
Sensors 22 00270 g004
Figure 5. Average throughput as a function of the number of users with 4 UAVs.
Figure 5. Average throughput as a function of the number of users with 4 UAVs.
Sensors 22 00270 g005
Figure 6. Energy transfer efficiency as a function of the number of users.
Figure 6. Energy transfer efficiency as a function of the number of users.
Sensors 22 00270 g006
Figure 7. Average throughput as a function of the energy arrival E m a x .
Figure 7. Average throughput as a function of the energy arrival E m a x .
Sensors 22 00270 g007
Figure 8. Average throughput as a function of the battery capacity C .
Figure 8. Average throughput as a function of the battery capacity C .
Sensors 22 00270 g008
Figure 9. Average throughput as a function of the energy transfer efficiency between two UAVs.
Figure 9. Average throughput as a function of the energy transfer efficiency between two UAVs.
Sensors 22 00270 g009
Figure 10. Convergence behavior for the reinforcement learning protocols.
Figure 10. Convergence behavior for the reinforcement learning protocols.
Sensors 22 00270 g010
Table 1. List of notations.
Table 1. List of notations.
NotationsDefinitions
M Number of UAVs
L Number of user sets
ρ Mean number of buildings per square kilometer
κ Scale parameter
α Fraction of area covered by buildings to the total area
L t Path loss
α L LOS path loss exponent
α N NLOS path loss exponent
C L Intercept of the LOS link
C N Intercept of the NLOS link
h ^ S Small-scale fading
N L Nakagami fading parameter for LOS link
N N Nakagami fading parameter for NLOS link
H i , j t Channel gain from the j-th UAV BS to a i-th ground user
G x r Directional antenna gain
G 0 Maximum antenna gain
θ c a Azimuth plane
θ c e Elevation plane
M c Mean lobe gain
m c Side lobe gain
γ i j t Signal-to-interference-plus-noise ratio from UAVj to user i
P i , j t Transmit power selected by UAVj
P m a x Maximum transmission power
I i , j Interference to UAVj
σ 2 Noise power level
N Total number of time slots
τ Te slot duration
E j Amount of harvested energy for UAVj
E m a x Maximum harvested energy
C Battery capacity of each UAVj
B j Battery state for UAVj
R i , j t Downlink   rate   of   user   i
W MmWave transmission bandwidth
U t Total throughput
j j   Energy transferred from UAVj to UAVj
β Energy transfer efficiency between two UAVs.
S State space
A j Action space
P State transition function
R j Reward function
Table 2. Simulation parameters.
Table 2. Simulation parameters.
ParametersValues
Number of UAVs4
Maximum flying altitude of UAVs100 m
Number of users12
Mean number of buildings per square kilometer is ρ 300/km2
Fraction of area covered by buildings to the total area α . 0.5
Scale parameter κ 20 m
LOS   intercept   C L 1.39
NLOS   intercept   C N 1.39
LOS   path   loss   exponent   α L 2
NLOS   path   loss   exponent   α N 3
LOS   Nakagami   fading   parameter   N L 3
NLOS   Nakagami   fading   parameter   N N 2
Available bandwidth W 1 GHz
Noise   figure   N F 10 dB
Noise   power   σ 2 170 + 10 l o g 10 ( W ) + N F dBm = −70 dBm
Transmission   power   P m a x (0,20) dBm
Battery capacity C4000 J
Energy   Arrival   E m a x (0,125) J
Energy transfer efficiency between two UAVs β 0.9
Number of episodes5000
Number of time slots per episode500
Batch size500
Replay memory size50,000
Learning rate for DQL 10 3
Learning rate for MADDPGActorCritic
10 4 10 3
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Domingo, M.C. Power Allocation and Energy Cooperation for UAV-Enabled MmWave Networks: A Multi-Agent Deep Reinforcement Learning Approach. Sensors 2022, 22, 270. https://doi.org/10.3390/s22010270

AMA Style

Domingo MC. Power Allocation and Energy Cooperation for UAV-Enabled MmWave Networks: A Multi-Agent Deep Reinforcement Learning Approach. Sensors. 2022; 22(1):270. https://doi.org/10.3390/s22010270

Chicago/Turabian Style

Domingo, Mari Carmen. 2022. "Power Allocation and Energy Cooperation for UAV-Enabled MmWave Networks: A Multi-Agent Deep Reinforcement Learning Approach" Sensors 22, no. 1: 270. https://doi.org/10.3390/s22010270

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop