Next Article in Journal
A Survey from Real-Time to Near Real-Time Applications in Fog Computing Environments
Previous Article in Journal
Multi-Tier Cellular Handover with Multi-Access Edge Computing and Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Interference Management in UAV-Based HetNets †

1
Department of Electrical and Computer Engineering, North Carolina State University, Raleigh, NC 27695, USA
2
Motorola Solutions, Inc., Plantation, FL 33322, USA
*
Author to whom correspondence should be addressed.
This paper is an extended version of our conference paper: Singh, S.; Kumbhar, A.; Güvenç, İ.; Sichitiu, M.L. Distributed Approaches for Inter-Cell Interference Coordination in UAV-Based LTE-Advanced HetNets. In Proceedings of the IEEE 88th Vehicular Technology Conference (VTC-Fall), Chicago, IL, USA, 27–30 August 2018.
Telecom 2021, 2(4), 472-488; https://doi.org/10.3390/telecom2040027
Submission received: 2 October 2021 / Revised: 17 November 2021 / Accepted: 17 November 2021 / Published: 24 November 2021

Abstract

:
Unmanned aerial vehicles (UAVs) can play a key role in meeting certain demands of cellular networks. UAVs can be used not only as user equipment (UE) in cellular networks but also as mobile base stations (BSs) wherein they can either augment conventional BSs by adapting their position to serve the changing traffic and connectivity demands or temporarily replace BSs that are damaged due to natural disasters. The flexibility of UAVs allows them to provide coverage to UEs in hot-spots, at cell-edges, in coverage holes, or regions with scarce cellular infrastructure. In this work, we study how UAV locations and other cellular parameters may be optimized in such scenarios to maximize the spectral efficiency (SE) of the network. We compare the performance of machine learning (ML) techniques with conventional optimization approaches. We found that, on an average, a double deep Q learning approach can achieve 93.46% of the optimal median SE and 95.83% of the optimal mean SE. A simple greedy approach, which tunes the parameters of each BS and UAV independently, performed very well in all the cases that we tested. These computationally efficient approaches can be utilized to enhance the network performance in existing cellular networks.

1. Introduction

An unmanned aerial vehicle (UAV) heterogeneous network (HetNet) consists of conventional stationary ground macro base stations (MBSs), supplemented by mobile UAV base stations (UABSs) and cells on wheels [1]. The agility of UAVs coupled with their ability to carry radios and communicate wirelessly has led their adoption in various applications to address network congestion and in public safety communications as a temporary substitute for damaged communication infrastructure. In the aftermath of hurricane Maria in 2017, ground base stations were destroyed and AT&T used UAVs to temporarily restore wireless voice, text, data, and multimedia services [2].
UABSs can complement existing MBSs in a UAV HetNet by providing wireless coverage to user equipment (UE) that are near cell borders, coverage holes, or far away from the MBS as shown in Figure 1. In this scenario, UABS 1 serves UE 5, while UABS 2 serves UE 6. These are the UEs that are far away from MBS 1, and thus these UABSs extends the coverage area of the network. On the other hand, though UABSs transmit at a lower power than MBSs, UABSs may cause interference to MBSs and vice-versa. Thus, the transmit powers of MBSs and UABSs needs to be adjusted accordingly to ensure sufficient signal to noise ratio (SINR) for all UEs.
As UABSs are generally powered by battery and have limited computational capabilities, any network optimization algorithms that are developed need to have low computational complexity. Additionally, the solutions should also have low time complexity so that the UABS can respond quickly to changes in cellular demand. Taking into account these constraints, we attempt to reduce the computational and time complexity of our proposed algorithms.
In this work, we develop computationally efficient algorithms to maximize the mean and median spectral efficiency (SE) in a UAV HetNet and evaluate the performance of these algorithms. This works extends a previous work written by the same authors [3]. In [3], we optimized the fifth percentile of spectral efficiency (5pSE) in a UAV-HetNet by employing a greedy algorithm and an artificial intelligence (AI) approach based on deep Q learning (DQN). In [3], it was observed that an AI approach, somewhat surprisingly, failed to find the optimum solution and was always out-performed by the greedy approach. In this work, we present an alternate AI solution, which is based on a double deep Q learning algorithm (DDQN) and uses a single AI agent to model all MBSs and UABSs in the UAV HetNet.
This single AI agent chooses the values of all relevant network parameters, unlike in [3], where each macro base station (MBS) and UAV base station (UABS) was modeled as a separate AI agent. Its performance is evaluated against the optimal exhaustive search and the computationally efficient sequential algorithm introduced in [3], which optimizes BS and UAV parameters in a greedy manner. Compared to [3], the performance of the AI approach is now closer to the optimal. Specifically, the AI algorithm can achieve 93.46 % of the optimal, when maximizing the median SE and 95.83 % of the optimal, when maximizing the mean SE.

2. Literature Review

Various studies in the literature have explored how UAVs can be used to supplement cellular networks. Approaches to calculate the best UABS position and trajectory have been studied in [4,5,6,7,8,9,10]. The authors in [4] calculated the best trajectory for a single UABS, assumed that the MBS and UABS operated at orthogonal frequencies, and thus the effect of interference was ignored. Greedy and unsupervised learning-based algorithms to position a fleet of UAVs to maximize aggregate user received power were presented in [5], while a sequential spiral algorithm was presented in [6].
However, none of these works considered interference between UAVs and BSs. Interference between UAVs and BSs was taken into account in [7,8]. Interference was minimized by using a statistical interference map and orthogonal frequency assignment in [7] to select optimal UAV locations, while multi-antenna beam-forming techniques were used in [8]. None of the studies leveraged LTE 3GPP Release-10/11 interference management techniques, nor any machine learning techniques. Machine learning techniques were used to predict potential network congestion and deploy UAVs accordingly in [10].
Interference between UAVs and BSs was assumed to be minimized using orthogonal frequency assignments and beamforming techniques rather than with LTE 3GPP Release-10/11 techniques. The authors in [11] followed the intuition that UAVs should be placed at coverage holes to maximize coverage, and used machine learning to locate coverage holes. A greedy approach to position UAVs was used in [12] to maximize the number of users covered, while ensuring that the QoS requirements of the users are satisfied.
Game-theory-based approaches were used in [13] The use of a UAV to provide wireless service to an Internet of Things (IoT) network was studied in [13], wherein IoT nodes harvest energy from a UAV before transmitting data on the uplink to the UAV. The nodes form coalitions with one node acting as the coalition head. The optimal trajectory of the UAV was calculated to maximize the energy available to the IoT coalition heads.
Novel approaches to optimally partition a geographical area into UABS and MBS cells were proposed in [14,15]. While [14] used optimal transport theory with the aim of minimizing the transmission delay for all users in the area, ref. [15] used a neural-based cost function in which user demand patterns are used to assign a cost and density to each area. Both studies, however, did not tackle interference mitigation challenges explicitly. Interference mitigation using 3GPP Release-10 enhanced inter-cell interference coordination techniques (eICIC) and Release-11 further enhanced inter-cell interference coordination (FeICIC) techniques in HetNets were studied in [16,17,18,19]. While [16] jointly optimized the ICIC parameters, the UE cell association rules, and the spectrum resource sharing between the macro and pico cells, it did not use 3GPP Release-11 FeICIC or cell range expansion (CRE) techniques.
Moreover, ref. [16] only studied LTE HetNets and not UAV HetNets. LTE UAV HetNets were also evaluated in [20], which optimized the allocation of LTE physical resource blocks in addition to the UAV position in order to maximize coverage. However, LTE interference management techniques were not utilized. In [17], the authors developed a stochastic-geometry-based framework to study and compare the effectiveness of 3GPP FeICIC techniques and eICIC techniques, but [17] also did not study UAV HetNets. The use of 3GPP Release-10/11 techniques along with UABS mobility in UAV HetNets was evaluated in [19]. However, this study did not individually optimize the 3GPP ICIC parameters, but rather applied the same ICIC parameter values to each MBS and UABS, which is sub-optimal, as we demonstrate. An overview of the existing literature that is related to our work is presented in Table 1.
To the best of our knowledge, low complexity approaches to optimize 3GPP Release-10/11 interference management parameters in UAV HetNets have not been studied in the literature. Our contribution is that we propose a greedy algorithm and a double deep Q learning-based algorithm, to individually optimize 3GPP Release-10/11 interference coordination parameters and UABS position in order to maximize the mean and median SE. We also compare these two computationally efficient algorithms with an optimal but computationally complex brute force algorithm.

3. System Model

We consider a HetNet with MBSs and UABSs operation in two tiers, within a simulation area of l × l square meters as shown in Figure 2. MBSs and UEs are randomly distributed in this area, according to a Poisson point process with the intensities λ mbs and λ ue , respectively. The number of MBSs and UEs in the simulation can be calculated as N mbs = λ mbs × l 2 and N ue = λ ue × l 2 , respectively. The 3D locations of all the MBSs, UABSs, and UEs are represented by the matrices X mbs R N mbs × 3 , X uabs R N uabs × 3 , and X ue R N ue × 3 , respectively. The transmission power of each MBSs is P mbs , while that of UABSs is P uabs . Given the antenna gains of MBS and UABS as K and K , respectively, the effective MBS and UABS transmission power is calculated as P mbs = K P mbs and P uabs = K P uabs , respectively.
We assume that the UABSs and MBSs exchange information over the X2 interfaces. We also consider that the downlink bandwidth available to an MBS or a UABS is shared equally among their served UEs. We assume that downlink data is always available for any UE—i.e., the downlink UE traffic buffers are always full. For an arbitrary UE n, where n { 1 , 2 , , N ue } , we define the macro-cell of interest (MOI) as the nearest MBS, and the UAV-cell of interest (UOI) as the nearest UABS.
For example, in the specific scenario shown in Figure 2, MBS 1 is the MOI for UEs 1 and 2, MBS 2 is the MOI for UEs 5 and 6, UABS 1 is the UOI of UE 2, and UABS 2 is the UOI of UE 3. We denote the reference symbol received power (RSRP) of the nth UE from the MOI and the UOI as S mbs ( d m n ) and S uabs ( d u n ) , respectively, where d m n is the distance from the nearest MOI, and d u n is the distance from the nearest UOI for the nth UE. We use the Okumura Hata suburban propagation model without any Rayleigh or Rician fading.
An arbitrary UE n is always assumed to connect to the nearest MBS or UABS, where n 1 , 2 , , N ue . Then, for the nth UE, the reference symbol received power (RSRP) from the macro-cell of interest (MOI) and the UAV-cell of interest (UOI) are given by
S mbs ( d m n ) = P mbs 10 φ / 10 , S uabs ( d u n ) = P uabs 10 φ / 10 ,
where φ is the path-loss observed from MBS in dB, φ is the path-loss observed from UABS in dB, d m n is the distance from the nearest MOI, and d u n is the distance from the nearest UOI.
The Okumura Hata path loss is a function of the carrier frequency, distance between the UE, serving cell, base station height, and UE antenna height. The path-loss (in dB) observed by the nth UE from MOI and UOI is given by:
φ = A + B log ( d m n ) + C ,
φ = A + B log ( d u n ) + C ,
where the distances d m n and d u n are in km, and the factors A, B, and C depend on the carrier frequency and antenna height. In a suburban environment, the factors A, B, and C are given by
A = 69.55 + 26.16 log ( f c ) 13.82 log ( h bs ) a ( h ue ) ,
B = 44.9 6.55 log ( h bs ) ,
C = 2 log ( f c / 28 ) 2 5.4 ,
where f c is the carrier frequency in MHz, h bs is the height of the base station in meters, and a ( h ue ) is the correction factor for the UE antenna height h ue in meters, which is defined as
a ( h ue ) = 1.1 log ( f c ) 0.7 h ue 1.56 log ( f c ) 0.8 .

3.1. SE with 3GPP Release-10/11 ICIC Techniques

In a HetNet, the MBSs transmit at higher powers and have higher ranges compared to the lower power UABSs. Nevertheless, the UABSs can extend their coverage and associate a larger number of UEs by using the cell range expansion (CRE) technique defined in 3GPP Release-8. The CRE of a UABS is defined as the factor by which UEs are biased to associate with that UABS. For example, in Figure 2, UABS 1 uses a CRE of 15 dB to force UE 4 to associate with itself. The use of CRE, however, results in increased interference to those UEs in the extended cell regions.
This interference from MBSs to UEs near the edge of range-extended UABS cells can be mitigated using time-domain-based ICIC techniques defined in 3GPP Release-10/11. These techniques require the MBS to transmit with reduced power during specific subframes on the physical downlink shared channel (PDSCH). Radio subframes with reduced power are termed coordinated subframes (CSF), and those with full power are termed uncoordinated subframes (USF).
We denote this power reduction factor by α where 0 α 1 . We note that α = 1 implies no ICIC, while eICIC techniques use α = 0 , and FeICIC techniques allow α to vary between 0 and 1. We use β to denote the USF duty cycle, and hence, the CSF duty cycle is given by ( 1 β ) . Figure 3 shows, for the scenario depicted in Figure 2, how MBS 1 and MBS 2 use power reduction factors, α 1 and α 2 respectively, to reduce interference to UE 3. We note that α 1 < α 2 , as MBS 2, being farther away from UE 3, can transmit at a higher power without degrading the performance of UE 3.
Individual MBSs or UABSs can schedule their UEs in USF or CSF based on the scheduling thresholds ρ and ρ , respectively. Then, a UE may be served either by an MOI or UOI and by the CSF or USF resources of the MOI/UOI, resulting in four different association categories. Let Γ denote the signal to interference ratio (SIR) at the MOI-USF, Γ denotes the SIR at the UOI-USF, and τ denotes the CRE that positively biases the UABS SIR to expand its coverage. Then, the four different resource categories where a UE may be scheduled can be summarized as follows. If Γ > τ Γ , we associate the UE with the MOI; otherwise, we schedule it with the UOI.
The intuition behind this condition is straightforward: associate the UE with the nearest base station that gives the best SIR, taking into account the CRE. On the other hand, if Γ > ρ or Γ ρ , we schedule the UE in CSF, and otherwise in USF. This condition is based on the following intuition: scheduling a UE in the CSF of an MOI degrades that UE’s SIR, whereas scheduling a UE in the CSF of a UOI improves that UE’s SIR. Thus, the “stronger” UEs that have sufficiently high SIRs and are close to the MBS should be scheduled in the CSFs of that MBS, as these “stronger” UEs can take the performance hit. Similarly, the “weaker” UEs that have low SIRs and are close to the cell edge of a UABS should be scheduled in the CSF of that UABS, as they need to be protected as a priority.
Using this framework of eICIC and FeICIC and following an approach similar to that of [17,22] for a HetNet, the SIR ( Γ , Γ csf , Γ , Γ csf ) and the SE ( C usf mbs , C csf mbs , C usf uabs , C csf uabs ) experienced by an arbitrary UE n can be defined for four different scenarios as follows:
( 1 ) UE associated with MOI and scheduled in USF:
Γ = S mbs ( d m n ) S uabs ( d u n ) + Z C usf mbs = β log 2 ( 1 + Γ ) N usf mbs .
( 2 ) UE associated with MOI and scheduled in CSF:
Γ csf = α S mbs ( d m n ) S uabs ( d u n ) + Z C csf mbs = ( 1 β ) log 2 ( 1 + Γ csf ) N csf mbs .
( 3 ) UE associated with UOI and scheduled in USF:
Γ = S uabs ( d u n ) S mbs ( d m n ) + Z C usf uabs = β log 2 ( 1 + Γ ) N usf uabs .
( 4 ) UE camped on UOI and scheduled in CSF:
Γ csf = S uabs ( d u n ) α S mbs ( d m n ) + Z C csf uabs = ( 1 β ) log 2 ( 1 + Γ csf ) N csf uabs .
In (8)–(11), Z and Z are, respectively, the interference at a UE from all the MBSs and the UABSs except the UOI and except the MOI. They have the same meaning but different values.

3.2. Performance Metrics

The objective of our study is to design algorithms to calculate the best FeICIC and eICIC parameters individually for all MBSs and UABSs as well as the positions of the UABSs to, thus, maximize our performance metric. We evaluate two performance metrics: the mean SE and median SE. The algorithm should find the best state, S , out of all possible states S such that:
S = arg C S max SE ( S ) ,
where C ( . ) denotes the function that calculates the mean or median SE over the whole network area for a given state S = [ X uabs , S mbs ICIC , S uabs ICIC ] . First, we calculate the SE for each UE as per Equations (1)–(4). As defined previously, X uabs is the matrix representing the location of the N uabs UABSs in three dimensions, S mbs ICIC = [ α , β , ρ ] R N mbs × 2 is a matrix that captures individual ICIC parameters for each MBS, and S uabs ICIC = [ τ , ρ ] R N uabs × 2 is a matrix that captures the individual ICIC parameters for each UABS. The vectors α = [ α 1 , , α N mbs ] T and ρ = [ ρ 1 , , ρ N mbs ] T capture the power reduction factors and scheduling thresholds, respectively, of each MBS. On the other hand, for each UABS, τ = [ τ 1 , , τ N uabs ] T and ρ = [ ρ 1 , , ρ N uabs ] T denote the CRE and scheduling threshold, respectively.

3.3. Parameter Optimization

The SE is a function of α , β , τ , ρ , ρ , and X uabs , which are our optimization space. Either these parameters can be optimized for individual MBSs and UABSs, or the same value of each parameter can be used for all MBS and UABS, which is sub-optimal but computationally less complex.
To show that optimizing the above parameters individually for each MBS and UABS gives a better performance than optimizing the ICIC parameters jointly, we consider the hypothetical situation depicted in Figure 2. Here, UE 3 is the critical UE to be protected from interference. Intuitively, as MBS 1 is closer to UE 3 compared to MBS 2, it is desirable for MBS 1 to transmit at a lower power during CSFs. Mathematically, α 1 = 0.4 < α 2 = 0.8 . As UE 4, which is served by UABS 2, is farther away from all the MBSs, UABS 2 does not have to use a large CRE to encourage UE 4 to associate with itself. In contrast, UABS 1 would have to use a larger CRE to encourage UE 3 to associate with itself. Mathematically, τ 1 = 15 dB > τ 2 = 0 dB. Therefore, optimizing the parameters individually for each MBS and UABS gives better performance.
The large size of the search space can be appreciated by referring to Table 2 and Table 3, which list the range and size of the parameters to be optimized. In this table, the parameters Δ α , Δ β , , Δ ρ , Δ ρ , Δ x , and Δ y denote the step sizes for α , β , ρ , ρ , x coordinate of a UABS’s location, and y coordinate of a UABS’s location, respectively, while ρ low and ρ low denote the lower bounds for ρ and ρ , respectively. Similarly, ρ high and ρ high denote the upper bounds for ρ and ρ , respectively. The actual size of the parameter space is depicted in Figure 4, which shows the number of possible states in S , over which the individual or joint optimization algorithms have to search, in order to find the best state, S .
Figure 4 shows that, as the geographical area and correspondingly the number of MBSs and UABSs increase, the size of the search space increases much more rapidly for individual optimization than for joint optimization. This behavior can be understood by calculating the number of possible permutations of α for two MBSs considered by joint and individual optimization approaches, assuming that α can take four different values from 0 to 1.
While the joint optimization approach will only compare the SE metric at these four values of α , the individual optimization approach will need to compare the SE metric for 4 2 = 16 different permutations of α values of the two MBSs. In this way, the parameter search size for joint optimization is agnostic to the number of MBSs or UABSs in the simulation area. The parameter search space of individual optimization, on the other hand, increases exponentially with number of MBSs and UABSs. We also observe that, for the individual optimization approach, the size of the state space exceeds a googol ( 10 100 ) of states as the simulation area increases beyond 15 km2.

4. UAV Location and Interference Management

We considered and compared the performance of three algorithms for FeICICs—a brute force algorithm, a greedy algorithm, and a DDQN learning algorithm. In this section, we present the brute force and greedy algorithms, while Section IV presents the DDQN learning algorithm.

4.1. Brute Force Algorithm

The brute force algorithm investigates the entire search state space of all possible values of UABS locations and individual eICIC and FeICIC parameters and returns the state with the best mean or median SE. This is illustrated in Algorithm 1. Since it individually searches for all possible parameter values, it is computationally infeasible for large areas and large number of MBSs and UABSs.
Algorithm 1 Brute force algorithm.
1:
Best state, S NULL
2:
bestSE 1
3:
for all State S do
4:
    currentSE C ( S )
5:
    if currentSE > bestSE then
6:
        bestSEcurrentSE
7:
         S S
8:
    end if
9:
end for

4.2. Greedy Algorithm

In order to reduce the time complexity, we utilized a heuristic algorithm that initially assumes that there is only one UABS in the system and finds the best location, eICIC, and FeICIC parameters for this UABS. It then considers that there are two UABSs in the system, with the parameters of the first UABS set to those found earlier, and finds the best parameters for this second UABS. Similarly, it then finds the best parameters for the third UABS considering the first two UABSs to be at their earlier determined states. The algorithm continues this procedure for the given number of UABSs.
A similar approach is used to optimize the parameters of MBSs. This algorithm is summarized in Algorithm 2. As we will see in the simulation results, this algorithm, somewhat surprisingly, performs close to the brute force search, and outperforms the DDQN algorithm, to be discussed in the next section. A main reason for this is that the large number of parameters in Table 1 allows to compensate for a non ideal selection of parameters in earlier stages, e.g., by tuning the scheduling thresholds and CRE, non-ideal location of a UABS can still result in good SE.
Algorithm 2 Greedy algorithm.
1:
for allUABSdo
2:
    Assume all previous UABS to be positioned at their best location and       operating at their best ICIC parameters.
3:
     Find best location and FeICIC parameters for current UABS.
4:
end for
5:
for allMBSdo
6:
     Assume all previous MBS to be operating at their best ICIC parameters.
7:
     Find best ICIC parameters for current MBS.
8:
end for

5. Machine Learning Approach for UAV ICIC and Placement

5.1. Q Learning

The Q learning algorithm is one of the commonly used reinforcement learning (RL) algorithms [23]. In reinforcement learning, an agent interacts with an environment by taking actions in different states and observing the costs or rewards of the actions. The agent starts out with random actions and. eventually, by observing the rewards and by exploring different states, it learns the best action to take in each state in order to optimize the cumulative reward. Figure 5 shows the interaction of the MBS and the UABS agent with the UAV HetNet environment considered in this paper.
Since RL and deep-learning techniques are relevant and are known to be effective in solving problems with large search spaces, it is instructive to investigate their performance in solving our UAV HetNet optimization problem. Q learning allows the agents to act and learn as follows: being in a state s after selecting action a, and receiving the immediate cost c, a UABS or MBS agent n updates its knowledge Q n ( s , a ) for the particular state-action pair through the following operation:
Q n ( s , a ) ( 1 η ) Q n ( s , a ) + η [ c n + λ min a Q n ( s , a ) ] ,
where η , the learning rate, is the agent’s willingness to learn from the UAV HetNet environment, λ is the discount factor, and s is the next state. Lower values of λ give more importance to immediate rewards. We note that, conventionally, α , rather than η , is used to denote the agent’s learning rate. We use η , as, in our work, we have already used α to denote the ICIC power reduction factor. After the Q learning agent has been trained, and when it is being evaluated, the exploration–exploitation factor, ϵ , determines the probability of the agent choosing a random action at a given state, (i.e., the agent explores), rather than choosing the optimal action, as learned by it (i.e., the agent exploits).
The state space is a sufficient representation of the environment at a point in time, containing all the information required by the agent to choose its next action. In our case, the state space is the set:
[ X ue , X mbs , X uabs , S mbs ICIC , S uabs ICIC ] ,
which captures the value of each ICIC parameter for each UABS and MBS, and also the location of each MBS, UABS and UE. The action space is the union of the possible actions for UABSs and MBSs in the system. At each state, along with its position, a UABS may choose any value for its ICIC parameters, τ and ρ . An MBS, similarly, may choose any value for its ICIC parameters, α , β and ρ .
Specifically, an MBS’s action space is the following set:
{ α , β , ρ } | α { 0 , Δ α , 2 Δ α , 1 } , β { 0 , Δ β , 2 Δ β , 1 } , ρ { ρ low , ρ low + Δ ρ , , ρ high } ,
and a UABS’s action space is the set:
{ τ , ρ , δ x , δ y } | τ { 0 , Δ τ , τ high + Δ τ , τ high } , ρ { ρ low , ρ low + Δ ρ , ρ high } , x uabs { l / 2 , l / 2 + Δ x , l / 2 + 2 Δ x , l / 2 } , y uabs { l / 2 , l / 2 + Δ x , l / 2 + 2 Δ x , l / 2 } ,
where x uabs and y uabs represent the X and Y coordinate, respectively, of the UABS, as a result of the action. Our chosen reward function is the median or mean SE.
As conventional Q learning algorithms maintain a Q table, with states as rows and actions as columns and with each cell representing the Q value of a specific action in a specific state, they cannot handle infinite state spaces, as the size of the Q table would become infinite. The Q table would take an extremely long duration to converge. Conventional Q learning suffers from issues of memory complexity, computational complexity, and sample complexity [24]. As illustrated in Figure 4, the parameter space for our scenario can become extremely large, and therefore a conventional Q learning approach is not feasible. Deep learning, relying on the powerful functional approximation and representational learning properties of deep neural networks, provides us with the tools for overcoming these problems [25].

5.2. Deep Q Learning

DQN, hailed as the first step towards general AI—an AI that can survive in a variety of environments, came to the forefront of machine learning when it was used by DeepMind and Google to train an agent to achieve professional level scores on 49 different Atari 2600 games [26]. DQN extends Q learning by using a neural network to model the Q function, instead of using the simple Q table. One approach is to design the network to accept the state and action as the inputs and provide the corresponding Q value as the output. Another approach is for the network to accept the current state as the input and provide the Q value of each possible action as the output.
The latter approach, used in [26], is found to be better as only a single forward pass through the network is needed when we want to do a Q value update or pick the action with the highest Q value [27]. This is the architecture that we use as well. This neural network can handle infinite state spaces and also recognize common patterns between similar states. Another advantage of deep Q learning is experience replay—the neural network is retrained after each action step, enabling the agent to adapt to changes in the environment. Thus, the deep Q learning agent can learn from its experience continuously.

5.3. Double Deep Q Learning (DDQN)

In our simulations, we use the double DQN (DDQN) [28] architecture, which leads to better policy evaluation in the presence of many similar-valued actions and reduces over-estimation in action values compared to DQN. The DDQN approach decouples the selection of an action from the evaluation of an action. The Q-learning equation in (13) is then modified as:
Q n ( s , a ) ( 1 η ) Q n ( s , a ) + η c n + λ Q n ( s , min a Q n ( s , a ) ) ,
where Q n is the evaluation network that is used to evaluate the policy, while Q n is the online network that is used to select the optimal actions. All MBSs and UABSs in the UAV HetNet are modeled by a single AI agent, which chooses the values of all relevant network parameters. The neural network parameters are listed in Table 4, while our DDQN approach is presented in Algorithm 10.
Algorithm 3 DDQN learning for UAV HetNets.
1:
while ( R e a l i z a t i o n n u m R e a l i z a t i o n s ) do
2:
   Intialize a new realization of the environment
3:
   while ( N u m b e r O f S t e p s n u m S t e p s ) do
4:
       The agent acts, observes the reward and the new state.
5:
       Update the weights of the model neural network
6:
       Update the weights of the target neural network
7:
     end while
8:
     Choose the best state encountered so far
9:
     Preserve the neural network weights
10:
end while

6. Simulation Results

The ICIC and UAV placement algorithms were evaluated in a UAV HetNet, and the parameters are as defined in Table 5. The DDQN-based AI algorithm was implemented using Intel RL Coach [29]. Intel RL Coach is a python framework, and it implements many state-of-the-art algorithms. These algorithms can be used through a set of application programming interfaces (APIs). The developer defines the environment and the optimization problem and calls the Intel RL Coach APIs to solve the problem. The UAV HetNet environment was created using python scripts, while Matlab scripts were used to simulate the path-loss model, associate UEs to BSs, and calculate the SE for a particular realization.
The agent was trained over 300 realizations and then evaluated over 250 realizations, where the location of UEs and BSs in each realization were generated using a random Poisson point process (PPP) with UE and MBS densities as indicated in Table 5. The CDF of the median and mean SE achieved by the algorithms, calculated over multiple realizations, are shown in Figure 6a,b, respectively, when the density of MBSs is 4 MBSs per km 2 , resulting in one MBS in the simulation area. Blue curves denote the CDF obtained using DDQN, red when using the greedy algorithm, and black when using the optimal exhaustive search.
It can be seen that the DDQN algorithm performs better when the value of the exploration–exploitation trade-off ( ϵ ) during evaluation is 0.5 (achieving 96.56 % of the optimal median SE and 98.31 % of the optimal mean SE), compared to when ϵ is 0.05 (achieving 89.1 % of the optimal mean SE and 82.85 % of the optimal median SE). Using a higher ϵ implies that the agent takes more random actions, compensating for under-training and helping it to come out of local minima.
For 8 MBSs per km 2 (two MBSs in the simulation area), the CDF of median and mean SE achieved by the algorithms are shown in Figure 7a,b, respectively. Again, the DDQN performance is better when using ϵ = 0.5 (achieving 93.46 % of the optimal median SE and 95.83 % of the optimal mean SE), than when using an ϵ = 0.05 (achieving 79.73 % of the optimal median SE and 82.89 % of the optimal mean SE). The greedy algorithm achieves the optimal results in the scenarios that we consider. The time complexity of the exhaustive search is exponential in the number of MBSs and UABSs and is given by:
C time = O ( 1 Δ α + 1 N mbs × ( 1 / Δ β + 1 ) N mbs × ρ high ρ low Δ ρ N mbs × ρ high ρ low Δ ρ N uabs × l 2 Δ x Δ y N uabs ) .
Unlike the brute force algorithm, the greedy algorithm has linear time complexity, expressed as:
O ( N mbs ) + O ( N uabs ) .
The time complexity of the DDQN approach during training is a function of the number of training steps,
O ( N train ) ,
where N train is the number of training steps. However, after training is complete and the resultant neural network is used to optimize UAV HetNet parameters, the time complexity is the number of steps in the evaluation phase:
O ( N eval ) ,
where N eval is the number of evaluation steps. To optimze all parameters for a realization consisting of two MBSs and one UABS, considering the range of parameters given in Table 5, the brute force algorithm searches over 409 , 600 parameter combinations, the greedy approach evaluates 464 parameter combinations, while the searches over 500 parameter combinations, as per the policy learned after training.

7. Conclusions

We studied how 3GPP LTE FeICIC parameters can be tuned in a UAV-HetNet to maximize the mean and median spectral efficiency of a cellular network. Computationally efficient AI and greedy approaches were compared with an optimal exhaustive search. The AI approach was implemented using DDQN RL with a single agent to model all UAVs and BSs in the network. We observed that, for the scenarios that we considered, the greedy algorithm achieved the optimal mean and median SE, while the AI approach achieved 93.46 % of the optimal median SE and 95.83 % of the optimal mean SE.

Author Contributions

Formal analysis, S.S., A.K., İ.G., M.L.S.; Investigation, S.S., A.K., İ.G., M.L.S.; Methodology, S.S., A.K., İ.G., M.L.S.; Writing—original draft, S.S.; Writing—review & editing, İ.G. and M.L.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by NSF under the grant numbers CNS-1453678 and CNS-1738093.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Technopedia. Cell on Wheels (COW). Available online: https://www.techopedia.com/definition/26001/cell-on-wheels-cow (accessed on 22 November 2021).
  2. Geib, C. An AT & T Drone Is Connecting Puerto Ricans to Wireless Service. Available online: https://futurism.com/att-drone-connecting-puerto-ricans-wireless-service/ (accessed on 22 November 2021).
  3. Singh, S.; Kumbhar, A.; Güvenç, I.; Sichitiu, M.L. Distributed Approaches for Inter-cell Interference Coordination in UAV-based LTE-Advanced Hetnets. In Proceedings of the 88th Vehicular Technology Conference (VTC-Fall), Chicago, IL, USA, 27–30 August 2018; pp. 1–6. [Google Scholar]
  4. Lyu, J.; Zeng, Y.; Zhang, R. Spectrum sharing and cyclical multiple access in UAV-aided cellular offloading. In Proceedings of the Global Communications Conference (GLOBECOM), Singapore, 4–8 December 2017; pp. 1–6. [Google Scholar]
  5. Sharafeddine, S.; Islambouli, R. On-demand deployment of multiple aerial base stations for traffic offloading and network recovery. Comput. Netw. 2019, 156, 52–61. [Google Scholar] [CrossRef] [Green Version]
  6. Lyu, J.; Zeng, Y.; Zhang, R.; Lim, T.J. Placement optimization of UAV-mounted mobile base stations. IEEE Commun. Lett. 2016, 21, 604–607. [Google Scholar] [CrossRef] [Green Version]
  7. Li, Y.; Cai, L. UAV-Assisted Dynamic Coverage in a Heterogeneous Cellular System. IEEE Netw. 2017, 31, 56–61. [Google Scholar] [CrossRef]
  8. Hanna, S.; Yan, H.; Cabric, D. Distributed UAV Placement Optimization for Cooperative Line-of-sight MIMO Communications. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 4619–4623. [Google Scholar] [CrossRef] [Green Version]
  9. Fouda, A.; Ibrahim, A.S.; Güvenç, Í.; Ghosh, M. Interference management in UAV-assisted integrated access and backhaul cellular networks. IEEE Access 2019, 7, 104553–104566. [Google Scholar] [CrossRef]
  10. Zhang, Q.; Mozaffari, M.; Saad, W.; Bennis, M.; Debbah, M. Machine Learning for Predictive On-Demand Deployment of Uavs for Wireless Communications. In Proceedings of the IEEE Global Communications Conference (GLOBECOM), Abu Dhabi, United Arab Emirates, 9–13 December 2018; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  11. Al-Ahmed, S.A.; Shakir, M.Z.; Zaidi, S.A.R. Optimal 3D UAV base station placement by considering autonomous coverage hole detection, wireless backhaul and user demand. J. Commun. Netw. 2020, 22, 467–475. [Google Scholar] [CrossRef]
  12. Huang, H.; Savkin, A.V.; Ding, M.; Kaafar, M.A. Optimized deployment of drone base station to improve user experience in cellular networks. J. Netw. Comput. Appl. 2019, 144, 49–58. [Google Scholar] [CrossRef]
  13. Sikeridis, D.; Tsiropoulou, E.E.; Devetsikiotis, M.; Papavassiliou, S. Wireless powered Public Safety IoT: A UAV-assisted adaptive-learning approach towards energy efficiency. J. Netw. Comput. Appl. 2018, 123, 69–79. [Google Scholar] [CrossRef]
  14. Mozaffari, M.; Saad, W.; Bennis, M.; Debbah, M. Optimal transport theory for cell association in UAV-enabled cellular networks. IEEE Commun. Lett. 2017, 21, 2053–2056. [Google Scholar] [CrossRef]
  15. Sharma, V.; Bennis, M.; Kumar, R. UAV-assisted heterogeneous networks for capacity enhancement. IEEE Commun. Let. 2016, 20, 1207–1210. [Google Scholar] [CrossRef] [Green Version]
  16. Deb, S.; Monogioudis, P.; Miernik, J.; Seymour, J.P. Algorithms for enhanced inter-cell interference coordination (eICIC) in LTE HetNets. IEEE/ACM Trans. Netw. 2014, 22, 137–150. [Google Scholar] [CrossRef] [Green Version]
  17. Merwaday, A.; Mukherjee, S.; Güvenç, I. Capacity analysis of LTE-Advanced HetNets with reduced power subframes and range expansion. EURASIP J. Wirel. Commun. Netw. 2014, 2014, 189. [Google Scholar] [CrossRef]
  18. Mukherjee, S.; Güvenç, I. Effects of range expansion and interference coordination on capacity and fairness in heterogeneous networks. In Proceedings of the 2011 Conference Record of the Forty Fifth Asilomar Conference on Signals, Systems and Computers (ASILOMAR), Pacific Grove, CA, USA, 6–9 November 2011; pp. 1855–1859. [Google Scholar]
  19. Kumbhar, A.; Binol, H.; Singh, S.; Guvenc, I.; Akkaya, K. Heuristic approach for jointly optimizing FeICIC and UAV locations in multi-tier LTE-advanced public safety HetNet. arXiv 2019, arXiv:2001.02760. [Google Scholar]
  20. Fahim, A.; Gadallah, Y. An Optimized LTE-Based Technique for Drone Base Station Dynamic 3D Placement and Resource Allocation in Delay-Sensitive M2M Networks. IEEE Trans. Mob. Comput. 2021. [Google Scholar] [CrossRef]
  21. Mozaffari, M.; Saad, W.; Bennis, M.; Debbah, M. Mobile unmanned aerial vehicles (UAVs) for energy-efficient Internet of Things communications. IEEE Trans. Wirel. Commun. 2017, 16, 7574–7589. [Google Scholar] [CrossRef]
  22. Kumbhar, A.; Guvenc, I.; Singh, S.; Tuncer, A. Exploiting LTE-Advanced HetNets and FeICIC for UAV-assisted Public Safety Communications. IEEE Access 2017, 6, 783–796. [Google Scholar] [CrossRef]
  23. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An introduction; MIT Press: Cambridge, MA, USA, 1998; Volume 1. [Google Scholar]
  24. Strehl, A.L.; Li, L.; Wiewiora, E.; Langford, J.; Littman, M.L. PAC model-free reinforcement learning. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; Volume 23, pp. 881–888. [Google Scholar]
  25. Arulkumaran, K.; Deisenroth, M.P.; Brundage, M.; Bharath, A.A. A brief survey of deep reinforcement learning. arXiv 2017, arXiv:1708.05866. [Google Scholar] [CrossRef] [Green Version]
  26. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
  27. Matiisen, T. Guest Post (Part I): Demystifying Deep Reinforcement Learning. Available online: https://ai.intel.com/demystifying-deep-reinforcement-learning/ (accessed on 22 November 2021).
  28. Van Hasselt, H.; Guez, A.; Silver, D. Deep Reinforcement Learning with Double Q-Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. [Google Scholar]
  29. Intel. Reinforcement Learning Coach by Intel AI Labs. Available online: https://github.com/IntelLabs/coach (accessed on 22 November 2021).
Figure 1. A UAV-assisted HetNet composed of UABSs and an MBS where certain UEs are offloaded to UABSs to improve the coverage and fairness.
Figure 1. A UAV-assisted HetNet composed of UABSs and an MBS where certain UEs are offloaded to UABSs to improve the coverage and fairness.
Telecom 02 00027 g001
Figure 2. Illustration of a UAV HetNet where the ICIC parameters of all MBSs and UABSs, and the locations of UABSs are optimized individually. This approach is in contrast to some existing work, such as [22], which, while optimizing locations for each UABS individually, assigns the same ICIC parameters to all UABSs and MBSs, Reprinted with permission from Ref. [3]. Copyright 2018 IEEE.
Figure 2. Illustration of a UAV HetNet where the ICIC parameters of all MBSs and UABSs, and the locations of UABSs are optimized individually. This approach is in contrast to some existing work, such as [22], which, while optimizing locations for each UABS individually, assigns the same ICIC parameters to all UABSs and MBSs, Reprinted with permission from Ref. [3]. Copyright 2018 IEEE.
Telecom 02 00027 g002
Figure 3. 3GPP Release-11 FeICIC with reduced power, almost blank subframes, and different power reduction factors for each MBS for the scenario in Figure 2.
Figure 3. 3GPP Release-11 FeICIC with reduced power, almost blank subframes, and different power reduction factors for each MBS for the scenario in Figure 2.
Telecom 02 00027 g003
Figure 4. Increase in the parameter search space size as a function of the total simulation area.
Figure 4. Increase in the parameter search space size as a function of the total simulation area.
Telecom 02 00027 g004
Figure 5. Interaction of the MBS or the UABS agent with the UAV HetNet environment for Q-learning.
Figure 5. Interaction of the MBS or the UABS agent with the UAV HetNet environment for Q-learning.
Telecom 02 00027 g005
Figure 6. SE results for an MBS density of 4 MBS per km 2 , which corresponds to one MBS in the simulation area. (a) CDF of median SE. DDQN achieves 96.56 % of the optimal when ϵ = 0.5 and 82.85 % of the optimal when ϵ = 0.05 . (b) CDF of mean SE. DDQN achieves 98.31 % of the optimal when ϵ = 0.5 and 89.1 % of the optimal when ϵ = 0.05 .
Figure 6. SE results for an MBS density of 4 MBS per km 2 , which corresponds to one MBS in the simulation area. (a) CDF of median SE. DDQN achieves 96.56 % of the optimal when ϵ = 0.5 and 82.85 % of the optimal when ϵ = 0.05 . (b) CDF of mean SE. DDQN achieves 98.31 % of the optimal when ϵ = 0.5 and 89.1 % of the optimal when ϵ = 0.05 .
Telecom 02 00027 g006
Figure 7. SE results for an MBS density of 8 MBS per km 2 , which corresponds to two MBSs in the simulation area. (a) CDF of median SE. DDQN achieves 93.46 % of the optimal when ϵ = 0.5 and 79.73 % of the optimal when ϵ = 0.05 . (b) CDF of mean SE. DDQN achieves 95.83 % of the optimal when ϵ = 0.5 and 82.89 % of the optimal when ϵ = 0.05 .
Figure 7. SE results for an MBS density of 8 MBS per km 2 , which corresponds to two MBSs in the simulation area. (a) CDF of median SE. DDQN achieves 93.46 % of the optimal when ϵ = 0.5 and 79.73 % of the optimal when ϵ = 0.05 . (b) CDF of mean SE. DDQN achieves 95.83 % of the optimal when ϵ = 0.5 and 82.89 % of the optimal when ϵ = 0.05 .
Telecom 02 00027 g007
Table 1. Comparison with related works.
Table 1. Comparison with related works.
Ref.Applicable to HetNetsUAVs Used as BSs3GPP Interference ManagementOptimization VariablesOptimization GoalOptimization Algorithm(s)
[4]🗸🗸🞨Bandwidth allocated to the UAV, user-cell association, radius of UAV’s trajectoryMaximize the minimum throughput across all usersBisection search
 [5]🗸🗸🞨Number and position of UAVsMinimize the number of UAVs, and maximize aggregate user received powerGreedy approach, a novel unsupervised learning approach
 [6]🞨🗸🞨Number and position of UAVsMinimize the number of UAVsA sequential algorithm that places UAVs along a spiral
 [8]🗸🗸🞨UAV positionMinimize the distance travelled by each UAV while maximizing the LoS MIMO channel capacityExhaustive search, gradient descent
 [9]🗸🗸🞨User-cell associations, downlink power allocations and UAV positionMaximize network sum rate, subject to a constraint on received SINRHybrid fixed-point iteration, particle swarm optimization
[10]🗸🗸🞨Location and area served by each UAVMinimize total downlink powerMachine learning framework based on a Gaussian mixture model and weighted expectation maximization
[14]🗸🗸🞨Area served by the BS and the UAVMinimize mean user transmission delayOptimal transport theory
[16]🗸🞨🗸Sub-frame radio resource sharing among macro BSs and pico cells, user-cell associationMaximize weighted proportional proportional fair throughputDual-based approach to solve a relaxed NLP followed by integer rounding
[17]🗸🞨🗸3GPP FeICIC parametersMaximize aggregate capacity and proportional fairness among usersExhaustive search
 [19]🗸🗸🗸3GPP FeICIC parameters. Parameters not optimized individually for each BS and UAV.Maximize fifth percentile of spectral efficiency (SE)Exhaustive search, genetic algorithm, and elitist harmony search
[13]🗸🗸🞨UAV trajectory and uplink transmission of each Internet of Things (IoT) nodeMaximize energy availability of IoT coalition headsExhaustive search for UAV trajectory and non-cooperative game theory-based approach to calculate uplink transmission power
[21]🗸🗸🞨Trajectory of multiple UAVs, association of devices to UAVs, and uplink powerMinimize total uplink transmit powerMain optimization problem decomposed into two sub-problems, which are solved iteratively together
 [20]🗸🗸🞨UAV position and allocation of LTE physical resource blocksMaximize number of users covered, while satisfying their delay requirementsA heuristic-based approach, which achieves near-optimal performance
 [12]🗸🗸🞨UAV positionMaximize number of users covered, while satisfying their QoS requirementsGreedy approach
 [11]🗸🗸🞨UAV positionMaximize number of users coveredReinforcement learning used to discover network coverage holes, where UAVs are then positioned
This work🗸🗸🗸3GPP FeICIC parameters. Parameters optimized individually for each BS and UAV.Mean and median SEDouble deep Q learning, greedy approach, exhaustive search
Table 2. UABS parameters and search space.
Table 2. UABS parameters and search space.
ParameterRangeSearch Space Size
UABS scheduling threshold ( ρ ) ρ low , ρ low + Δ ρ , ρ low + 2 Δ ρ ρ high ( ρ high ρ low ) ( Δ ρ )
Cell range expansion ( τ ) 0 , Δ τ , 2 Δ τ , τ high τ high Δ τ
X coordinate of UABS l / 2 , l / 2 + Δ x , l / 2 + 2 Δ x , l / 2 l Δ x
Y coordinate of UABS l / 2 , l / 2 + Δ y , l / 2 + 2 Δ y , l / 2 l Δ y
Table 3. MBS parameters and search space.
Table 3. MBS parameters and search space.
ParameterRangeSearch Space Size
MBS power reduction factor ( α ) 0 , Δ α , 2 Δ α , 1 1 / Δ α + 1
MBS duty cycle ( β ) 0 , Δ β , 2 Δ β , 1 1 / Δ β + 1
MBS scheduling threshold ρ ρ low , ρ low + Δ ρ , ρ low + 2 Δ ρ ρ high ( ρ high ρ low ) ( Δ ρ )
Table 4. Parameters relevant to the DDQN RL algorithm.
Table 4. Parameters relevant to the DDQN RL algorithm.
LayerNumber of Neurons
Input layerSize of state space
Middle layer512, densely connected
Output LayerSize of action space
Table 5. Simulation parameters.
Table 5. Simulation parameters.
ParameterValue
MBS and UE density8 or 4 per km 2 and 100 per km 2
MBS and UABS transmit powers 46 dBm and 30 dBm
Path-loss exponent4
Altitude of UABSs 121.92 m ( 400 feet )
Simulation area 0.5 × 0.5 km 2
Range expansion bias in dB0 to 15
Power reduction factor for MBS during ( α )0 to 1
Duty cycle for the transmission of USF ( β )0 to 1
Scheduling threshold for UEs served by MBSs ( ρ )35 dB or 45 dB
Scheduling threshold for UEs served by UABSs ( ρ ) 20 dB to 5 dB
Downlink frequency763 MHz
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Singh, S.; Kumbhar, A.; Güvenç, İ.; Sichitiu, M.L. Intelligent Interference Management in UAV-Based HetNets. Telecom 2021, 2, 472-488. https://doi.org/10.3390/telecom2040027

AMA Style

Singh S, Kumbhar A, Güvenç İ, Sichitiu ML. Intelligent Interference Management in UAV-Based HetNets. Telecom. 2021; 2(4):472-488. https://doi.org/10.3390/telecom2040027

Chicago/Turabian Style

Singh, Simran, Abhaykumar Kumbhar, İsmail Güvenç, and Mihail L. Sichitiu. 2021. "Intelligent Interference Management in UAV-Based HetNets" Telecom 2, no. 4: 472-488. https://doi.org/10.3390/telecom2040027

Article Metrics

Back to TopTop