Next Article in Journal
Model Predictive Control Design and Hardware in the Loop Validation for an Electric Vehicle Powertrain Based on Induction Motors
Previous Article in Journal
A Review of Research on Spectrum Sensing Based on Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Energy Efficient Power Allocation in Massive MIMO Based on Parameterized Deep DQN

Department of Electrical and Computer Engineering, Ajou University, Suwon 16499, Republic of Korea
*
Authors to whom correspondence should be addressed.
Electronics 2023, 12(21), 4517; https://doi.org/10.3390/electronics12214517
Submission received: 29 September 2023 / Revised: 30 October 2023 / Accepted: 30 October 2023 / Published: 2 November 2023
(This article belongs to the Special Issue Deep Reinforcement Learning and Its Latest Applications)

Abstract

:
Machine learning offers advanced tools for efficient management of radio resources in modern wireless networks. In this study, we leverage a multi-agent deep reinforcement learning (DRL) approach, specifically the Parameterized Deep Q-Network (DQN), to address the challenging problem of power allocation and user association in massive multiple-input multiple-output (M-MIMO) communication networks. Our approach tackles a multi-objective optimization problem aiming to maximize network utility while meeting stringent quality of service requirements in M-MIMO networks. To address the non-convex and nonlinear nature of this problem, we introduce a novel multi-agent DQN framework. This framework defines a large action space, state space, and reward functions, enabling us to learn a near-optimal policy. Simulation results demonstrate the superiority of our Parameterized Deep DQN (PD-DQN) approach when compared to traditional DQN and RL methods. Specifically, we show that our approach outperforms traditional DQN methods in terms of convergence speed and final performance. Additionally, our approach shows 72.2% and 108.5% improvement over DQN methods and the RL method, respectively, in handling large-scale multi-agent problems in M-MIMO networks.

1. Introduction

With the increasing demand for mobile communications and Internet of Things technologies, wireless networks are facing increased data traffic and resource management issues owing to the rapid growth of wireless applications. Fifth-generation cellular networks have gained considerable attention for achieving spectrum efficiency and storage capacity. Massive multiple-input multiple-output (M-MIMO) networks are reliable options to overcome data storage and capacity issues to satisfy diverse user requirements. The main concept in M-MIMO technology is to equip the base stations (BSs) with a large number (i.e., 100 or more) of wireless antennas to simultaneously serve numerous users, enabling significant improvement in spectrum efficiency [1,2].
The presence of a huge number of antennas in M-MIMO, data multiplexing, and management would make MIMO transceiver optimization more challenging compared to single-antenna networks. The multi-objective nature of M-MIMO transceivers has resulted in various optimization strategies being performed in the past, including user association [3], power allocation [4], and user scheduling [5]. A joint user association and resource allocation problem was investigated in [6,7,8]. Given the non-convex multiple objective function in the M-MIMO problem, achieving the Pareto optimal solution set in a multi-objective environment becomes more challenging. Recently, proposed methods to solve multi-objective problems include approaches based on linear programming [9], game-theory [10], and Markov approximation [11,12]. Success of these methods requires complete knowledge of the system, which is rarely available.
Thus, emerging machine learning (ML) is an efficient tool to solve such complex multi-objective problems. In this ML field, Reinforcement Learning (RL) is the most appropriate branch to solve a non-convex problem. In RL-based optimization methods, three major elements (i.e., agents, reward, and action) of the proposed solution enable the self-learning abilities from the environment.
The Q-learning algorithm is one of the widely used RL methods because it requires minimal computation. It can be expressed using single equations and does not need to know the state transition probability. The RL agents maximize the long-term rewards over the current optimal reward function [13,14] using a Q-learning algorithm [14,15,16]. The agents are free to change their actions independently in a single-agent RL method, leading to a fluctuation in the overall action space, as well as action and rewards of the different agents in the process [16]. Q-learning methods have been used for power and resource allocation in heterogeneous and cellular networks [17]. However, it may be considerably difficult to handle such large state and action spaces in M-MIMO systems using Q-learning methods. To handle these issues of RL methods, deep reinforcement learning (DRL) methods are coupled with deep learning and RL to enhance the performance of RL for large-scale scenario problems. Nowadays, DRL methods [18] are promising to handle these complicated objective functions. DRL methods have already been applied to several tasks, such as resource allocation, fog radio access networks, dynamic channels, access, and mobile computing [19,20,21].
In DRL, the deep Q-network (DQN) method is mostly employed to train the agents to achieve an optimal scheme from a large state and action space. In [22], Rahimi et al. gave an algorithm of DQN, which was based on deep neural networks and has been previously used in the literature. Zhao et al. used the DRL method for the efficient management of user association and resource allocation for maximizing the network utility and maintaining the quality of service (QoS) requirements [23]. In [24,25,26,27], the authors proposed a DQN algorithm to allocate power using a multi-agent DRL method. Recent advancements in DRL, particularly techniques like Deep Q-Networks (DQNs), have opened up new avenues for addressing resource allocation challenges in wireless networks. However, when it comes to applying DQNs to solve the combined problem of power allocation and user association, a critical step involves converting the continuous action space for power allocation into a discrete action space. This quantization process can potentially result in suboptimal power allocation decisions, limiting overall performance. Additionally, the complexity of a DQN grows significantly as the dimensions of the action space increase. This exponential complexity can lead to high power consumption and slow convergence rates, which are highly undesirable in practical applications. The motivation for our research stems from the evolving landscape of wireless communication networks, particularly resource allocation in massive MIMO systems. As the demand for high-speed, reliable, and energy-efficient wireless transmission increases, optimizing resource allocation becomes dominant. This challenge becomes even more noticeable when dealing with complex circumstances that involve both discrete and continuous decision-making. EE has appeared as a critical metric in modern wireless networks, as it directly effects sustainability and operational costs. Achieving high EE while ensuring seamless user association and power allocation is a complicated problem that necessitates advanced solutions. Our motivation, therefore, lies in addressing this pressing need for more efficient and sustainable wireless networks. We are driven by the potential to make a substantial impact on the way resources are managed, energy is conserved, and user satisfaction is enhanced. By introducing the PD-DQN algorithm and applying it to the joint user association and power allocation problem, we aim to contribute valuable insights and practical solutions to the field of wireless communication.
We are motivated by the vision of network operators and engineers with a versatile tool that can adapt to the workings of hybrid action spaces, ultimately leading to more high-performing wireless networks. To address these challenges, our paper introduces the use of Parameterized Deep Q-Network (PD-DQN) techniques, which deal with parameterized state spaces. However, it falls short in terms of estimation capabilities and tends to produce suboptimal policies due to its tendency to overestimate Q-values [28]. Our approach outperforms traditional DQN and RL methods with significant improvements in convergence speed and final performance. In [29], authors concentrate on hybrid beamforming in integrated satellite–terrestrial networks. Their research lies in achieving both secrecy and energy efficiency through carefully designed beamforming strategies. In this work [30], authors focus on the security aspects of resource allocation within cognitive satellite–terrestrial networks. They delve into the design of secure transmission techniques through strategic power and resource allocation. In [31], authors present a comprehensive examination of the issue of malicious reflector-assisted beamforming in IoT networks. Their primary focus is on devising power allocation strategies to counteract the impact of destructive beamforming. In [32], authors explore the advantages of incorporating reconfigurable intelligent surfaces (RISs) and their role in power allocation and system performance. PD-DQN is well suited for solving problems involving hybrid action spaces, making it a more efficient choice for the joint power allocation and user association problem, which uses discrete and continuous action space [28]. This hybrid approach is designed to address the challenges posed by a mixed discrete–continuous action space.
In this study, we introduced a novel approach with a PD-DQN algorithm [28]. The main contributions of this paper are as follows:
  • Our work presents an advanced approach to the user association and power allocation problem, with the overarching objective of maximizing EE in the context of a massive MIMO network. By aiming for EE as the primary goal, we address a critical aspect of modern wireless communication networks, highlighting the practical relevance of our research.
  • To effectively address power allocation challenges, we have carefully considered the design of the action space, the state space, and the reward function. Our approach leverages the model-free Deep Q-Network (DQN) framework and, notably, the PD-DQN framework, which empowers us to update policies within the hybrid discrete–continuous action space. This innovation is pivotal in enabling adaptive decision-making within complex real-world scenarios, underscoring the practicality of our methodology.
  • The simulation results show that the proposed user association and power allocation method based on PD-DQN performs better than DRL and RL methods.

2. System Model

In this study, we considered a single-cell massive MIMO network, which consists of N remote radio heads (RRHs) and single antenna users. Here, RRHs are connected to a baseband unit via backhaul connections. Each RRH is equipped with M m a x antennas. There are U single-antenna users served by N RRHs together operating in the same time frequency domain. It is assumed that M m a x > U. In this network, we associate each user with a single RRH [7]. The set of users is denoted by U. Figure 1 shows the network architecture based on a DRL. These approaches utilize a credit assignment system to reward parameters based on their recent impact on the search process. They also employ an operator selection mechanism to determine which parameters to use, guided by the assigned credits.
It is assumed that the channel between the uth users and the nth RRH is given by
h n , u = β n , u   g n , u
where β n , u signifies the large-scale fading coefficient, and g n , u   signifies the small-scale fading coefficient. g n , u is also known as Rayleigh fading, and the elements are independent and identically distributed (i.i.d) random variables having zero mean and unit variance [8].
The received signal of the uth user on the nth RRH can be given by [7]
Y n , u = p u h n , u w n , u s n , u + j = 1 , j u U p j h n , u w n , u s n , j + z n , u
where p u is power transmitted through the uth user; s n , u is the N r × 1 data symbol of the uth user on the nth RRH; w n , u is the N t × N r beamforming vector of the uth user on the nth RRH, which is given by w n , u = h n , u E h ^ n , u 2 ; and h n , u is the channel matrix of the uth user on the nth RRH. z n , u is the noise vector of independent and identically distributed (i.i.d.) additive complex Gaussian noise having zero mean and variance of σ . Without loss of generality, we set
E s m , u s m , u H   = I ,   E s m , u s m , j H   = 0 ,   ( k j ) ,   E s m , u n m , u H = 0

Power Consumption

We considered the downlink phase for power consumption. The overall P c is expressed as the sum of the transmit power and fixed power consumption of the RRHs and base units denoted as P F I X and the power consumed by the components of the active antennas [5]. The total P c can then be given by
P c = P F i x + P a n = 1 N M n + n = 1 N u = 1 U 1 ν P n , u
where P a denotes power assumed for active antenna, and ν denotes power amplifier efficiency, ν 0,1 .

3. Problem Formulation

According to the system model, the ergodic achievable rate of the uth user is given by [10]
R n , u = log 2 1 + Γ u , n
where Γ u , n is signal-to-interference-plus-noise ratio of the uth and nth RRH and is given by
R n , u = M n p n Γ n , u j = 1 N q = 1 U p j , q β j , u + σ 2
To deal with the above problem, the association between the uth users and nth RRH is given by
x n , u = 1 , u L l 0 , u U L l l = 1 , . . . L
The system energy efficiency (EE) can be expressed as
m a x X , P   η = u = 1 U R u P F I X + p a n = 1 N M n + n = 1 N u = 1 U 1 v p n , u
The optimization problem maximizing the system EE can be formulated as
P 1 :   m a x   X , P η = u = 1 U R u ( X , P ) P c X , P ,   s . t   C 1 :   0 < P u , m P u m a x C 2 :   a u , m 0,1 C 3 :   m a u , m = 1 C 4 :   R u R min
In the above problem, C1 indicates that the transmitted power consumption is smaller than the transmit power limit of each RRH. Constraints C2 and C3 indicate that one user can only be associated with one RRH. C4 maintains the QoS requirement of each user and signifies the lower limit of the required transmit rate of users. Problem (8) is NP-hard and is usually difficult to find a feasible solution for [22]. Therefore, a multi-agent DRL approach was used to solve this problem, as described in the next section.

4. Multi-Agent DRL Optimization Scheme

The problem P1 is a non-convex problem where user association as well as power allocation approaches are involved. To solve this tractable problem, a multi-agent DQN-based RL technique was applied. The major component of the RL approach is based on the Markov decision-making process (MDP), which is a new proposed reward function, prior to the application of the multi-agent DQN approach.

4.1. Overview of RL Method

In this section, we present the overview of RL. In RL, the aim is to find optimal policy. The problem P1 is converted into a MDP (s, a, r, P s s n e w ) similar to the existing work [26,27], where s, a, and r represent the set of state, set of action, and reward functions, respectively. P s s n e w is the transition probability from state s to s n e w with reward r. In the DRL, these state variables are defined as follows:
State space: In problem P1, the users as agents select the BSs for communication at time t. The network consists of U agents. The state space can be expressed as
s ( t ) = s 1 ( t ) , s 2 ( t ) , . . . s u ( t ) , . . . , s U ( t )
Action space: At time t, the action of the agent is to control the transmit power level between the user association and BS. The action space consisting of each user can be defined as
a ( t ) = a 1 ( t ) , a 2 ( t ) , . . a s u ( t ) , . . . , a U ( t )
Reward function: The energy efficiency of all users can be expressed as a system reward function
r ( t ) = u U r u ( t ) = u = 1 U η u ( t )
where is r ( t ) the reward function, which is maximized to achieve the optimal policy with interaction with the outer environment.
Therefore, within the RL framework, the problem P1 can be transformed into problem P2, as follows:
P 1 :   max X , P , M r n
where X represents the user association matrix, and P denotes the power allocation vector. The agent identifies its state s(t) at time t and follows a policy π to perform an action a(t), that is, a ( t ) = π ( s ( t ) ) . Following this, the users communicate with the BSs, and the reward function becomes r t = r t | s = s ( t ) , a = a ( t ) . Therefore, the future cumulative discounted reward at time t can be given by
R ( t ) = τ = t T γ τ t r ( τ )
where γ ∈ [0, 1] denotes the discount factor for the upcoming rewards. To solve the P2, a value function for policy π u is defined as
V u π u = Ε τ = t T γ τ t r u ( τ ) | s u τ , a u ( τ )
where E[·] denotes the expectation operator. Using the Markov property, the value function is defined as
V u π u = r ( s u τ , π u ) + γ s n e w S P s s n e w ( π u ) V u π u ( s u τ n e w )
The Q-function when performing action a u ( τ ) in state s u τ with policy π u can be expressed as in [24], that is,
Q π s τ , a ( τ ) = Ε R τ | s τ , a ( τ )
The optimal Q-value function satisfies the Bellman equation [33,34] derived as
Q π * s τ , a ( τ ) = r s τ , a ( τ ) + γ s S P s s a ( τ ) V π * ( s τ )
Accordingly, the Bellman optimality Equation (17) [24], V u π u * ( s u τ n e w ) can be obtained as
V u π u * ( s u τ n e w ) = m a x a u Q π u * s u τ n e w , a u τ n e w ( τ )
Adding Equations (17) and (18), we get
Q π * s τ , a ( τ ) = r s τ , a ( τ ) + γ s S P s s a ( τ ) m a x a u Q π u * s u τ n e w , a u τ n e w ( τ )
The update of the Q-value function is given by [14] as
Q π * s τ , a ( τ ) = ( 1 α ) Q π s τ , a ( τ ) + α r s τ , a ( τ ) + γ m a x a Q π s τ , a ( τ )
where α is learning rate scaled between 0 and 1 and updating speed of Q π u s u τ , a u ( τ ) . The RL algorithm shows good performance if size of states is small. In case of high dimensional state space, classical RL approaches fail to perform. Some states are not sampled because of the high dimensional state space, and they require several restrictions. First, the convergence rate might become slow, and storage of the lookup table becomes impractical. Thus, the use of the DRL method was explored to solve the problem with the large space.

4.2. Multi-Agent DQN Frameworks

In contrast to the classical Q-learning approach, the author in [23] proposed the DQN method, which was basically a DRL method. This DQN method relies on two components, e.g., replay memory and target network. The agent stores transitions s u τ , a u ( τ ) , r u ( τ ) , s u τ n e w   in a replay memory D. Then, it extracts this transition from memory D by using random sampling to compute the Q-value function. The agent uses the memory D in a part of mini-batch to train the Q-network, and then a gradient descent method is applied to update the weight parameter θ of the behavior network:
π u = max Q π u * s u τ , a u n e w ( τ ) a
In the DQN method, the types of networks are included; that is, DQN sets the θ target target networks. The learning model calculates the target value y j   with a weight parameter θ target for a certain time t, which can mitigate the volatility of the learning scheme. During the learning process, after several iterations of H the weight parameter θ is synchronized with the target network, θ θ target . The agent utilizes a greedy random policy, which means that the agent randomly selects an action a u ( τ ) parameter θ for the behavior network. Consequently, a ( τ ) value and θ are updated iteratively using the minimum loss function [35]:
L ( θ ) = y j Q π u ( s u , a u ; θ ) ,
where y j = r j + γ m a x a u n e w ( j ) Q π u ( s u j n e w , a u n e w ( j ) ; θ target ) .
The proposed DQN algorithm for user association and power allocation is shown in Algorithm 1.
Algorithm 1: DQN Based for User Association and Power Allocation Algorithm
1. Initialize Q(s, a) = 0; learning rate α, target network and replay memory D.
2. Set the weight, discount factor
3.  for each training episode do
4.      Initialize state s
5.      Choose a random number
6.       if   x < ε then
7.         Choose action randomly;
8.       else
9.        Select action  a u ( τ ) = m a x a n e w s u τ , a u n e w ( τ ) ; θ
10.      end if
11.      Execute action  a u ( τ ) and next state s u τ n e w
12.      Calculate energy efficiency using Equation (13).
13.      Store transition s u τ , a u ( τ ) , r u ( τ ) , s u τ n e w in D
14.        If the replay memory is full then
15.        Random sampling a mini-batch from D
16.        Perform gradient descent on y j Q π u ( s u , a u ; θ ) 2   w . r . t .   parameter   θ
17        end if
18 Update target network
19 End for

4.3. Parameterized Deep Q-Network Algorithm

The combined user association and power allocation procedure in a hybrid action space can be solved via parameterization, but it still has generalization problems. In order to solve this problem, we used the epsilon greedy exploration method, which enables the DQN to explore a wide variety of states and actions and improves generalization. In a hybrid action space, the Q-value is recast as Q ( s , a ) = Q ( a , β , χ ) where β denotes a discrete action, and χ denotes a continuous action. The reward is defined as parameterized double DQN with replay buffer EE. Whether UE is associated with BS or not, the user association will only have the two discrete values β = 0 and 1. On the other hand, χ is a matrix that represents various levels of power distribution. The PD-DQN for the user association and power allocation algorithm is presented in Algorithm 2.
Algorithm 2: PD-DQN for User Association and Power Allocation Algorithm
1. Initialize primary and target Deep Q-Networks (DQN) with random weights.
2. Set up a Mini-batch and a Replay Buffer for experience storage.
3. For each episode:
Generate an initial state (s_i) by selecting a random action.
4. Inside each episode loop:
      While the episode is ongoing:
         Choose an action based on an epsilon-greedy policy.
      If a random number is less than epsilon:
      Select a discrete action from a predefined set ( β ).
    Otherwise:
         Estimate Q-values for discrete actions using the primary Q-network and choose the highest.
      If a random number is less than epsilon:
         Choose a continuous action from a predefined set ( χ ).
    Otherwise:
      Estimate Q-values for continuous actions using the primary Q-network and choose the highest.
      Execute the selected action in the environment following an epsilon-greedy policy.
5. After each episode loop, sample a minibatch of experiences from the replay buffer.
6. For each experience in the minibatch:
      Calculate the target Q-value using the target network.
        If the episode is ongoing, compute the target Q-value for the next state.
        If the episode is complete, compute the target Q-value with the reward.
        Calculate the predicted Q-value for the current state.
        Determine the difference between predicted and target Q-values and update the primary Q-network accordingly.
        Update the current state, target network weights.
7. Update the environment with user associations and power allocations.
8. Repeat the process for the next episode if needed.
9. End

5. Simulation Results

In this section, the results of the simulation using the DRL algorithms are presented. We consider a distributed M-MIMO with three RRHs equipped with 300 antennas in the cell with diameter of 2000 m. We consider K = 20 randomly distributed users within the cell. The PC of each RRH is set to 10,000 mW.
The power consumed by the component of active antennas is 200 mW, and the power amplifier efficiency, ν, is 0.25. We assume a transmission bandwidth of 10 MHz [7]. The other parameters are given in Table 1. We consider the Hata-COST231 propagation model [8]. The large-scale-fading β in Equation (1) is borrowed from the literature [8] and is given by 10 log 10 β n , u = P L n , u + Ω where PL is path loss, and Ω represents standard deviation. Here, dn,u is the distance between uth users and the nth RRH.
Figure 2 shows the energy efficiency of the proposed DRL algorithm and RL algorithm. Figure 2 demonstrates two observations that indicate that the EE achieved by the DRL method outperforms the RL methods. As the number of episodes increases up to 50, the system EE increases and tends to converge after 250 episodes for both schemes. Additionally, the learning speed of the Q-learning method is lower than that of the multi-agent DQN algorithm. For the Q-learning method, there is a slight improvement in the system EE at episode 120, whereas in the DRL approach, the system EE tends to be stable at episode 257. The EE is unstable at the beginning as seen in the DRL scheme, and the stability increases as the episodes increase, and thereafter increases slowly. This is because the agent selects actions in a random manner and stores the transition information in D.
Figure 3 shows the EE versus the number of users fixed at M = 20. From the figure we can see that, the EE generally first increases with K from 5 to 45 and then decreases flatten. This is due to the fact that when scheduling more users, more RRHs are activated to serve users creating more interference noise. Furthermore, proposed DRL algorithm performs superior to QL.
Figure 4 compares the EE performance at different discounted factors, γ = {0.1, 0.5, and 0.8}. It can be observed that a lower discount factor results in higher EE through different discount factors. Figure 4 illustrates that the PD-DQN method optimizes the user association and power allocation. Moreover, when the number of epochs increases, the EE performance of each user is better at a different discount factor.
From the Figure 5, we can observe that PDQN consistently outperforms both DQN and RL in terms of energy efficiency. The EE values achieved by PDQN show a steady increase, starting from 1.5 and reaching 5.25, indicating significant improvement. DQN and RL also exhibit improvements, but their EE values remain below that of PDQN. In terms of percentage improvement, PDQN surpasses DQN by an average of around 35%, while PDQN outperforms RL by approximately 40%. Additionally, DQN exhibits a slight advantage over RL, with an average improvement of about 5%.

6. Conclusions

In this paper, we have studied the user association and power allocation problem in a massive MIMO based on a PD-DQN framework. The numerical results indicate that the PD-DQN approach performs better than the DQN and classical Q-learning schemes. The main motivation of this paper is to study the resource allocation scheme in M-MIMO. In addition, for the simulation results in this study, we considered the DQN approach to tackle the problem of user association and power allocation in M-MIMO. The aforementioned optimization problem was formulated to maximize the EE in the downlink network, and the convergence of the multi-agent DRL (DQN) algorithm was studied. Furthermore, convergence analyses confirmed that the proposed methods perform better in terms of EE than the RL method. Establishing the upper bound is important to set a benchmark for the achievable EE. In our study, this upper bound signifies the theoretical maximum EE that can be achieved in the given massive MIMO resource allocation problem under ideal conditions. Our study has showcased the potential of our PD-DQN framework to enhance energy efficiency in massive MIMO systems. The inclusion of the upper bound in our evaluation has not only validated the effectiveness of our approach but has also shed light on areas for further improvement. The convergence rate indicates that the proposed algorithm excels in terms of energy efficiency. Furthermore, additional simulation outcomes demonstrate superior energy efficiency across varying user counts, number of antennas, and diverse learning rates. The enhancement of the proposed PD-DQN on average may reach 72.2% and 108.5% over the traditional DQN and RL methods, respectively.

Author Contributions

Methodology, S.S.; Formal analysis, W.Y.; Investigation, S.S. and W.Y.; Writing—original draft, S.S.; Writing—review & editing, W.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Research Foundation of Korea (NRF), Ministry of Education, Science and Technology (Grant No. 2016R1A2B4012752).

Data Availability Statement

Data is confidential and can be made available on request from corresponding authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, H.; Wang, Z.; Wang, H. An energy-efficient power allocation scheme for Massive MIMO systems with imperfect CSI. Digit. Signal Process. 2021, 112, 102964. [Google Scholar] [CrossRef]
  2. Rajoria, S.; Trivedi, A.; Godfrey, W.W.; Pawar, P. Resource Allocation and User Association in Massive MIMO Enabled Wireless Backhaul Network. In Proceedings of the IEEE 89th Vehicular Technology Conference (VTC2019-Spring), Kuala Lumpur, Malaysia, 28 April–1 May 2019; pp. 1–6. [Google Scholar] [CrossRef]
  3. Ge, X.; Li, X.; Jin, H.; Cheng, J.; Leung, V.C.M. Joint user association and user scheduling for load balancing in heterogeneous networks. IEEE Trans. Wireless Commun. 2018, 17, 3211–3225. [Google Scholar] [CrossRef]
  4. Liang, L.; Kim, J.; Jha, S.C.; Sivanesan, K.; Li, G.Y. Spectrum and power allocation for vehicular communications with delayed CSI feedback. IEEE Wirel. Commun. Lett. 2017, 6, 458–461. [Google Scholar] [CrossRef]
  5. Bu, G.; Jiang, J. Reinforcement Learning-Based User Scheduling and Resource Allocation for Massive MU-MIMO System. In Proceedings of the 2019 IEEE/CIC International Conference on Communications in China (ICCC), Changchun, China, 11–13 August 2019; pp. 641–646. [Google Scholar] [CrossRef]
  6. Yang, K.; Wang, L.; Wang, S.; Zhang, X. Optimization of resource allocation and user association for energy efficiency in future wireless networks. IEEE Access 2017, 5, 16469–16477. [Google Scholar] [CrossRef]
  7. Dong, G.; Zhang, H.; Jin, S.; Yuan, D. Energy-Efficiency-Oriented Joint User Association and Power Allocation in Distributed Massive MIMO Systems. IEEE Trans. Veh. Technol. 2019, 68, 5794–5808. [Google Scholar] [CrossRef]
  8. Ngo, H.Q.; Ashikhmin, A.; Yang, H.; Larsson, E.G.; Marzetta, T.L. Cell-free massive MIMO versus small cells. IEEE Trans. Wirel. Commun. 2017, 16, 1834–1850. [Google Scholar] [CrossRef]
  9. Elsherif, A.R.; Chen, W.-P.; Ito, A.; Ding, Z. Resource Allocation And Inter-Cell Interference Management For Dual-Access Small Cells. IEEE J. Sel. Areas Commun. 2015, 33, 1082–1096. [Google Scholar] [CrossRef]
  10. Sheng, J.; Tang, Z.; Wu, C.; Ai, B.; Wang, Y. Game Theory-Based Multi-Objective Optimization Interference Alignment Algorithm for HSR 5G Heterogeneous Ultra-Dense Network. IEEE Trans. Veh. Technol. 2020, 69, 13371–13382. [Google Scholar] [CrossRef]
  11. Zhang, X.; Sun, S. Dynamic scheduling for wireless multicast in massive MIMO HetNet. Phys. Commun. 2018, 27, 1–6. [Google Scholar] [CrossRef]
  12. Nassar, A.; Yilmaz, Y. Reinforcement Learning for Adaptive Resource Allocation in Fog RAN for IoT with Heterogeneous Latency Requirements. IEEE Access 2019, 7, 128014–128025. [Google Scholar] [CrossRef]
  13. Sun, Y.; Feng, G.; Qin, S.; Liang, Y.-C.; Yum, T.P. The Smart Handoff Policy for Millimeter Wave Heterogeneous Cellular Networks. IEEE Trans. Mob. Comput. 2018, 17, 1456–1468. [Google Scholar] [CrossRef]
  14. Watkins, C.J.C.H.; Dayan, P. Q-Learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  15. Zhai, Q.; Bolić, M.; Li, Y.; Cheng, W.; Liu, C. A Q-Learning-Based Resource Allocation for Downlink Non-Orthogonal Multiple Access Systems Considering QoS. IEEE Access 2021, 9, 72702–72711. [Google Scholar] [CrossRef]
  16. Amiri, R.; Mehrpouyan, H.; Fridman, L.; Mallik, R.K.; Nallanathan, A.; Matolak, D. A machine learning approach for power allocation in HetNets considering QoS. In Proceedings of the 2018 IEEE International Conference on Communications (ICC), Kansas City, MO, USA, 20–24 May 2018; pp. 1–7. [Google Scholar] [CrossRef]
  17. Ghadimi, E.; Calabrese, F.D.; Peters, G.; Soldati, P. A reinforcement learning approach to power control and rate adaptation in cellular networks. In Proceedings of the 2017 IEEE International Conference on Communications (ICC), Paris, France, 21–25 May 2017; pp. 1–7. [Google Scholar] [CrossRef]
  18. Meng, F.; Chen, P.; Wu, L. Power allocation in multi-user cellular networks with deep Q learning approach. In Proceedings of the ICC 2019—2019 IEEE International Conference on Communications (ICC), Shanghai, China, 20–24 May 2019; pp. 1–7. [Google Scholar] [CrossRef]
  19. Ye, H.; Li, G.Y.; Juang, B.F. Deep reinforcement learning based resource allocation for v2v communications. IEEE Trans. Veh. Technol. 2019, 68, 3163–3173. [Google Scholar] [CrossRef]
  20. Wei, Y.; Yu, F.R.; Song, M.; Han, Z. Joint optimization of caching, computing, and radio resources for fog-enabled IOT using natural actor critic deep reinforcement learning. IEEE Internet Things J. 2019, 6, 2061–2073. [Google Scholar] [CrossRef]
  21. Sun, Y.; Peng, M.; Mao, S. Deep reinforcement learning-based mode selection and resource management for green fog radio access networks. IEEE Internet Things J. 2019, 6, 960–1971. [Google Scholar] [CrossRef]
  22. Rahimi, A.; Ziaeddini, A.; Gonglee, S. A novel approach to efficient resource allocation in load-balanced cellular networks using hierarchical DRL. J. Ambient. Intell. Humaniz. Comput. 2021, 13, 2887–2901. [Google Scholar] [CrossRef]
  23. Zhao, N.; Liang, Y.-C.; Niyato, D.; Pei, Y.; Wu, M.; Jiang, Y. Deep reinforcement learning for user association and resource allocation in heterogeneous cellular networks. IEEE Trans. Wirel. Commun. 2019, 18, 5141–5152. [Google Scholar] [CrossRef]
  24. Nasir, Y.S.; Guo, D. Multi-Agent Deep Reinforcement Learning for Dynamic Power Allocation in Wireless Networks. IEEE J. Sel. Areas Commun. 2019, 37, 2239–2250. [Google Scholar] [CrossRef]
  25. Xu, Y.; Yu, J.; Headley, W.C.; Buehrer, R.M. Deep Reinforcement Learning for Dynamic Spectrum Access in Wireless Networks. In Proceedings of the 2018 IEEE Military Communications Conference (MILCOM), Los Angeles, CA, USA, 29–31 October 2018; pp. 207–212. [Google Scholar] [CrossRef]
  26. Li, M.; Zhao, X.; Liang, H.; Hu, F. Deep reinforcement learning optimal transmission policy for communication systems with energy harvesting and adaptive mqam. IEEE Trans. Veh. Technol. 2019, 68, 5782–5793. [Google Scholar] [CrossRef]
  27. Su, Y.; Lu, X.; Zhao, Y.; Huang, L.; Du, X. Cooperative communications with relay selection based on deep reinforcement learning in wireless sensor networks. IEEE Sens. J. 2019, 19, 9561–9569. [Google Scholar] [CrossRef]
  28. Xiong, J.; Wang, Q.; Yang, Z.; Sun, P.; Han, L.; Zheng, Y.; Fu, H.; Zhang, T.; Liu, J.; Liu, H. Parametrized deep q-networks learning: Reinforcement learning with discrete-continuous hybrid action space. arXiv 2018, arXiv:1810.06394. [Google Scholar]
  29. Lin, Z.; Lin, M.; Champagne, B.; Zhu, W.-P.; Al-Dhahir, N. Secrecy-Energy Efficient Hybrid Beamforming for Satellite-Terrestrial Integrated Networks. IEEE Trans. Commun. 2021, 69, 6345–6360. [Google Scholar] [CrossRef]
  30. Lin, Z.; Niu, H.; An, K.; Hu, Y.; Li, D.; Wang, J.; Al-Dhahir, N. Pain without Gain: Destructive Beamforming from a Malicious RIS Perspective in IoT Networks. IEEE Internet Things J. 2023. [Google Scholar] [CrossRef]
  31. An, K.; Lin, M.; Ouyang, J.; Zhu, W.-P. Secure Transmission in Cognitive Satellite Terrestrial Networks. IEEE J. Sel. Areas Commun. 2016, 34, 3025–3037. [Google Scholar] [CrossRef]
  32. Lin, Z.; Niu, H.; An, K.; Wang, Y.; Zheng, G.; Chatzinotas, S.; Hu, Y. Refracting RIS-Aided Hybrid Satellite-Terrestrial Relay Networks: Joint Beamforming Design and Optimization. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 3717–3724. [Google Scholar] [CrossRef]
  33. Hsieh, C.-K.; Chan, K.-L.; Chien, F.-T. Energy-Efficient Power Allocation and User Association in Heterogeneous Networks with Deep Reinforcement Learning. Appl. Sci. 2021, 11, 4135. [Google Scholar] [CrossRef]
  34. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, UK, 1998. [Google Scholar]
  35. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
Figure 1. System model based on parameterized DRL.
Figure 1. System model based on parameterized DRL.
Electronics 12 04517 g001
Figure 2. Convergence of energy efficiency values.
Figure 2. Convergence of energy efficiency values.
Electronics 12 04517 g002
Figure 3. EE versus number of users.
Figure 3. EE versus number of users.
Electronics 12 04517 g003
Figure 4. EE versus number of Epochs.
Figure 4. EE versus number of Epochs.
Electronics 12 04517 g004
Figure 5. EE versus number of antennas.
Figure 5. EE versus number of antennas.
Electronics 12 04517 g005
Table 1. Simulation parameters.
Table 1. Simulation parameters.
ParameterValues
Standard deviation8 dB
Path loss model PLPL = −140.6−35log10(d)
Episodes500
Steps T500
Discount rate γ0.9
Mini-batch size b8
Learning rate0.01
Replay memory size D5000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sharma, S.; Yoon, W. Energy Efficient Power Allocation in Massive MIMO Based on Parameterized Deep DQN. Electronics 2023, 12, 4517. https://doi.org/10.3390/electronics12214517

AMA Style

Sharma S, Yoon W. Energy Efficient Power Allocation in Massive MIMO Based on Parameterized Deep DQN. Electronics. 2023; 12(21):4517. https://doi.org/10.3390/electronics12214517

Chicago/Turabian Style

Sharma, Shruti, and Wonsik Yoon. 2023. "Energy Efficient Power Allocation in Massive MIMO Based on Parameterized Deep DQN" Electronics 12, no. 21: 4517. https://doi.org/10.3390/electronics12214517

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop