Next Article in Journal
A Malicious Code Detection Method Based on Stacked Depthwise Separable Convolutions and Attention Mechanism
Previous Article in Journal
Architecture for a Mobile Robotic Camera Positioning System for Photogrammetric Data Acquisition in Hydroelectric Tunnels
Previous Article in Special Issue
Research on Satellite Network Traffic Prediction Based on Improved GRU Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Slime Mould Optimization with Deep-Learning-Based Resource Allocation in UAV-Enabled Wireless Networks

1
Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
2
School of Automation, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
3
RUDN University, 6 Miklukho-Maklaya Street, 117198 Moscow, Russia
4
Institute of Computer Technologies and Information Security, Southern Federal University, 347922 Taganrog, Russia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(16), 7083; https://doi.org/10.3390/s23167083
Submission received: 25 April 2023 / Revised: 6 July 2023 / Accepted: 22 July 2023 / Published: 10 August 2023
(This article belongs to the Special Issue Resource Allocation for Cooperative Communications)

Abstract

:
Unmanned aerial vehicle (UAV) networks offer a wide range of applications in an overload situation, broadcasting and advertising, public safety, disaster management, etc. Providing robust communication services to mobile users (MUs) is a challenging task because of the dynamic characteristics of MUs. Resource allocation, including subchannels, transmit power, and serving users, is a critical transmission problem; further, it is also crucial to improve the coverage and energy efficacy of UAV-assisted transmission networks. This paper presents an Enhanced Slime Mould Optimization with Deep-Learning-based Resource Allocation Approach (ESMOML-RAA) in UAV-enabled wireless networks. The presented ESMOML-RAA technique aims to efficiently accomplish computationally and energy-effective decisions. In addition, the ESMOML-RAA technique considers a UAV as a learning agent with the formation of a resource assignment decision as an action and designs a reward function with the intention of the minimization of the weighted resource consumption. For resource allocation, the presented ESMOML-RAA technique employs a highly parallelized long short-term memory (HP-LSTM) model with an ESMO algorithm as a hyperparameter optimizer. Using the ESMO algorithm helps properly tune the hyperparameters related to the HP-LSTM model. The performance validation of the ESMOML-RAA technique is tested using a series of simulations. This comparison study reports the enhanced performance of the ESMOML-RAA technique over other ML models.

1. Introduction

The rising demand for higher-quality wireless services drives upcoming wireless transmission systems to be responsible for widespread coverage and connectivity over any mobile device [1,2]. Also, the variety of network applications greatly demands energy consumption, network capability, and service latency for masses of mobile devices. For realizing the vision of limitless access to wireless information anytime and anywhere for everything, the recently developed unmanned aerial vehicle (UAV)-based flying platform is capable of breaking the limitation of the conventional network structure that drives us to rethink the advancement of transmission systems in the upcoming generation [3]. UAVs—in other words, drones—have gained special consideration due to their low-cost deployment, simplicity, and prominent flexibility. Due to their higher flying attitude, drone-based platforms have established effectual Line-of-Sight (LoS) connection with ground users (GUs), thereby reducing the power utilization for reliable connectivity [4]. Figure 1 demonstrates an overview of UAV-enabled WSNs.
Thus, the UAV-based flying mobile transmission technique gives energy- and cost-effective solutions with a limited territorial cellular structure for the GU. Formulating drone-assisted wireless communication systems has attracted more and more research interest [5]. The present study on the UAV-related wireless communication mechanism primarily focuses on resource optimization and drone placement, assuming that drones serve as aerial relays or aerial BSs to assist GUs [6]. The altitude of a drone can be enhanced for the trajectory model without or with the horizontal location related to various QoS requirements and considerations.
Although UAV-related communications have numerous merits in real time, certain technical issues should be solved to unlock the promising performance gains [5]. Initially, stringent power limitations were a bottleneck to effective drone communications. The energy storage of the onboard batteries of a drone was generally small because of the size and weight limitations of the drone [7]. Moreover, transmission and flight power consumption are based on the drone’s velocity and trajectory. Thus, energy-efficient drones have drawn critical research attention in the literature. Binding the above-mentioned advantages of drone-based networks faces several technical challenges in resource allocation models [8,9]. To be specific, UAV-based networks perform optimally if drones’ trajectories or positions are adequately planned, drones’ power transfer is adequately assigned, and the UAV-UE relationship is properly managed to handle the dynamic of channel state information (CSI) among UEs and UAVs [10]. Such joint designs frequently need a precise CSI prediction. But perfect CSI is only sometimes probable since drones can flexibly move in space, making CSI rapidly change location over time. Moreover, UAV-based networks still suffer similar difficulties to multicast communication networks and coordinated multipoint (CoMP) [11]. More realistic methods for efficiently advancing drone-assisted networks are of timely importance.
This paper presents an Enhanced Slime Mould Optimization with Deep-Learning-based Resource Allocation Approach (ESMOML-RAA) in a UAV-enabled wireless network. The ESMOML-RAA technique considers UAVs as learning agents with the formation of resource assignment decisions as actions and designs a reward function to minimize weighted resource consumption. For resource allocation, the presented ESMOML-RAA technique employs a highly parallelized long short-term memory (HP-LSTM) model with an ESMO algorithm as a hyperparameter optimizer. Using the ESMO algorithm helps properly tune the hyperparameters related to the HP-LSTM model. The performance validation of the ESMOML-RAA technique is tested using a series of simulations. In short, the key contributions are listed as follows:
  • Developing a new ESMOML-RAA technique for the optimal allocation of resources in UAV-assisted wireless networks, which comprises the effective allocation of restricted resources like bandwidth, power, and computing resources to multiple UAVs to optimize the system performance.
  • Employing HP-LSTM for resource allocation, where LSTM is a type of recurrent neural network that can effectively capture long-term dependencies and temporal patterns in sequential data, making it suitable for modelling the dynamic nature of UAV networks.
  • Designing an ESMO algorithm by integrating the concept of elite oppositional-based learning (EOBL), which enhances the exploration and exploitation capabilities of the SMO algorithm. Hyperparameter tuning using the ESMO algorithm helps improve the HP-LSTM model’s performance.

2. Related Works

Nguyen et al. [12] presented reconfigurable intelligent surface (RIS)-supported drone networks that either benefit the drone’s agility or employ RIS reflection to improve the network performance. A deep reinforcement learning (DRL) system was presented to resolve the continuous optimization issue with time-varying channels in a centralized way. Luong et al. [13] examined a new technique for developing the deep Q-learning (DQL) technique to consider the hassles of inaccessible CSI to determine the location of drones, and invoked the variance of convex (DC)-related optimization technique to proficiently overcome drone transmit beamforming and drone–user association to specify the determined location of drones. In [14], research scholars considered the minimized sum power issue by cooperatively enhancing power control, RA, and user association in an MEC system that includes many drones. The author developed a centralized multiagent RL (MARL) technique since the issue was nonconvex. But, essential problems, namely privacy concerns and distributed frameworks, were ignored by the centralized method. The authors modelled a multiagent federated RL (MAFRL) technique in a semidistributed structure.
Cui et al. [15] examined the dynamic RA of many drone-based transmission networks to maximize long-term benefits. The authors developed the long-run RA issue as a stochastic game to maximize the anticipated rewards and to design the uncertainty and dynamics in surroundings, in which all drones will be learning agents and all RA solutions correspond to an activity engaged in by the drones. Then, as per its local observations utilizing learning, the authors developed a MARL structure where all agents find their optimal method. In [16], the drone as a flying BS was considered for an emergency transmission system, including 5G mMTC network slicing, to enhance the service quality. The drone-related mMTC makes a BS selection method to maximize the system’s energy efficiency. Afterward, utilizing the Markov decision process (MDP) theory, the system method can be minimized into the stochastic-optimization-related issue. The authors devised an approach to optimize energy efficiency to solve the RA problem, a Dueling-Deep-Q-Network (DDQN)-related method related to the RL method.
The authors in [17] proposed a drone-assisted distributed routing framework focusing on quality-of-service provision in IoT environments (D-IoT). The study in this paper focused on highly dynamic flying ad hoc network environments. This model was utilized to develop a distributed routing framework. A neuro-fuzzy interference system was applied to achieve reliable and efficient route selection. A quality-of-service provisioning framework for a UAF-assisted aerial ad hoc network environment (QSPU) was proposed in [18] to achieve reliable aerial communications. UAV-centric mobility models were utilized to develop a complete aerial routing framework, and it was proved that a number of service-oriented performance metrics in a UAV-assisted aerial ad hoc network environment achieved better performance. Furthermore, for a UAV-assisted aerial ad hoc network environment, a quality-of-service provisioning framework was proposed to focus on reliable aerial communication [19]. UAV-centric mobility models were utilized to develop aerial routing frameworks.
Li et al. [20] devised a novel DRL-related flight RA framework (DeFRA) method for reducing the overall loss of data packets in continual action spaces. The abovementioned method depends on deep deterministic policy gradient (DDPG), optimally controlled speeds, and instantaneous headings of the drone and chooses ground devices for collecting data. Additionally, for predicting network dynamics resulting from energy arrivals and time-varying airborne channels in ground devices, a state characterization layer using LSTM was formulated. In [21], the authors deployed a clustered multidrone for providing RA and computing task offloading services to IoT gadgets. The author developed a multiagent DRL (MADRL)-related method for minimizing the overall network computational cost while assuring QoS necessities for UEs or IoT gadgets in the IoT platform. To reduce long-term computation costs with regard to energy and delay, the authors developed the issue as a natural extension of the Markov decision process (MDP) considering a stochastic game.

3. The Proposed Model

In this study, we developed a new ESMOML-RAA technique for resource assignment in UAV-enabled wireless networks. Figure 2 shows the working process of the proposed model. The presented ESMOML-RAA technique attained energy-effective and computationally effective decisions proficiently. At the same time, the ESMOML-RAA technique considered the UAV as a learning agent with the formation of resource assignment decisions as actions and designed a reward function with the intention of minimizing weighted resource consumption.

3.1. System Model

Consider a multi-UAV MEC network, wherein N mobile users (MUs) are randomly distributed with M UAVs flying in a specific place. Every UAV assumes computation and communication abilities, which enables MUs to offload the task [22,23]. The study aims to minimize the computation and energy utilization of UAVs via resource allocation (RA) using the QoS constraint of MUs concerning latency. The sets of UAVs and MUs are correspondingly indicated by M = 1 , 2 ,   ,   M and N   =   1 , 2 ,   ,   N . Then, the network functions are considered during a time K that comprises T γ time intervals represented as T = 1 , 2 , ,   T γ , and the location of the MU is considered to be constant in time intervals, signified as q i t = x i t ,   y i t . Moreover, the UAV is considered to be moving at a constant altitude H with R radius coverage, and the coordinate of the jth UAV is denoted by p j = X j t ,   Y j t . Additionally, the kth tasks of ith MU are specified as R i k = S i k ,   F i k ,   D , whereby S i k ,   F j k ,   D represents the input size, the necessary CPU circle, and the maximal tolerant time of k t h task, respectively. It should be noted that D is similar to every task that characterizes the Q o S constraint of latency-intensive tasks. Table 1 shows some of the notation used in the proposed approach.
Communication model
The channel gain from jth UAV to ith MUs was modelled as well as the representation of a multi-UAV network infrastructure, where every UAV independently makes an RA decision.
h i j t = β 0 d 2 t = β 0 H 2 + | q i t p j t | 2 ,    
In Equation (1), β 0 indicates the channel gain at the reference distance d 0 = 1 m. Once UAV j handles kth tasks from MU i , the data of task should be transferred from jth to ith, and the throughput is given as follows:
r i , k j t = B log 2 1 + a i , k j t P h i j t σ 2 + I i u t ,    
In Equation (2), a i , k j t 0 , 1 indicates the energy allocation indicator, B represents the overall bandwidth allocated to M U , and P indicates the maximal transmission power of the UAV. Now, the radio resource utilized by the smaller cells is assumed to overlap; hence, mutual interference takes place after the similar task is transferred to distinct UAV servers, using I i u t = u ϵ M ,   u   j p u t h i u t . σ 2 represents the background noise power. Thus, the communication time of kth tasks in tth interval is given by the following:
L i , k j , s t = S i k t r i , k j t ,  
In Equation (3), S i k t represents the size of the implemented task in the tth time interval. The transmission power consumption of the jth UAVs in the tth time interval is formulated by the following:
p i ; k j , s t = a i ; k j t P L i ; k j , s t .  
Generally, the computational resource of each UAV is considered to be the same, indicated as C circles for every second. Therefore, the execution time of the kth task in the jth UAVs is formulated by
L i , k j , c t = F i k t b i , k j t C ,
In Equation (5), b i , k j t 0 , 1 signifies the computational RA decision, and it can be formulated as
p i , k j , c t = γ 0 b i , k j t C .  
In Equation (6), γ 0 refers to the constant associated with the hardware structure.

3.2. Problem Formulation

By mutually considering energy consumption and computation complexity, the study focuses on minimizing the resource consumption of UAVs using the constraints of the M U s Q o S with respect to latency:
min a , b ω 0 j = 1 M t = 1 T l i , j , t p i , k j , s t + 1 ω 0 j = 1 M t = 1 T l i , j , t p i , k j , c t , i N , k K  
s . t .   C 1 :   a i , k j t 0 , 1 C 2 :   b i , k j t 0 , 1 C 3 :   t i , k j , c t + t i , k j , s t D i N , k K                                               C 4 :   | q i t p j t | 2 R 2                                 C 5 :   i = 1 M l i , j , t 1 j M , t T                                
From the expression, l i , j , t = 0 , 1 indicates either MU i offloads tasks to UAV j at time interval t .   C 1 and C 2 show the constraint on the radio RA and computation RA, respectively. C 3 shows that the overall processing time of task k must fulfil the maximal tolerant time D . Meanwhile, the radius coverage of the UAV denotes R , and C 4 guarantees the designated MU and UAV in the transmission range. C 5 represents the fact that every UAV could implement the individual task at a particular time interval.

3.3. Resource Allocation Using the HP-LSTM Model

For resource allocation, the presented ESMOML-RAA technique employed the HP-LSTM model. For enabling the LSTM to compute 0 t in parallel, the HPLSTM utilizes a bag-of-words representation s t of previous tokens for the computation of gates and HL [24]:
s t = k = 1 t 1 i k
whereas s 1 refers to the zero vector. The BoW representation s t is attained effectually using the cumulative sum function. Figure 3 showcases the structure of LSTM.
The proposed neural network consists of several layers as depicted. The first layer is a unit vector layer used to change the input from number to vector form as required by the long-short-term memory (LSTM) implementation. The next layer is a recurrent LSTM layer with a memory parameter. The LSTM layer consists of several elements according to input, output, and forget gates which use logistic sigmoid, while the memory gate uses the tanh activation function. A single LSTM layer with all the gates is illustrated in Figure 3, where Xt is the input, ht is the output on iteration, and Ct is the state of the layer on iteration. Last is a sequence last layer that returns the last element of the sequence. After that, there is a linear layer with n inputs and m outputs. The last layer is a softmax layer that normalizes the outputs.
Afterward, it concatenates the input i and the equivalent layer standardized BoW representation L N s for succeeding in computation:
v = i L N s
Here, layer normalization is introduced to prevent potential explosions due to accumulation in Equation (8) to stabilize the training process.
Then, it calculates the input gate, forget gate, and HL:
i g = σ L N W i v + b  
f g = σ L N W f v + b f  
h = α L N W h v + b    
Then, v is calculated on the order before the computation of these gates and HLs; Equations (10)–(12) are only needed to calculate the entire order, allowing the effective sequence-level parallelization of higher-cost linear transformation. But the BoWs context representation s t lacks a weighting process related to the preceding step output o t 1 of the novel LSTM; therefore, utilizing a two-layer feed-forward network for HL computation was also attempted to alleviate potentially related disadvantages:
h = W h 2 α L N W h 1 v + b h 1 + b h 2
Afterwards, it can be upgraded to the HL h with input gate i g :
h r = h i g
whereas h r implies the upgraded HL.
With h r and f g , LSTM cells are calculated across the order:
c t = c t 1 f g t + h r t  
Equation (15) keeps the step-by-step recurrence upgrade of the LSTM cell and could not be parallelized across the order, and then it only comprises element-wise multiplication–addition functions that are lightweight and, related to linear transformation, could be calculated very quickly on modern hardware.
Different from the original LSTM that calculates the output gate o g dependent upon the concatenated vector v t , it calculates the output gate with the recently created cell state c and the input to LSTM, as c is expected to have superior quality to BoW representations.
o g = σ L N W o i | c + b 0      
Lastly, it executes the output gate to cells and attains the resultant HPLSTM layer.
o = c 0 g  
Either Equation (16) (comprising the linear transformation to compute the output gate) or Equation (17) is also effectually parallelized in the order.

3.4. Parameter Tuning Using the ESMO Algorithm

In this work, the ESMO algorithm helped to properly tune the hyperparameters related to the HP-LSTM model, i.e., learning rate. Shimin Li et al. [25] proposed the SMO algorithm, which is inspired by the diffusion and behaviour conduct of slime Mold in nature. The different steps and phases of SMO comprise approaching food, oscillation, and wrapping food. The mathematical expression of the SMO algorithm is given as follows:
Approach food: in this stage, the odour of food stimulates slime mould (SM) for searching, which leads to massive oscillation and position upgrading and is mathematically given as follows:
X t + 1 = X b t + v b W . X A t X B t , r < p ; v c X t r p    
In Equation (18), v c linearly declines from one to zero. a ,   a defines the magnitude of v b . ‘t’ denotes the current iteration, ‘W’ shows the index of weight, X b represents the maximal concentration of odour, and X A and X B denote the two random SMs. ‘p’ is evaluated by means of =   tan h q i F b e s t . Furthermore, the weight ‘W’ is evaluated as follows:
W ( i ) s m e l l i n d e x = 1 + r log   F c b e s t q i F c b e s t F c w o r s t + 1 ;   C o n d i t i o n 1 r log   F c b e s t q i F c b e s t F c w o r s t + 1 ;   O t h e r
In Equation (19), the condition specifies that S i orders the first half of the population, r signifies the random value in the range of [0,1], b F represents optimum fitness obtained in the existing iterative method, and w F signifies the worst fitness value [26].
Wrap food: the wrap food process in SMO to update the position of SMO is attained as follows:
X = r a n d . U B L B + L B , r a n d < z X b t + v b . W . X t A X t B , r < p v c . X t , r p    
where L B and U B symbolize the lower and upper limitations nge; r a n d and r represent the random integer 0 ,   1 .
Grabble food: Here, the value of ν b randomly oscillates amongst a ,   a and progressively approaches 0 as the iteration increases. The values of v c oscillate amongst [−1, 1] and eventually tend to zero.
The ESMO technique was designed based on the elite opposition-based learning (EOBL) system, which is an effective and stable system that enhances population variation, broadens the searching area, avoids premature convergence, and strengthens global searching [27]. The searching system employs the possible or reversed solution for assessing the fitness value of prey, afterward ordering the optimum individual for completing the iteration. Considering that the searching agent with the optimum fitness value was regarded as an elite individual, the elite individual was calculated as x e = x e , 1 ,   x e , 2 , , x   e ,   D , the possible solution was calculated as x i = x i , 1 ,   x i , 2 , , x i ,   D   , and the reverse solution was calculated as x j = x i , 1 ,   x i , 2 , , x i ,   D . The equation was calculated as:
x i , j = k d a j + d b j x e , j ,   i = 1 , 2 , , n ; j = 1 , 2 , , D  
whereas n indicates the population size, D refers to the dimensional problem, k implies the arbitrary value that k O , 1 , and d a j and d b j signify the dynamic limits of jth decision variable, respectively; this can be calculated as:
d a j =   min   x i , j ,     d b j =   max   x i , j    
The dynamic restriction saves an optimum solution and modifies the searching area of the inverse solution.
The searching agent x i , j was calculated as:
x i , j = r a n d d a j , d b j ,     i f   x i , j d a j o r x i , j d b j

4. Performance Evaluation

In this section, the resource allocation performance of the ESMOML-RAA model is investigated. To evaluate the performance of the proposed resource allocation scheme, we used Python and TensorFlow for simulation experiments and analysis. The hardware set includes a processor of an i5-4590S CPU@ 3.00 GHz, 1 TB HDD, and 8 GB RAM. The number of base stations (BSs) was 15 with 25 users. Table 2 highlights the system performance assessment of the ESMOML-RAA model with varying numbers of BSs.
Figure 4 represents the system throughput (ST) inspection of the ESMOML-RAA technique with several BSs. The results implied that the ESMOML-RAA method obtained an improved ST with UAVs. For instance, with three BSs, the ESMOML-RAA system with UAVs obtained a higher ST of 117 Mbps. Meanwhile, with seven BSs, the ESMOML-RAA technique with UAVs attained a superior ST of 293 Mbps. Moreover, with 11 BSs, the ESMOML-RAA method with UAVs achieved a higher ST of 617 Mbps.
Figure 5 signifies the energy consumption (ECON) inspection of the ESMOML-RAA technique with numerous BSs; the outcome shows that the ESMOML-RAA system achieved enhanced ECON with UAVs. For example, with three BSs, the ESMOML-RAA method with UAVs attained a higher ECON of 150 Mbps. Meanwhile, with seven BSs, the ESMOML-RAA algorithm with UAVs achieved a greater ECON of 357 Mbps. Furthermore, with 11 BS, the ESMOML-RAA approach with UAVs acquired a maximum ECON of 537 Mbps.
In Figure 6, a comparative system energy efficiency (EE) analysis of the ESMOML-RAA method with other models such as deep Q-Network (DQN), Q-learning, random, and maximum [8] is provided. The experimental outcomes state that the ESMOML-RAA technique reached higher EE values than the other ones. For example, with 10 users, the ESMOML-RAA method attained an improved EE of 147,715,335 bit/j, while the DQN, Q-learning, random, and maximum models obtained a reduced EE of 141,671,107 bit/j, 117,116,431 bit/j, 94,450,577 bit/j, and 84,250,942 bit/j, respectively. Simultaneously, with 25 users, the ESMOML-RAA method achieved a better EE of 26,453,013 bit/j, while the DQN, Q-learning, random, and maximum models acquired a decreased EE of 23,808,663 bit/j, 18,519,963 bit/j, 17,764,435 bit/j, and 15,875,614 bit/j, respectively.
Figure 7 provides a comparative EE examination of the ESMOML-RAA model with other models. The experimental results specify that the ESMOML-RAA technique obtained greater EE values than the others. For example, with three BSs, the ESMOML-RAA method attained an improved EE of 39,213,816 bit/j, while the DQN, Q-learning, random, and maximum models reached decreased EEs of 36,459,435 bit/j, 33,882,755 bit/j, 32,016,884 bit/j, and 28,995,950 bit/j, respectively. Simultaneously, with eight BSs, the ESMOML-RAA technique achieved an enhanced EE of 56,362,059 bit/j, whereas the DQN, Q-learning, random, and maximum approaches obtained reduced EEs of 54,762,741 bit/j, 50,497,893 bit/j, 46,499,598 bit/j, and 42,856,707 bit/j, respectively.
Figure 8 shows an inspection of the overall computation time (CT) of the ESMOML-RAA technique. The obtained values imply that the ESMOML-RAA method attained reduced values of CT under all BSs. For example, with three BSs, the ESMOML-RAA approach provided a minimal CT of 184 s, while the DQN, Q-learning, random, and maximum models reached maximum CTs of 223 s, 261 s, 320 s, and 372 s, respectively. Also, with 15 BSs, the ESMOML-RAA algorithm provided the lowest CT of 393 s, while the DQN, Q-learning, random, and maximum models reached the highest CTs of 455 s, 460 s, 485 s, and 511 s, respectively.
Figure 9 shows an examination of the overall packet loss ratio (PLR) of the ESMOML-RAA model. The attained values show that the ESMOML-RAA model reached decreased values of PLR under all BSs. For example, with three BSs, the ESMOML-RAA model provided a minimal PLR of 30.02%, while the DQN, Q-learning, random, and maximum models attained maximal PLRs of 36.94%, 52.32%, 65.39%, and 75.90%, respectively. Also, with 15 BSs, the ESMOML-RAA model reached the lowest PLR of 9%, whereas the DQN, Q-learning, random, and maximum models provided the highest PLRs of 18.48%, 27.45%, 32.58%, and 38.99%, respectively.
From the detailed results, it is apparent that the ESMOML-RAA model accomplished effectual resource allocation performance.

5. Conclusions

In this study, we employed the ESMOML-RAA technique for resource assignment in UAV-enabled wireless networks. The presented ESMOML-RAA technique attained energy-effective and computationally effective decisions proficiently. At the same time, the ESMOML-RAA technique considered the UAV as a learning agent with the formation of resource assignment decisions as actions and designed a reward function with the intention of minimizing of weighted resource consumption. The presented ESMOML-RAA technique employed the HP-LSTM model with the ESMO algorithm as a hyperparameter optimizer for resource allocation. Using the ESMO algorithm helped to properly tune the hyperparameters related to the HP-LSTM module. The performance validation of the ESMOML-RAA technique was tested using a series of simulations. This comparison study reports the enhanced performance of the ESMOML-RAA technique over other ML models.
In future, new resource allocation approaches can be developed to dynamically adapt to varying network conditions in real time such as UAV mobility, varying network traffic, and environmental changes to make proactive decisions and optimize resource allocation accordingly. In addition, the integration of edge computing and federated learning can be investigated in resource allocation for UAV networks.

Author Contributions

Conceptualization, R.A. and E.M.; methodology, A.K. and M.S.A.M.; software, R.A.; validation, A.M. and E.M.; formal analysis, R.A.; investigation, A.K. and M.S.A.M.; resources, A.R. and E.M.; data curation, E.M.; writing—original draft preparation, A.M. and M.S.A.M.; writing—review and editing, R.A. and E.M.; visualization, A.M.; supervision, A.M.; project administration, R.A.; funding acquisition, A.K. and R.A. All authors have read and agreed to the published version of the manuscript.

Funding

This paper has been supported by the RUDN University Strategic Academic Leadership Program (recipient Abdukodir Khakimov), and in part by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R323), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are contained within the article and/or available from the corresponding author upon reasonable request.

Acknowledgments

The research was funded by the RUDN University Strategic Academic Leadership Program (recipient Abdukodir Khakimov), and in part the authors express their gratitude to Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R323), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Seid, A.M.; Boateng, G.O.; Anokye, S.; Kwantwi, T.; Sun, G.; Liu, G. Collaborative Computation Offloading and Resource Allocation in Multi-UAV-Assisted IoT Networks: A Deep Reinforcement Learning Approach. IEEE Internet Things J. 2021, 8, 12203–12218. [Google Scholar] [CrossRef]
  2. Rafiq, A.; Ping, W.; Min, W.; Hong, S.H.; Josbert, N.N. Optimizing energy consumption and latency based on computation offloading and cell association in MEC enabled Industrial IoT environment. In Proceedings of the 2021 6th International Conference on Intelligent Computing and Signal Processing (ICSP), Xi’an, China, 9–11 April 2021; pp. 10–14. [Google Scholar]
  3. Dai, Z.; Zhang, Y.; Zhang, W.; Luo, X.; He, Z. A Multi-Agent Collaborative Environment Learning Method for UAV Deployment and Resource Allocation. IEEE Trans. Signal Inf. Process. Netw. 2022, 8, 120–130. [Google Scholar] [CrossRef]
  4. Peng, H.; Shen, X. Multi-Agent Reinforcement Learning Based Resource Management in MEC- and UAV-Assisted Vehicular Networks. IEEE J. Sel. Areas Commun. 2020, 39, 131–141. [Google Scholar] [CrossRef]
  5. Munaye, Y.Y.; Juang, R.-T.; Lin, H.-P.; Tarekegn, G.B.; Lin, D.-B. Deep Reinforcement Learning Based Resource Management in UAV-Assisted IoT Networks. Appl. Sci. 2021, 11, 2163. [Google Scholar] [CrossRef]
  6. Hu, J.; Zhang, H.; Song, L.; Han, Z.; Poor, H.V. Reinforcement Learning for a Cellular Internet of UAVs: Protocol Design, Trajectory Control, and Resource Management. IEEE Wirel. Commun. 2020, 27, 116–123. [Google Scholar] [CrossRef]
  7. Zhao, N.; Liu, Z.; Cheng, Y. Multi-Agent Deep Reinforcement Learning for Trajectory Design and Power Allocation in Multi-UAV Networks. IEEE Access 2020, 8, 139670–139679. [Google Scholar] [CrossRef]
  8. Qi, W.; Song, Q.; Guo, L.; Jamalipour, A. Energy-Efficient Resource Allocation for UAV-Assisted Vehicular Networks with Spectrum Sharing. IEEE Trans. Veh. Technol. 2022, 71, 7691–7702. [Google Scholar] [CrossRef]
  9. Rafiq, A.; Wang, P.; Wei, M.; Muthanna, M.S.A.; Josbert, N.N. Mitigation Impact of Energy and Time Delay for Computation Offloading in an Industrial IoT Environment Using Levenshtein Distance Algorithm. Secur. Commun. Netw. 2022, 2022, 6469380. [Google Scholar] [CrossRef]
  10. Chen, X.; Liu, X.; Chen, Y.; Jiao, L.; Min, G. Deep Q-Network based resource allocation for UAV-assisted Ultra-Dense Networks. Comput. Netw. 2021, 196, 108249. [Google Scholar] [CrossRef]
  11. Khan, N.A.; Jhanjhi, N.; Brohi, S.N.; Usmani, R.S.A.; Nayyar, A. Smart traffic monitoring system using Unmanned Aerial Vehicles (UAVs). Comput. Commun. 2020, 157, 434–443. [Google Scholar] [CrossRef]
  12. Nguyen, K.K.; Khosravirad, S.R.; da Costa, D.B.; Nguyen, L.D.; Duong, T.Q. Reconfigurable Intelligent Surface-Assisted Multi-UAV Networks: Efficient Resource Allocation with Deep Reinforcement Learning. IEEE J. Sel. Top. Signal Process. 2021, 16, 358–368. [Google Scholar] [CrossRef]
  13. Luong, P.; Gagnon, F.; Labeau, F. Resource allocation in UAV-Assisted wireless networks using reinforcement learning. In Proceedings of the 2020 IEEE 92nd Vehicular Technology Conference (VTC2020-Fall), Victoria, BC, Canada, 18 November–16 December 2020; pp. 1–6. [Google Scholar]
  14. Nie, Y.; Zhao, J.; Gao, F.; Yu, F.R. Semi-Distributed Resource Management in UAV-Aided MEC Systems: A Multi-Agent Federated Reinforcement Learning Approach. IEEE Trans. Veh. Technol. 2021, 70, 13162–13173. [Google Scholar] [CrossRef]
  15. Cui, J.; Liu, Y.; Nallanathan, A. Multi-Agent Reinforcement Learning-Based Resource Allocation for UAV Networks. IEEE Trans. Wirel. Commun. 2019, 19, 729–743. [Google Scholar] [CrossRef]
  16. Gupta, R.K.; Kumar, S.; Misra, R. Resource allocation for UAV-assisted 5G mMTC slicing networks using deep reinforcement learning. Telecommun. Syst. 2022, 82, 141–159. [Google Scholar] [CrossRef]
  17. Kumar, K.; Kumar, S.; Kaiwartya, O.; Kashyap, P.K.; Lloret, J.; Song, H. Drone assisted Flying Ad-Hoc Networks: Mobility and Service oriented modeling using Neuro-fuzzy. Ad Hoc Netw. 2020, 106, 102242. [Google Scholar] [CrossRef]
  18. Kumar, K.; Kumar, S.; Kaiwartya, O.; Sikandar, A.; Kharel, R.; Mauri, J.L. Internet of Unmanned Aerial Vehicles: QoS Provisioning in Aerial Ad-Hoc Networks. Sensors 2020, 20, 3160. [Google Scholar] [CrossRef]
  19. Cao, Y.; Kaiwartya, O.; Li, T. (Eds.) Secure and Digitalized Future Mobility: Shaping the Ground and Air Vehicles Cooperation, 1st ed.; CRC Press: Boca Raton, FL, USA, 2022. [Google Scholar] [CrossRef]
  20. Li, K.; Ni, W.; Dressler, F. LSTM-Characterized Deep Reinforcement Learning for Continuous Flight Control and Resource Allocation in UAV-Assisted Sensor Network. IEEE Internet Things J. 2021, 9, 4179–4189. [Google Scholar] [CrossRef]
  21. Seid, A.M.; Boateng, G.O.; Mareri, B.; Sun, G.; Jiang, W. Multi-Agent DRL for Task Offloading and Resource Allocation in Multi-UAV Enabled IoT Edge Network. IEEE Trans. Netw. Serv. Manag. 2021, 18, 4531–4547. [Google Scholar] [CrossRef]
  22. Wang, M.; Shi, S.; Gu, S.; Zhang, N.; Gu, X. Intelligent resource allocation in UAV-enabled mobile edge computing networks. In Proceedings of the 2020 IEEE 92nd Vehicular Technology Conference (VTC2020-Fall), Victoria, BC, Canada, 18 November–16 December 2020; pp. 1–5. [Google Scholar]
  23. Rafiq, A.; Muthanna, M.S.A.; Muthanna, A.; Alkanhel, R.; Abdullah, W.A.M.; Abd El-Latif, A.A. Intelligent edge computing enabled reliable emergency data transmission and energy efficient offloading in 6TiSCH-based IIoT networks. Sustain. Energy Technol. Assess. 2022, 53, 102492. [Google Scholar] [CrossRef]
  24. Xu, H.; Liu, Q.; van Genabith, J.; Xiong, D.; Zhang, M. Multi-head highly parallelized LSTM decoder for neural machine translation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Bangkok, Thailand, 3–7 August 2021; pp. 273–282. [Google Scholar]
  25. Li, S.; Chen, H.; Wang, M.; Heidari, A.A.; Mirjalili, S. Slime mould algorithm: A new method for stochastic optimization. Future Gener. Comput. Syst. 2020, 111, 300–323. [Google Scholar] [CrossRef]
  26. Houssein, E.H.; Mahdy, M.A.; Shebl, D.; Manzoor, A.; Sarkar, R.; Mohamed, W.M. An efficient slime mould algorithm for solving multi-objective optimization problems. Expert Syst. Appl. 2022, 187, 115870. [Google Scholar] [CrossRef]
  27. Zhang, J.; Zhang, G.; Kong, M.; Zhang, T. Adaptive infinite impulse response system identification using an enhanced golden jackal optimization. J. Supercomput. 2022, 79, 10823–10848. [Google Scholar] [CrossRef]
Figure 1. Overview of UAV-enabled WSN.
Figure 1. Overview of UAV-enabled WSN.
Sensors 23 07083 g001
Figure 2. Working process of ESMOML-RAA technique.
Figure 2. Working process of ESMOML-RAA technique.
Sensors 23 07083 g002
Figure 3. Architecture of LSTM.
Figure 3. Architecture of LSTM.
Sensors 23 07083 g003
Figure 4. ST analysis of ESMOML-RAA approach with varying numbers of BSs.
Figure 4. ST analysis of ESMOML-RAA approach with varying numbers of BSs.
Sensors 23 07083 g004
Figure 5. ECON analysis of ESMOML-RAA approach with varying BS.
Figure 5. ECON analysis of ESMOML-RAA approach with varying BS.
Sensors 23 07083 g005
Figure 6. EE analysis of ESMOML-RAA approach with distinct users.
Figure 6. EE analysis of ESMOML-RAA approach with distinct users.
Sensors 23 07083 g006
Figure 7. EE analysis of ESMOML-RAA approach with varying numbers of BSs.
Figure 7. EE analysis of ESMOML-RAA approach with varying numbers of BSs.
Sensors 23 07083 g007
Figure 8. CT analysis of ESMOML-RAA approach with different numbers of BSs.
Figure 8. CT analysis of ESMOML-RAA approach with different numbers of BSs.
Sensors 23 07083 g008
Figure 9. PLR analysis of ESMOML-RAA approach with varying numbers of BSs.
Figure 9. PLR analysis of ESMOML-RAA approach with varying numbers of BSs.
Sensors 23 07083 g009
Table 1. Notations and descriptions.
Table 1. Notations and descriptions.
NotationDescription
RAResource allocation
K Network function
T γ Time interval
i Input
d 0 Reference distance
T γ Time interval
D Latency-intensive tasks
B Bandwidth
p i ; k j , s Power consumption
Table 2. Result analysis of ESMOML-RAA approach with varying numbers of BSs.
Table 2. Result analysis of ESMOML-RAA approach with varying numbers of BSs.
Number of Base StationsUnmanned Aerial VehicleWithout Unmanned Aerial Vehicle
System Throughput (Mbps)Energy Consumption (j)System Throughput (Mbps)Energy Consumption (j)
31171502750
52032723656
72933575171
93694577097
1146853784118
13531622102137
15617676120152
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alkanhel, R.; Rafiq, A.; Mokrov, E.; Khakimov, A.; Muthanna, M.S.A.; Muthanna, A. Enhanced Slime Mould Optimization with Deep-Learning-Based Resource Allocation in UAV-Enabled Wireless Networks. Sensors 2023, 23, 7083. https://doi.org/10.3390/s23167083

AMA Style

Alkanhel R, Rafiq A, Mokrov E, Khakimov A, Muthanna MSA, Muthanna A. Enhanced Slime Mould Optimization with Deep-Learning-Based Resource Allocation in UAV-Enabled Wireless Networks. Sensors. 2023; 23(16):7083. https://doi.org/10.3390/s23167083

Chicago/Turabian Style

Alkanhel, Reem, Ahsan Rafiq, Evgeny Mokrov, Abdukodir Khakimov, Mohammed Saleh Ali Muthanna, and Ammar Muthanna. 2023. "Enhanced Slime Mould Optimization with Deep-Learning-Based Resource Allocation in UAV-Enabled Wireless Networks" Sensors 23, no. 16: 7083. https://doi.org/10.3390/s23167083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop