You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

10 January 2024

MEC Server Sleep Strategy for Energy Efficient Operation of an MEC System

and
School of Information Convergence, Kwangwoon University, Seoul 01897, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Applications of Deep Learning and Artificial Intelligence Methods: 2nd Edition

Abstract

Optimizing the energy consumption of an MEC (Multi-Access Edge Computing) system is a crucial challenge for operation cost reduction and environmental conservation. In this paper, we address an MECS (MEC Server) sleep control problem that aims to reduce the energy consumption of the system while providing users with a reasonable service delay by adjusting the number of active MECSs according to the load imposed on the system. To tackle the problem, we identify two crucial issues that influence the design of an effective sleep control technique and propose methods to address each of these issues. The first issue is accurately predicting the system load. Changes in system load are spatio-temporally correlated among MECSs. By leveraging such correlation information with STGCN (Spatio-Temporal Graph Convolutional Network), we enhance the prediction accuracy of task arrival rates for each MECS. The second issue is rapidly selecting MECSs to sleep when the load distribution over an MEC system is given. The problem of choosing sleep MECS is a combinatorial optimization problem with high time complexity. To address the issue, we employ a genetic algorithm and quickly determine the optimal sleep MECS with the predicted load information for each MECS. Through simulation studies, we verify that compared to the LSTM (Long Short-Term Memory)-based method, our method increases the energy efficiency of an MEC system while providing a compatible service delay.

1. Introduction

The increasing prevalence of high-speed wireless networks, smart mobile devices, and digital services has led to a connected and data-centric society, driving innovation and economic growth across various industries. This digital transformation poses complex challenges such as low-latency data processing, efficient bandwidth utilization, and improved service quality. In response, Multi-Access Edge Computing (MEC) systems have garnered significant attention. MEC systems deploy numerous low-capacity MEC servers (MECSs) at the network edge, utilizing computing and storage resources for low-latency services. This is crucial for applications such as autonomous vehicles, augmented reality, real-time gaming, and healthcare. MEC systems efficiently manage network bandwidth and reduce congestion by processing data at the edge before transmitting it to central data centers. However, the benefit of an MEC system does not come without costs. With the expansion of the service area in MEC systems, there is an associated increase in the required number of MECSs. This escalation in the number of MECS contributes to a higher energy consumption, resulting in an augmented operational expenditure for the MEC system. In addition, networks and data centers are expected to account for 59.8% of the CO2 emissions in the information and communications technology sector by 2030 [1]. Therefore, it is essential to increase the energy efficiency of an MEC system to reduce the operational cost and tackle the environmental conservation problem.
An MECS undertakes the reception, processing, and subsequent transmission of results for tasks offloaded from users. Consequently, the energy consumed within an MEC system is categorized into the communication energy and computing energy of MECS. In the task-offloading process, a user device incurs transmission power consumption. Given the typically small size of service results, the downlink transmission power of an MECS is often assumed to be negligible. However, if an MECS cannot handle a service for processing the requested task, task migration to a cloud server or another MECS becomes necessary, constituting a significant portion of the communication energy within an MECS. To mitigate the energy impact of task migration, various service caching methods have been proposed [2,3,4]. Given the potential minimization of communication energy using appropriate service caching methods, this paper concentrates on addressing the computing energy consumption issue. Since an MEC system involves complex interactions among diverse elements, a multitude of approaches have been proposed to mitigate computing energy consumption across various facets of MEC systems. For instance, the energy consumed in an MEC system can be reduced via an optimal offloading decision [5], resource management [6], network selection [7], and dynamic voltage and frequency scaling [8]. These methodologies presuppose the active mode for all MECSs within the system, overlooking the temporal and spatial variations in tasks offloaded to each MECS. Generally, the workload imposed on an MEC system is unevenly distributed across MECSs. Thus, when the total system load is below the combined capacity of all MECSs, some MECSs may remain lightly loaded or idle. However, an MECS consumes a substantial amount of energy even in the idle state [9]. To address this, minimizing the total energy consumption of an MEC system involves placing unnecessary MECSs into a sleep state. However, since only active MECSs process loads in an MEC system, the service delay increases with the number of MECSs in the sleep state. Consequently, determining the operation mode of each MECS based on workload distribution becomes crucial for effectively reducing energy consumption while maintaining a reasonable service quality.
Various MECS sleep control methods have been proposed to strike a delicate balance between two conflicting performance metrics: the amount of energy consumed and the service latency delivered by an MEC system. Threshold methods [10,11], commonly employed in determining the operational mode of an MECS, involve transitioning an MECS into sleep mode when its load falls below a predefined threshold. This approach, wherein each MECS independently decides based on its own workload, is straightforward to implement. However, the uneven distribution of the load across MECSs at each time step prompts the exploration of comprehensive load distribution considerations to optimize system-wide energy efficiency. Addressing this, cooperative sleep decision methods have been proposed [9,12], where MECSs form clusters, exchange status information, and make sleep decisions, considering the status of other MECSs in the same cluster. Despite the potential enhancement in energy efficiency, such cooperative methods introduce additional signaling overhead, and when MECS decisions differ, an iterative consensus process ensues, potentially leading to delayed decision making. To tackle these challenges, deep learning models have been used [13,14]. These approaches leverage the current load information of MEC at a specific time point to predict the load of each MECS in the subsequent time step. Subsequently, decisions are made regarding which MECS should sleep based on these predictions. To forecast the MECS load, these techniques often model the MECS load as a time-series sequence and predominantly employ the LSTM (Long Short Term Memory) model as it is suitable for time-series data prediction. In this approach, the accuracy of the determined set of sleep MECS based on predicted values depends on the precision of load prediction for each MECS. Therefore, the optimization of the sleep MECS decision problem can be reframed as a precise workload prediction problem.
However, since LSTM was originally developed for the prediction of Euclidean data, its performance may degrade when applied to non-Euclidean data with graph structures like MEC systems. Therefore, in this paper, we improve the previous method based on the LSTM predictor in two major aspects. Firstly, we improve the accuracy of the predicted task arrival rate. The task arrival rate to an MECS varies in time and space. However, LSTM exploits only the correlation in the time domain, neglecting the valuable correlation information in the space domain. We enhance the accuracy of task arrival rate prediction by using STGCN (Spatio-Temporal Graph Convolution Network) [15]. STGCN extends the convolution operations commonly used in graph neural networks to both spatial and temporal dimensions to effectively capture information between neighboring nodes and detects temporal changes, providing enhanced capabilities in modeling spatio-temporal data. Secondly, we reduce the prediction delay. Since LSTM has a recurrent architecture, it processes inputs sequentially. Therefore, the methods based on LSTM are slow in producing predicted values. On the contrary, since we use the STGCN model that can process inputs in parallel, we can reduce the time needed to produce predicted values. We can summarize our contributions as follows.
  • We propose a framework for the MECS sleep decision by using STGCN. We define an input graph for this framework and enhance the prediction accuracy for the workload distribution in an MEC system by utilizing not only the workload correlations in each MECS in the time domain, but also the relationships among MECSs in the space domain.
  • Despite the availability of the workload distribution information, determining the operational modes of individual MECS poses a computationally challenging combinatorial optimization problem. To address this, we utilize a genetic algorithm (GA) to fast compute the optimal operation mode for each MECS at the start of each time slot, taking into account both energy consumption and service latency factors for all MECSs in the system.
  • Comprehensive simulation studies show that our approach is better than the conventional LSTM-based method in terms of both energy efficiency and the time required for determining sleep MECSs.
The organization of this paper is as follows. We present related works in Section 2. In Section 3, we describe the system model and cast the sleep control problem. We describe our MECS sleep decision method in Section 4. In Section 5, we verify the proposed method by evaluating its performance via extensive simulation studies. We conclude the paper with future research directions in Section 6.

3. System Model

In this paper, we consider an MEC system in a wireless network composed of a set of base stations (BSs), a set of MECSs, and a remote controller. We assume that each BS is located with an MECS. Henceforth, we will use MECS and BS interchangeably unless stated otherwise. We divide the system time into a series of time slots with equal length Δ t and consider a discrete time controller. Compared with a cloud server, an MECS is resource constrained in terms of the computing power and the memory size. Therefore, an MECS cannot host all the services that a cloud server has at the same time. Since the issue of service caching in an MEC system is dealt in [2,3,4], in this paper, we focus on MECS sleep control by assuming that an MECS contains the necessary services for processing tasks offloaded to it. In other words, tasks offloaded to an MEC system is served not by a cloud server but by an MECS.
We denote the set of MECSs in the system as M and the task processing capacity of an MECS i as F i . We also denote the service region of an MECS i as S i . If there is any overlap between S i and S j , MECS i and MECS j are neighboring MECSs. We denote the set of neighboring MECSs of an MECS i as M i . A user u offloads its task to the nearest MECS, which is called the serving MECS of u. We denote the set of users in S i during a time slot t as U i ( t ) . Each MECS has a task queue to accommodate the offloaded tasks. We denote the task queue length of an MECS i at the beginning of a time slot t as Q i ( t ) .
We introduce a control variable a i ( t ) that indicates the operation mode of an MECS i during a time slot t. Specifically, when a i ( t ) = 0 , an MECS i is in a sleep mode during a time slot t. On the contrary, a i ( t ) = 1 represents the case where an MECS i is in an active mode during a time slot t. We denote the number of tasks generated from the users in S i during a time slot t as λ ^ i ( t ) . If a i ( t ) = 1 , λ ^ i ( t ) s are offloaded to their active serving MECS i. However, if a i ( t ) = 0 , they are offloaded to active neighbors of the MECS i. Therefore, if we denote the number of tasks newly offloaded to an MECS i during a time slot t as λ i ( t ) , it is zero if a i ( t ) = 0 . In contrast, when a i ( t ) = 1 , λ i ( t ) becomes the sum of λ ^ i ( t ) and the tasks offloaded to i from the sleeping neighbors of the MECS i. In other words, if we denote the amount of tasks offloaded to an active MECS i from its sleeping neighbor j during a time slot t as λ ^ j i ( t ) , λ i ( t ) = λ ^ i ( t ) + j M i ( 1 a j ( t ) ) λ ^ j i ( t ) . Then, the dynamics of the task queue in an MECS i is described as
Q i ( t + 1 ) = max { 0 , Q i ( t ) + λ i ( t ) F i } .
Bounding the completion time for the tasks offloaded to an MEC system is crucial for delivering a satisfactory service to users. Since we assume that each task is served by an MECS once it is offloaded to an MEC system, the time spent to complete a task in an MEC system is determined by the task queuing delay and the task service delay in a serving MECS, which depends on the service scheduling policy, task size, and CPU capacity. We assume that the task service delay is constant because modern servers use dynamic frequency scaling. If we assume that each MECS processes tasks in a first-in, first-out (FIFO) manner, the queuing delay is proportional to the queue length. As shown in Equation (1), the change in the task queue length in an MECS i during a time slot t is determined by the queue length at the beginning of the time slot t, the amount of new tasks entering during the time slot t, and the service capacity of an MECS i. Among these factors, λ i ( t ) is unknown at the start of the time slot. Since λ i ( t ) depends on a i ( t ) , λ ^ i ( t ) , and j M i ( 1 a j ( t ) ) λ ^ j i ( t ) , we do not know the characteristics of the task arrival process to each MECS. Therefore, we take a conservative approach to bound the queue length. In other words, we aim to maintain Q i ( t ) s under Q t h .
The energy consumed by an MECS depends on its load. According to [9], the energy consumed by an MECS in an active state during a time slot t is given as
e i ( t ) = α P m + ( 1 α ) P m ρ i ( t ) ,
where P m is the maximum power consumed when the load is the highest, α [ 0 , 1 ] determines the fraction of power consumed when an MECS is idle, and ρ i ( t ) is the utilization of a MECS during a time slot t. Since the amount of tasks that an MECS can process during a time slot is given as η i ( t ) = min { Q i ( t ) + λ i ( t ) , Δ t F i } , the utilization of an MECS i during a time slot t is given as
ρ i ( t ) = η i ( t ) Δ t F i .
Therefore, the total energy consumed in an MEC system during a time slot is given as
E ( t ) = i M a i ( t ) e i ( t ) .
Our goal is to find an optimal sleep control vector Ω ( t ) = { a 1 ( t ) , , a | M | ( t ) } at the start of each time slot that minimizes the total energy consumption in an MEC system while providing a reasonable service to the users. Thus, our MECS sleep control problem is formulated as follows.
P 1 : Ω ( t ) = arg min Ω ( t ) E ( t ) s . t . Q i ( t ) < Q t h , t , i M . j M i a j ( t ) 1 , t , i M whose a i ( t ) = 0 .
The first constraint is that each MECS bounds its queue length. Wireless communication is constrained mainly by the distance between a user and a BS. Thus, to offload a task from a user to an MECS, the distance between them should be below a certain value. In this paper, we assume that MECSs are provisioned so that each user is able to communicate with an MECS located at a distance of two hops. Thus, the second constraint is that at least one of the neighboring MECS of the sleeping MECS must be active. The problem (5) is a combinatorial problem and is NP-hard in general. In addition, to solve the problem at the start of each time slot, λ i ( t ) ( i M ) should be given. However, λ i ( t ) cannot be known at the time when we need to determine Ω ( t ) . Therefore, to resolve these issues, we adopt a deep learning model STGCN to predict the task arrival rate for each MECS by considering the spatio-temporal relationship among them. Given the set of predicted task arrival rates, we employ a genetic algorithm to solve the problem (5) fast.

4. Deep Learning-Inspired Sleep Control Strategy

In Figure 1, we show the overall procedure of our MECS sleep control strategy. Our MECS sleep control method is composed of two modules. In the first module, at the beginning of each time slot t, a controller predicts { λ ^ i ( t + 1 ) : i M } by using the spatio-temporal correlations among the recent past { { λ ^ i ( t ) } , { λ ^ i ( t 1 ) } , , { λ ^ i ( t h + 1 ) } } , where h is the length of the past task arrival rate history. We denote the predicted λ ^ i ( t ) as λ ˜ i ( t ) . The second module uses the genetic algorithm to determine the MECS sleep control vector Ω ( t ) by using λ ˜ i ( t ) , Q i ( t + 1 ) , and F i , for all i M . We explain the operation of each module in the following subsections.
Figure 1. Deep learning-inspired MECS sleep control framework.

4.1. Task Arrival Rate Prediction

The amount of tasks offloaded to an MECS i is determined by the number of users in U i ( t ) and their service preferences. Since the distance that a user can move during a time slot is limited, U i ( t + 1 ) is influenced mainly by U i ( t ) and U j ( t ) s, where j M i . Therefore, λ ^ i ( t + 1 ) is affected not only by { λ ^ i ( t ) , , λ ^ i ( t h + 1 ) } but also λ ^ j ( t ) . Consequently, to increase the prediction accuracy, the spatio-temporal correlation among { λ ^ i ( t ) , , λ ^ i ( t h + 1 ) } and λ ^ j ( t ) must be exploited. To make use of the spatio-temporal relationship, we model an MEC system as a graph G = ( V , E ) . The set of nodes V is the set of MECSs M . The elements of the edge set E are the links which reflect the adjacency between MECSs. In other words, the element ( i , j ) in E is 1 if j M i , otherwise it is zero.
Our goal is to predict the amount of tasks imposed on each MECS during the next time slot. Thus, we use the history of the workload generated in each S i as a node feature. We denote the node feature vector as x ( t ) = { λ ^ i ( t ) : i M } . Therefore, our task arrival prediction problem becomes a problem to find a mapping function f p such that
[ λ ˜ 1 ( t + 1 ) , , λ ˜ | M | ( t + 1 ) ] = f p ( G ; [ x ( t ) , , x ( t h + 1 ) ] ) ,
where h is the length of the task-load history.
Since an MEC system has a graph structure, we adopt a STGCN model to approximate f p by capturing and exploiting the spatio-temporal dependency among λ ^ i ( t a ) s for all i M and a { 1 , , h 1 } .
As we can see in Figure 1, STGCN is composed of two consecutive ST-Conv blocks and one output layer. The ST-Conv block is composed of a spatial Graph-Conv module in between two temporal Gated-Conv modules (Figure 2). The Gated-Conv module is composed of a 1-D convolution unit having a width K t kernel followed by GLU (gated linear unit). We denote the input to the k-th temporal Gated-Conv module in the l-th ST-Conv block as v k , l = ( v k , l ( t ) , , v k , l ( t h + 1 ) where k , l { 1 , 2 } . Then, for each node in G, the 1-D convolution unit performs temporal convolution on v k , l by exploring K t neighbors, which reduces the sequence length from h to h K t + 1 . Two linear layers in the GLU take the temporal features extracted by the 1-D convolution unit and produces P k , l and Q k , l , respectively. Then, GLU outputs c k , l by computing P k , l σ ( Q k , l ) , where ⊙ is the element-wise Hadamard product. Therefore, if we denote the temporal convolution kernel for the k-th temporal Gated-Conv module in the l-th ST-Conv block as Γ k , l and the temporal gated convolution operator as T , the temporal gated convolution is expressed as
c k , l = Γ k , l T v k , l = P k , l σ ( Q k , l ) ,
where σ ( · ) is a sigmoid function.
Figure 2. ST-Conv block pipeline.
The spatial Graph-Conv module is designed based on a ChebNet [38], which is a spectral-based GNN. The goal of the spatial Graph-Conv module is to derive the spatial features from the temporal features extracted by the temporal Gated-Conv module. To achieve the goal, the spatial Graph-Conv module performs graph convolution on c k , l . We denote the graph convolution operator as g and the graph convolution kernel used by the spatial Graph-Conv module in the l-th ST-Conv block as Θ l ( l { 1 , 2 } ) . We also denote the input size and the output size of the feature maps as C i and C o , respectively. Since the output of the first temporal Gated-Conv module is fed into the spatial Graph-Conv module, the graph convolution on the input c 1 , l is given as
Θ l g c 1 , l = i = 1 C i Θ i , j ( L ) c 1 , l , 1 j C o ,
where Θ i , j is the Chebychev coefficients and L is the normalized graph Laplacian.
Since the output of the spatial Graph-Conv module is fed into the second temporal Gated-Conv module, the final output of the l-th ST-Conv block becomes
v l + 1 = Γ 2 , l T ReLU ( Θ l g ( Γ 1 , l T v 1 , l ) ) ,
where ReLU is the rectified linear units’ function. The output of the second ST-Conv module is fed into the fully connected output layer. The output layer performs temporal convolution on the comprehensive features obtained from two ST-Conv blocks and produces a one-step prediction { λ ˜ i ( t + 1 ) : i M } .
To train the STGCN, the following L2 loss function is used.
L ( λ ˜ | W θ ) = t | | λ ˜ ( t + 1 ) λ ^ ( t + 1 ) | | 2 ,
where W θ are all trainable parameters in the STGCN model.

4.2. MECS Sleep Control Vector Determination

After collecting { Q i ( t ) , λ ˜ i ( t ) , F i : i M } at the beginning of each time slot t, a controller determines an MECS sleep control vector by using GA. To exploit GA, we define a fitness function for a combination x Ω ( t ) as f ( x ) = E ( t ) if x satisfies the queue length constraint Q i ( t ) < Q t h , i M . When x violates the queue length constraint, we set f ( x ) = .
We create a population X c by randomly selecting n combinations from the possible combinations of Ω ( t ) . Among the combinations in X c , we find the best combination x c as follows.
x c = a r g m a x x X c f ( x ) .
To construct a crossover set Z 1 from X c , we create a temporary set Y by selecting the best n / 2 combinations from X c based on the fitness values of x X c . Then, we randomly select two combinations x and y from Y and crossover them to make two children, z 1 and z 2 . Specifically, we randomly select an index k from [ 1 , m ] , where m = | M | is the number of MECSs in an MEC system. Then, we make a child combination z 1 by concatenating x [ 1 : k ] and y [ k + 1 : m ] , where x [ 1 : k ] denotes the first k elements in x and y [ k + 1 : m ] represents the last m k elements in y. We also make another child combination z 2 by concatenating y [ 1 : k ] and x [ k + 1 : m ] and put both z 1 and z 2 into the crossover set Z 1 . To construct a mutation set Z 2 , we mutate the children combination z 1 and z 2 as follows. After randomly selecting an index k [ 1 , m ] , we mutate z a ( a { 1 , 2 } ) by changing z a [ k ] from 0 to 1 or vice versa. Then, we add the mutated z 1 and z 2 into the mutation set Z 2 .
After repeating the crossover process and the mutation process n / 2 times, we find the best combination x n in X c Z 1 Z 2 . If f ( x c ) < f ( x n ) , we replace x c with x n . We construct a new population P n by randomly choosing n combinations from X c Z 1 Z 2 and replace P c with P n . We repeat the whole process n g times and output the final x c as  Ω ( t ) . We summarize the MECS sleep decision algorithm in Algorithm 1.
Algorithm 1 MECS Sleep Decision Algorithm
1:At the end of a time slot t:
2:  Input:  { Q i ( t ) , λ ˜ i ( t ) , F i } , i M .
3:  Output: Optimal MECS sleep control vector x c .
4:  Init: n, n g .
5:      Construct X c by randomly selecting n elements from Ω ( t ) .
6:      Get x c = a r g m a x x X c f ( x ) .
7:      i = 0
8:while  i n g   do
9:     Create Y by choosing the n / 2 elements in X c with the highest fittness value.
10:      Z 1 = Z 2 = .
11:      j = 0 .
12:     for  j n  do
13:          Randomly select x and y from Y.
14:          Make z 1 and z 2 by crossing over x and y.
15:           Z 1 = Z 1 { z 1 , z 2 }
16:          Mutate z 1 and z 2 .
17:           Z 2 = Z 2 { mutated z 1 , mutated z 2 }
     Find x n = a r g m a x x X c Z 1 Z 2 f ( x )
18:     if  f ( x c ) < f ( x n )  then
19:           x c = x n
20:     Construct X n by randomly selecting n elements from X c Z 1 Z 2 .
21:     Replace X c with X n
22:Return x c .

5. Performance Evaluation

In this section, we verify the proposed method via simulation studies. We compare the performance of our method with that of a conventional method using the LSTM model as a task arrival rate predictor.

5.1. Simulation Environment Setup

We deploy nine MECSs uniformly across a grid with dimensions of 600 m by 600 m (Figure 3). The service region of each MECS is configured as the surrounding four grids with itself at the center. At the beginning of the simulation, we uniformly deploy 275 users on the topology. Users move around over time according to a random mobility model. In other words, at the start of each time slot, a user changes its moving direction from ( 0 , 2 π ] , according to the Uniform distribution, and selects its speed from the Normal distribution with a mean 15 m per time slot and a variance of 2.5 m per time slot. To maintain the number of users in the topology, we assume that the top and bottom, as well as left and right, of the system topology are connected. Therefore, for example, when a user moves to the right and exits the topology, we add a new user to the left side of the topology.
Figure 3. Simulation topology.
Users offload their tasks to the nearest MECS. Initially, we configure that each user generates one task. After the initialization, we dynamically adjust the task generation rate of each user to change the total load imposed on the system over time. If Q i ( t ) Q t h for all MECS i M at the start of each time slot t, we increase the task generation rate of each user. The increase rate is randomly selected from { 0.05 , 0.10 , 0.15 , 0.20 } for each user. When the load imposed on the MEC system exceeds its capacity, the queue length of at least one MECS exceeds Q t h . Then, after randomly choosing the decrease rate from { 0.05 , 0.10 , 0.15 , 0.20 } for each user, we reduce the task generation rate of each user by the chosen rate. As the task arrival rate decreases, the number of sleeping MECS increases. However, in accordance with the second condition of the problem P1, at least three MECSs must remain active to serve users in the MEC system. Therefore, after the reduction in the task generation rate initiates, we consistently decrease the task arrival rate every time slot. This continues until the active MECS count reaches three. At this point, we resume increasing the user’s task generation rate by randomly selecting the increase rate from { 0.05 , 0.10 , 0.15 , 0.20 } for each user every time slot.
After conducting the data generation process over 50,000 time slots, we construct the training dataset using task generation rates observed in the initial 35,000 time slots. The subsequent 7500 time slots are designated as the validation dataset, and the final 7500 time slots are allocated for the test dataset. In each dataset, a data sample is defined as a pair ( [ x ( t ) , , x ( t h + 1 ) ] , x ( t + 1 ) ) , where x ( t ) = { λ ^ i ( t ) : i M } . These datasets are used to train the STGCN. After completing the training of STGCN, we generate x ( t ) for 1000 time slots via the same data generation process. We then evaluate the performance of the proposed method by inputting them into the trained STGCN.
We set P m = 20 W , α = 0.5 , and F i = 75 tasks per time slot. We also set the maximum queue length of each MECS ( Q M ) to 75 and Q t h = Q M / 2 . We configure the length of the task-load history as h = 12 . For the genetic algorithm, we set n = | X c | = 20 and n g = 20 . We use the publicly known values of the LSM model and STGCN model for configuring their hyperparameters. Specifically, we use the hyperparameters in [39] to configure the hyperparameters of the LSTM model and employ the hyperparameters in [40] to configure the STGCN model. One hidden layer with 64 units is used for the LSTM model. Since one LSTM model has 16,961 parameters and there are nine MECS, a total of 152,649 parameters are used when the LSTM method is used. The STGCN model is composed of two ST-Conv blocks and one output layer. The units in the first ST-Conv block is (64, 32, 64) and the units in the second ST-Conv block is (64, 32, 128). The output layer has (256,128) units. Since one STGCN model is used for all MECSs, the total parameters when STGCN is used is 193,246. For our simulation study, we use the Colab Pro with CUDA version 12.0. We use Nvidia Tesla T4 GPU and 51 GB random access memory. When we run these models, we use Python version 3.10.12, Tensorflow 2.14.0, and Pytorch 2.1.0.

5.2. Task Arrival Rate Prediction Accuracy

To quantitatively compare the accuracy of the task arrival rate prediction, we calculate the mean absolute error (MAE) and the root mean square error (RMSE) and show the results in Table 3. We observe that STGCN can predict the task arrival rate more accurately than LSTM regardless of the location of an MECS in the topology. The difference stems from the fact that, unlike LSTM, which independently considers the temporal correlation of the task arrival rate at each MECS, STGCN comprehensively considers the spatio-temporal relationships of the task arrival rates across all MECSs when predicting the task arrival rate. For example, compared to the LSTM predictor, the STGCN predictor decreases MAE and RMSE by 11.24 % and 11.01 % , respectively, when all MECSs are considered.
Table 3. Comparison of task arrival rate prediction accuracy. MECS 3 is located at the side, MECS 4 is located at the center, and MECS 8 is located at the vertex of the topology. Total represents the case where all the MECSs are considered when calculating MAE and RMSE.
To further understand the prediction behavior, we inspect the distribution of the prediction error (i.e., λ ^ i ( t ) λ ˜ i ( t ) ) in Figure 4. In this figure, for ease of visual comparison, we plot all subfigures with the same ranges for both the x-axis and y-axis. As we can see in Figure 4, the prediction errors are more densely clustered around zero when using STGCN compared to LSTM. In addition, we observe that STGCN exhibits smaller variations in the errors compared to LSTM.
Figure 4. Comparison of task arrival rate error distributions. The term ‘Total’ represent the case where all MECSs are considered. (a) LSTM, MECS 3, (b) STGCN, MECS 3, (c) LSTM, MECS 4, (d) STGCN, MECS 4, (e) LSTM MECS 8, (f) STGCN, MECS 8, (g) LSTM, Total, and (h) STGCN, Total.
We also examine the prediction delay. Since LSTM processes inputs sequentially, the LSTM method takes an average of 450 ms to predict task arrival rates. On the contrary, STGCN processes inputs in parallel by using graph convolutions. Therefore, the time required to predict task arrival rates decreases to as small as 10 ms when STGCN is used. This corresponds to a 93.33 % reduction compared to the LSTM method.

5.3. Performance of Sleep Control Vector Decision Method

We use GA to determine Ω ( t ) . To focus on the influence of GA on the energy consumption and the queue length, we use true task arrival rates (i.e., { λ ^ i ( t ) } ) instead of the predicted task arrival rates (i.e., { λ ˜ i ( t ) } ) when determining Ω ( t ) at the beginning of each time slot. Given λ ^ i ( t ) s, we compare the performance of GA and that of the exhaustive search (ES) method. Since ES explores all possible instances in the solution space, Ω ( t ) found by ES is globally optimal for the given input state.
We measure the difference between E ( t ) when ES is applied and that when GA is used and draw the distribution of the difference in E ( t ) in Figure 5. The circles in the box plot (Figure 5a) represent the outliers. We observe that the distribution of E ( t ) when GA is applied is very similar to that of E ( t ) when ES is applied. Specifically, the Q-values in the box plot are Q 1 = 12.46 , Q 2 = 0.43 , and Q 3 = 0.50 . In addition, we observe in Figure 5b that the difference in E ( t ) is densely concentrated around zero. We also examine the Q-values of E ( t ) when ES and GA were applied, respectively. When ES is used, it is Q 1 = 8.53 , Q 2 = 10.64 , and Q 3 = 12.54 , but while GA is employed, it is Q 1 = 8.61 , Q 2 = 10.69 , and Q 3 = 12.70 . Thus, the difference between Q-values obtained by GA and those when ES is applied is marginal.
Figure 5. Distribution of the difference in E ( t ) between ES and GA. (a) Box plot; (b) Histogram.
In Figure 6, we show the distribution of the differences between the average queue length (i.e., Q ¯ ( t ) = 1 | M | i M Q i ( t ) ) when ES is applied and Q ¯ ( t ) when GA is utilized. When compared to the ideal case where the error is zero, the median value ( Q 2 ) of Q ¯ ( t ) is 0.54. When we inspect the histogram, the average difference in Q ¯ ( t ) is 0.87. To further inspect the variance of Q ¯ ( t ) , we examine the interquartile range (IQR), which is the difference between the 75th percentile and the 25th percentile. In the case of ES, the IQR for Q ¯ ( t ) is 4.40, whereas with GA, the IQR is 3.91, resulting in a difference of 0.49 between them. Considering that Q t h = 37.5 , such a difference can be regarded as very small.
Figure 6. Distribution of the difference in Q ¯ ( t ) between ES and GA. (a) Box plot; (b) Histogram.
We also compare ES and GA in terms of their run time. ES takes an average of 80 ms to find Ω ( t ) while GA takes an average of 29 ms. By using GA, we reduce the time needed to determine Ω ( t ) by 63.75 % .

5.4. Combined Effect

To evaluate the combined effects of the task arrival rate prediction and the sleep control vector determination, we compare the performance of the proposed method (STGCN-GA) with that of the following two techniques. The first method denoted by True-ES is an ideal method that determines Ω ( t ) via the exhaustive search by using the true task arrival rates { λ ^ i ( t ) } . The second method uses GA to determine Ω ( t ) by exploiting the task arrival rate predicted by the LSTM model. Henceforth, we denote the second method as LSTM-GA.
We denote the energy consumed at an MECS i during a time slot t when True-ES is used as e i T E ( t ) . We also denote by e i L G ( t ) the energy consumed at an MECS i during a time slot t when LSTM-GA is used. We represent the energy consumed at an MECS i by our method during a time slot t as e i S G ( t ) . In Figure 7, we show the box plots of δ i e , L G ( t ) = e i T E ( t ) e i L G ( t ) and δ i e , S G ( t ) = e i T E ( t ) e i S G ( t ) . To facilitate the comparison, the ranges of the y-axis in all subfigures are shown to be the same. We observe that the energy consumed by the proposed method is closer to the ideal value obtained by the True-ES method than the energy consumed by LSTM-GA. Specifically, Q 2 of δ i e , L G ( i M ) is −0.52 while Q 2 of δ i e , S G ( i M ) becomes −0.28. In addition, the IQR of δ i e , L G ( i M ) is 14.10 and the IQR of δ i e , S G ( i M ) decreases to 12.74.
Figure 7. Comparison of energy difference distributions for each MECS. (a) Box plots for δ i e , L G ( t ) ; (b) Box plots for δ i e , S G ( t ) .
To investigate the difference between the queue length generated by our method and LSTM-GA and the ideal queue length generated by True-ES, we introduce the following symbols. We denote the difference between Q ¯ ( t ) obtained by True-ES and that acquired by our method as δ i q , S G ( t ) . We also denote the difference between Q ¯ ( t ) obtained by True-ES and Q ¯ ( t ) when LSTM-GA is used as δ i q , L G ( t ) . Then, we show the distribution of δ i q , S G ( t ) and δ i q , L G ( t ) in Figure 8. We observe that δ i q , S G ( t ) is closer to zero than δ i q , L G ( t ) , which means that the proposed method makes the queue length more similar to the ideal Q ¯ ( t ) obtained by Tru-ES compared to LSTM-GA. Specifically, the median of δ i q , L G ( t ) is 0.56 while the median of δ i q , S G ( t ) is 0.47, which corresponds to a 16.07 % reduction. In addition, the IQR of δ i q , L G ( t ) is 4.20 while the IQR of δ i q , S G ( t ) is 3.87. Thus, the proposed method decreases the IQR of the queue length difference by 7.86 % .
Figure 8. Comparison of queue length difference distributions. (a) Box plots for δ i q , L G ( t ) ; (b) Box plots for δ i q , S G ( t ) .
To compare the time taken for sleep control, we examine the end-to-end inference time, starting from predicting the task input rates to determining the sleep control vector. In Figure 9, we show the distribution of the end-to-end inference time as the box plots. When we inspect the Q-values, LSTM-GA obtains Q 1 = 440 ms, Q 2 = 449 ms, and Q 3 = 463 ms. The proposed method reduces the Q-values significantly. When our method is used, Q 1 = 18 ms, Q 2 = 21 ms, and Q 3 = 22 ms. In other words, compared to LSTM-GA, the proposed method decreases the median end-to-end inference time from 449 ms to 21 ms, which is a 95.32 % reduction. With respect to the IQR, our method reduces the IQR of the end-to-end inference time from 23 ms to 4 ms, which corresponds to an 82.16 % decrease.
Figure 9. End-to-end inference time distribution comparison. (a) LSTM-GA; (b) Proposed method.

6. Conclusions and Future Works

In this paper, we address the MECS sleep control problem in a multi-access edge computing system. To increase the energy efficiency of an MEC system while bounding the queue length of each MECS within a given value, we comprehensively exploit the spatio-temporal correlation structure in the task arrival rate distribution among the MECSs by using the STGCN model. By using only one STGCN model, we predict the task arrival rate for each MECS at the same time. In addition, we reduce the time to find an optimal sleep control vector given a system load state. We evaluate the performance of the proposed method via extensive simulation studies. The results verify that compared to a conventional method based on the LSTM model, the proposed method increases the energy efficiency of a system and decreases the average queue length.
As our future works, we will investigate methods to deal with the prediction error involved in the task arrival rate prediction. Any predictor introduces error inevitably. We will scrutinize the trend of errors that the STGCN predictor causes and analyze the tradeoff between energy consumption and queue length according to the type of errors. Then, we will study ways to improve the performance by controlling errors adaptively according to the operation state of an MEC system.

Author Contributions

Methodology, J.P.; Software, M.K.; Validation, M.K.; Writing—original draft, J.P.; Supervision, J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Research Foundation of Republic of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2022R1F1A1065371).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Freitag, C.; Lee, M.B.; Widdicks, K.; Knowles, B.; Blair, G.; Friday, A. The Climate Impact of ICT: A Review of Estimates, Trends and Regulations, Lancaster University. arXiv 2020, arXiv:2102.02622. [Google Scholar]
  2. Liy, N.; Zhuy, X.; Liy, Y.; Wangy, L.; Zhai, L. Service Caching and Task Offloading of Internet of Things Devices Guided by Lyapunov Optimization. In Proceedings of the 2022 IEEE ISPA/BDCloud/SocialCom/SustainCom, Melbourne, Australia, 17–19 December 2022. [Google Scholar]
  3. Yan, J.; Bi, S.; Duan, L.; Zhang, Y.-A. Pricing-Driven Service Caching and Task Offloading in Mobile Edge Computing. IEEE Trans. Wirel. Commun. 2021, 20, 4495–4512. [Google Scholar] [CrossRef]
  4. Bae, H.; Park, J. Proactive Service Caching in a MEC System by Using Spatio-Temporal Correlation among MEC Servers. Appl. Sci. 2023, 13, 12509. [Google Scholar] [CrossRef]
  5. Wu, H.; Chen, J.; Nguyen, T.N.; Tang, H. Lyapunov-Guided Delay-Aware Energy Efficient Offloading in IIoT-MEC Systems. IEEE Trans. Ind. Inform. 2023, 19, 2117–2128. [Google Scholar] [CrossRef]
  6. Wang, Y.; Chen, M.; Li, Z.; Hu, Y. Joint Allocations of Radio and Computational Resource for User Energy Consumption Minimization Under Latency Constraints in Multi-Cell MEC Systems. IEEE Trans. Veh. Technol. 2023, 72, 3304–3320. [Google Scholar] [CrossRef]
  7. Thananjeyan, S.; Chan, C.A.; Wong, E.; Nirmalathas, A. Mobility-Aware Energy Optimization in Hosts Selection for Computation Offloading in Multi-Access Edge Computing. IEEE Open J. Commun. Soc. 2020, 1, 1056–1065. [Google Scholar] [CrossRef]
  8. Li, S.; Sun, W.; Sun, Y.; Huo, Y. Energy-Efficient Task Offloading Using Dynamic Voltage Scaling in Mobile Edge Computing. IEEE Trans. Netw. Sci. Eng. 2021, 8, 588–598. [Google Scholar] [CrossRef]
  9. Wang, S.; Zhang, X.; Yan, Z.; Wang, W. Cooperative Edge Computing With Sleep Control Under Nonuniform Traffic in Mobile Edge Networks. IEEE Internet Things J. 2019, 6, 4295–4306. [Google Scholar] [CrossRef]
  10. Amer, A.A.; Talkhan, I.E.; Ismail, T. Optimal Power Consumption on Distributed Edge Services Under Non-Uniform Traffic with Dual Threshold Sleep/Active Control. In Proceedings of the 3rd Novel Intelligent and Leading Emerging Sciences Conference (NILES), Giza, Egypt, 23–25 October 2021. [Google Scholar]
  11. Xu, J.; Wu, X.; Huang, Q.; Sun, P. How Should the Server Sleep?—Age-Energy Tradeoff in Sleep-Wake Server Systems. IEEE Trans. Green Commun. Netw. 2023, 7, 1620–1634. [Google Scholar] [CrossRef]
  12. Park, J.; Lim, Y. Bio-Inspired Sleep Control for Improving the Energy Efficiency of a MEC System. Appl. Sci. 2023, 13, 2620. [Google Scholar] [CrossRef]
  13. Wang, Q.; Xie, Q.; Yu, N.; Huang, H.; Jia, X. Dynamic Server Switching for Energy Efficient Mobile Edge Networks. In Proceedings of the 2019 IEEE International Conference on Communications (ICC), Shanghai, China, 20–24 May 2019. [Google Scholar]
  14. Wang, Q.; Tan, L.T.; Hu, R.Q.; Qian, Y. Hierarchical Energy-Efficient Mobile-Edge Computing in IoT Networks. IEEE Internet Things J. 2020, 7, 11626–11639. [Google Scholar] [CrossRef]
  15. Yu, B.; Yin, H.; Zhu, Z. Spatio-Temporal Graph Convolutional Networks: A Deep Learning Framework for Traffic Forecasting. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018. [Google Scholar]
  16. Suárez-Varela, J.; Almasan, P.; Ferriol-Galmés, M.; Rusek, K.; Geyer, F.; Cheng, X.; Shi, X.; Xiao, S.; Scarselli, F.; Cabellos-Aparicio, A.; et al. Graph Neural Networks for Communication Networks: Context, Use Cases and Opportunities. IEEE Netw. 2023, 37, 146–153. [Google Scholar] [CrossRef]
  17. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph Neural Networks: A Review of Methods and Applications. KeAi AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  18. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2001, 32, 4–24. [Google Scholar] [CrossRef] [PubMed]
  19. Lei, B.; Zhang, P.; Suo, Y.; Li, N. SAX-STGCN: Dynamic Spatio-Temporal Graph Convolutional Networks for Traffic Flow Prediction. IEEE Access 2022, 10, 107022–107031. [Google Scholar] [CrossRef]
  20. Wu, W.; Tu, F.; Niu, M.; Yue, Z.; Liu, L.; Wei, S.; Li, X.; Hu, Y.; Yin, S. STAR: An STGCN ARchitecture for Skeleton-Based Human Action Recognition. IEEE Trans. Circuits Syst. I 2023, 70, 2370–2383. [Google Scholar] [CrossRef]
  21. Nguyen, P.-D.; Ha, V.N.; Le, L.B. Computation Offloading and Resource Allocation for Backhaul Limited Cooperative MEC Systems. In Proceedings of the 2019 IEEE 90th Vehicular Technology Conference (VTC2019-Fall), Honolulu, HI, USA, 22–25 September 2019. [Google Scholar]
  22. Vu, T.T.; Huynh, N.V.; Hoang, D.T.; Nguyen, D.N.; Dutkiewicz, E. Offloading Energy Efficiency with Delay Constraint for Cooperative Mobile Edge Computing Networks. In Proceedings of the 2018 IEEE Global Communications Conference (GLOBECOM), Abu Dhabi, United Arab Emirates, 9–13 December 2018. [Google Scholar]
  23. Wang, J.; Feng, D.; Zhang, S.; Liu, A.; Xia, X.-G. Joint Computation Offloading and Resource Allocation for MEC-enabled IoT Systems with Imperfect CSI. IEEE Internet Things J. 2020, 8, 3462–3475. [Google Scholar] [CrossRef]
  24. Yang, J.; Shah, A.A.; Pezaros, D. A Survey of Energy Optimization Approaches for Computational Task Offloading and Resource Allocation in MEC Networks. Electronics 2023, 12, 3548. [Google Scholar] [CrossRef]
  25. Moon, S.; Park, J.; Lim, Y. Task Migration Based on Reinforcement Learning in Vehicular Edge Computing. Wirel. Commun. Mob. Comput. 2021, 2021, 9929318. [Google Scholar] [CrossRef]
  26. Park, J.; Lim, Y. Balancing Loads among MEC Servers by Task Redirection to Enhance the Resource Efficiency of MEC Systems. Appl. Sci. 2021, 11, 7589. [Google Scholar] [CrossRef]
  27. Li, M.; Zhang, Q.; Liu, F. Finedge: A Dynamic Cost-efficient Edge Resource Management Platform for NFV Network. In Proceedings of the IEEE/ACM 28th International Symposium on Quality of Service (IWQoS’20), Hangzhou, China, 15–17 June 2020. [Google Scholar]
  28. Park, J.; Lim, Y. Online Service-Time Allocation Strategy for Balancing Energy Consumption and Queuing Delay of a MEC Server. Appl. Sci. 2022, 12, 4539. [Google Scholar] [CrossRef]
  29. Pan, L.; Wang, L.; Chen, S.; Liu, F. Retention-Aware Container Caching for Serverless Edge Computing. In Proceedings of the IEEE Conference on Computer Communications (INFOCOM’22), London, UK, 2–5 May 2022. [Google Scholar]
  30. Wu, H.; Fan, Y.; Wang, Y.; Ma, H.; Xing, L. A Comprehensive Review on Edge Caching from the Perspective of Total Process: Placement, Policy and Delivery. Sensors 2021, 21, 5033. [Google Scholar] [CrossRef] [PubMed]
  31. Malazi, H.T.; Chaudhry, S.R.; Kazmi, A.; Palade, A.; Cabrera, C.; White, G.; Clarke, S. Dynamic Service Placement in Multi-Access Edge Computing: A Systematic Literature Review. IEEE Access 2022, 10, 32639–32688. [Google Scholar] [CrossRef]
  32. Sonkoly, B.; Czentye, J.; Szalay, M.; Németh, B.; Toka, L. Survey on Placement Methods in the Edge and Beyond. IEEE Commun. Surv. Tutor. 2021, 23, 2590–2629. [Google Scholar] [CrossRef]
  33. Wu, T.; Li, X.; Ji, H.; Zhang, H. An Energy-Efficient Sleep Management Algorithm for UDN with Edge Caching. In Proceedings of the 2017 IEEE Globecom Workshops (GC Wkshps), Singapore, 4–8 December 2017. [Google Scholar]
  34. Ranadheera, S.; Maghsudi, S.; Hossain, E. Computation Offloading and Activation of Mobile Edge Computing Servers: A Minority Game. IEEE Wirel. Commun. Lett. 2018, 7, 688–691. [Google Scholar] [CrossRef]
  35. Chen, L.; Zhou, S.; Xu, J. Energy Efficient Mobile Edge Computing in Dense Cellular Networks. In Proceedings of the 2017 IEEE International Conference on Communications (ICC), Paris, France, 21–25 May 2017. [Google Scholar]
  36. Merluzzi, M.; Pietro, N.; Lorenzo, P.; Strinati, E.C.; Barbarossa, S. Network Energy Efficient Mobile Edge Computing with Reliability Guarantees. In Proceedings of the 2019 IEEE Global Communications Conference (GLOBECOM), Waikoloa, HI, USA, 9–13 December 2019. [Google Scholar]
  37. Ma, R.; Zhou, X.; Zhang, H.; Yuan, D. Joint Optimization of Energy Consumption and Latency Based on DRL: An Edge Server Activation and Task Scheduling Scheme in IIoT. In Proceedings of the 14th International Conference on Wireless Communications and Signal Processing (WCSP), Nanjing, China, 1–3 November 2022. [Google Scholar]
  38. Defferrard, M.; Bresson, X.; Vandergheynst, P. Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering. In Proceedings of the Thirtieth Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  39. Brownlee, J. A Gentle Introduction to LSTM Autoencoders. Available online: https://machinelearningmastery.com/lstm-autoencoders (accessed on 27 August 2020).
  40. Yu, G. STGCN-PyTorch. Available online: https://github.com/Aguin/STGCN-PyTorch (accessed on 27 December 2019).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.