Next Article in Journal
Interval Type-2 Fuzzy PID Controller Using Disassembled Gradational Optimization
Next Article in Special Issue
Heuristic Path Search and Multi-Attribute Decision-Making-Based Routing Method for Vehicular Safety Messages
Previous Article in Journal
Research on Similarity Simulation Experiment of Mine Pressure Appearance in Surface Gully Working Face Based on BOTDA
Previous Article in Special Issue
Capture-Aware Dense Tag Identification Using RFID Systems in Vehicular Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computing Offloading Based on TD3 Algorithm in Cache-Assisted Vehicular NOMA–MEC Networks

1
School of Information Engineering, East China Jiaotong University, Nanchang 330013, China
2
School of Information Engineering, Nanchang University, Nanchang 330031, China
3
School of Information Science and Engineering, Southeast University, Nanjing 210096, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(22), 9064; https://doi.org/10.3390/s23229064
Submission received: 10 October 2023 / Revised: 2 November 2023 / Accepted: 7 November 2023 / Published: 9 November 2023

Abstract

:
In this paper, in order to reduce the energy consumption and time of data transmission, the non-orthogonal multiple access (NOMA) and mobile edge caching technologies are jointly considered in mobile edge computing (MEC) networks. As for the cache-assisted vehicular NOMA–MEC networks, a problem of minimizing the energy consumed by vehicles (mobile devices, MDs) is formulated under time and resource constraints, which jointly optimize the computing resource allocation, subchannel selection, device association, offloading and caching decisions. To solve the formulated problem, we develop an effective joint computation offloading and task-caching algorithm based on the twin-delayed deep deterministic policy gradient (TD3) algorithm. Such a TD3-based offloading (TD3O) algorithm includes a designed action transformation (AT) algorithm used for transforming continuous action space into a discrete one. In addition, to solve the formulated problem in a non-iterative manner, an effective heuristic algorithm (HA) is also designed. As for the designed algorithms, we provide some detailed analyses of computation complexity and convergence, and give some meaningful insights through simulation. Simulation results show that the TD3O algorithm could achieve lower local energy consumption than several benchmark algorithms, and HA could achieve lower consumption than the completely offloading algorithm and local execution algorithm.

1. Introduction

With the rapid development of information and communication technologies, the data traffic generated by vehicles (mobile devices, MDs) has also significantly increased [1]. For wireless communication networks, more spectrum resources are required for data traffic transmission [2,3,4,5]. In addition, higher computing power is required by MDs for supporting large amounts of task calculation. However, due to the limited battery capacity of MDs, it may be challenging to process these computation tasks for them. By deploying edge computing servers at base stations (BSs), mobile edge computing (MEC) can support MDs in processing tasks at the adjacent edge servers [6,7]. Compared with cloud computing (CC), which requires tasks to be uploaded to a remote cloud, MEC can provide additional computing resources for MDs within its coverage area and thus reduce their computing overhead [8,9,10,11,12,13,14].
Although the edge servers can reduce the computing overhead of MDs by providing more computing resources, the extra time and energy consumption caused by offloading tasks through wireless channels cannot be ignored, especially for high-size computation tasks. In order to further reduce the time and energy consumption caused by offloading tasks, edge caching technology is also introduced into MEC networks. By caching tasks of MDs at edge servers in advance, the overhead caused by offloading tasks could be greatly reduced [15,16,17,18,19].
To upload tasks from MDs to edge servers, orthogonal multiple access (OMA) is often used, but it may be greatly challenging to provide a high transmission rate and support massive connections. As another type of resource utilization management, non-orthogonal multiple access (NOMA) technologies can let multiple users share the same frequency bands, achieve higher spectral efficiency and support extensive connections [20,21,22,23]. It is evident that NOMA is a good type of resource utilization management for reducing the cost of task transmission in MEC networks.
Although the application of caching and NOMA technologies in MEC networks can reduce time and energy consumption, such a framework will make the design of computation offloading and edge caching schemes more complex. To the best of our knowledge, until now, how to jointly perform the device association, computation offloading, edge caching, subchannel selection and resource allocation is still an important and open topic in cache-assisted NOMA–MEC networks.

1.1. Related Work

So far, a lot of work has been conducted on joint computation offloading and resource optimization in NOMA–MEC networks. In [20], joint radio and computation resource allocation was optimized to maximize the offloading energy efficiency in NOMA–MEC-enabled IoT networks, and a solution based on a multi-layer iterative algorithm was proposed. In [21], local computation resource, offloading ratio, uplink transmission time and power and subcarrier assignment were jointly optimized to minimize the sum of weighted energy consumed by users in NOMA–MEC networks, and some effective iterative algorithms were designed for single-user and multi-user cases. In [24], joint task offloading, power allocation and computing resource allocation were optimized to achieve delay minimization using a deep reinforcement learning (DRL) algorithm in NOMA–MEC networks. In [25], the joint optimization of offloading decisions, local and edge computing resource allocation and power and subchannel allocation were realized to minimize energy consumption in heterogeneous NOMA–MEC networks, and an effective iterative algorithm was designed. In [26], power and computation resource allocations were jointly optimized to minimize overall computation and transmission delay for massive MIMO and NOMA-assisted MEC systems, and a solution based on an interior-point algorithm was given. In [27], the channel resource allocation and computation offloading policy were jointly optimized to minimize the sum of weighted energy and latency in NOMA–MEC networks, and some efficient solutions were found using a DRL algorithm based on actor–critic and deep Q-network (DQN) methods.
To further reduce the offloading time and energy consumption, edge caching technology is introduced into conventional MEC networks. Such a framework has attracted more and more attention. In [28], the offloading and caching decisions, uplink power and edge computing resources were jointly optimized to minimize the sum of weighted local processing time and energy consumption in two-tier cache-assisted MEC networks, and a distributed collaborative iterative algorithm was proposed. In [29], a problem of adaptive request scheduling and cooperative service caching was studied in cache-assisted MEC networks. After formulating the optimization problems as partially observable Markov decision process (MDP) problems, an online DRL algorithm was proposed to improve the service hitting ratio and latency reduction rate. In [30], optimal offloading and caching strategies were established to minimize overall delay and energy consumption of all regions using a deep deterministic policy gradient (DDPG) framework in cache-assisted multi-region MEC networks. In [31], joint MD association and resource allocation were performed to minimize the sum of MDs’ weighted delay in heterogeneous cellular networks with MEC and edge caching functions, and an effective iterative algorithm was developed using coalitional game and convex optimization theorems. In [32], to minimize the content transmission delay in vehicular edge computing networks, a cooperative vehicular edge computing and caching scheme based on asynchronous federated and deep reinforcement learning was proposed to predict the popular content and the optimal cooperative caching location of the content. In [33], to reduce the cost of the cloud service center through the asynchronous advantage actor–critic algorithm, the offloading decision, service caching and resource allocation strategies were jointly optimized in the three-tier mobile cloud–edge computing structure combining computation offloading and service caching mechanisms. In [34], a logistic function-based service reliability probability (SRP) estimation model was built, and the average SRP maximization problem of a virtual machine-based edge computing server was studied for such a model. At last, a low-complexity heuristic alternative optimization algorithm was proposed.
To enhance spectral efficiency and support massive connections, NOMA technology has attracted increasing attention in cache-assisted MEC networks. In [35], the multi-agent deep deterministic policy gradient method was used to dynamically optimize the user association, power control and cache placement of BSs and satellites to improve the network energy efficiency in a NOMA-enabled satellite integrated with a terrestrial network scenario. In [36], joint optimization of offloading and caching decisions and computation resource allocation was performed to maximize long-term reward in cache-assisted NOMA–MEC networks under the predicted task popularity, and single-agent and multi-agent Q-learning algorithms were proposed to find feasible solutions. In [37], joint optimization of offloading and caching decisions was performed to minimize the system delay in cache-assisted NOMA–MEC networks, and a multi-agent DQN algorithm was used for finding efficient solutions under the predicted popularity. In [38], local task processing time was minimized by jointly optimizing offloading and caching decisions and the allocation of edge computing resources and uplink power in cache-assisted NOMA–MEC networks with single BS, and the blocking successive upper-bound minimization method was utilized to achieve efficient solutions.
Although the framework of cache-assisted (vehicular) NOMA–MEC networks can greatly reduce the task processing time and energy consumption and support massive connections, there exist very few relevant efforts. Unlike the above-mentioned work, we jointly optimize the edge computing resource allocation, subchannel selection, device association, offloading and caching decisions for the cache-assisted vehicular NOMA–MEC networks with multiple BSs, minimizing the energy consumed by MDs under time and resource constraints. In addition, unlike existing efforts, we develop an effective dynamic joint computation offloading and task-caching algorithm based on the twin-delayed deep deterministic policy gradient algorithm (TD3) to find efficient solutions, named the TD3-based offloading (TD3O) algorithm.

1.2. Contribution and Organization

In this paper, we jointly optimize the edge computing resource allocation, subchannel selection, device association, offloading and caching decisions in cache-assisted vehicular NOMA–MEC networks, minimizing the energy consumed by MDs under time and resource constraints. Specifically, the main contributions and work of this paper can be listed as follows.
  • Edge computing resource allocation, subchannel selection, device association, computation offloading and edge caching are jointly performed in cache-assisted vehicular NOMA–MEC networks. To the best of our knowledge, work that concerns subchannel selection is a new investigation for cache-assisted vehicular NOMA–MEC networks with multi-server scenarios. Meanwhile, as far as this problem is concerned, the goal is to minimize the energy consumed by MDs under the constraints of time, computing resources, caching capacity, the number of MDs associated with each BS and the number of MDs associated with each subchannel. As far as we know, such an optimization problem is a new concentration in cache-assisted vehicular NOMA–MEC networks.
  • We design effective algorithms to find feasible solutions to the formulated problem. Considering that the formulated problem is in a mixed-integer, nonlinear, multi-constraint form, a simple map between actions and actual policies in a conventional twin-delayed deep deterministic policy gradient (TD3) algorithm cannot be well applied. In addition, too large an action space will cause the TD3 algorithm to fail to search for correct actions and thus fail to converge. In view of these concerns, we develop an effective TD3O algorithm integrating with the AT algorithm to solve the formulated problem. Moreover, in order to solve this problem in a non-iterative manner, an effective heuristic algorithm (HA) is also designed.
  • Performance analyses of the designed algorithms. Some analyses are made for the computation complexity and convergence of the designed algorithms in detail. In addition, some meaningful simulation analyses are also made by introducing other benchmark algorithms for comparison, and some good results and insights are achieved.
The rest of the paper is organized as follows. Section 2 introduces the system model. Section 3 formulates a problem of minimizing local energy consumption in cache-assisted vehicular NOMA–MEC networks. Section 4 designs the HA and TD3O algorithm. Section 5 gives the computation complexity and convergence analyses for the designed algorithms. Section 6 investigates the performance of the designed algorithms through simulation. Section 7 gives conclusions and discussions.

2. System Model

2.1. Network Model

Figure 1 shows the cache-assisted vehicular NOMA–MEC networks. In such network, there exist M MDs, and the index set of them is denoted as M = { 1 , 2 , , M } ; B BSs are deployed, and the index set of them is given by I = 1 , 2 , , I . In addition, each BS is equipped with one edge computing server and one edge caching server, and these BSs connect to each other through wired links. We assume that each MD has one computation task at any timeslot, which can be processed by itself, its associated BS or another auxiliary BS selected by this associated BS. When tasks have been cached at the BSs used for processing them, they do not need to be uploaded to these BSs; when the associated BSs have not cached tasks, MDs need to upload tasks to these BSs; when the auxiliary BSs have not cached tasks, the associated BSs need to upload tasks to their selected auxiliary BSs.
Assume that the association index between MD m and BS i is x m , i 0 , 1 , where X = x m , i m M , i I . x m , i = 1 if MD m is associated with BS i, otherwise x m , i = 0 . In addition, we assume that the caching index of the task of MD m at BS i is denoted as y m , i 0 , 1 , where Y = y m , i | m M , i I . y m , i = 1 if the task of MD m is cached at BS i, otherwise y m , i = 0 . We also assume that the offloading (execution) index of the task of MD m at BS i is denoted as u m , i , where U = u m , i | m M , i I . u m , i = 1 if the task of MD m is executed at BS i, otherwise u m , i = 0 . At last, we assume that the association index between MD m and subchannel k of BS i is denoted as z m , i , k , where Z = z m , i , k | m M , i I , k K . If x m , i ( 1 y m , i ) ( 1 y m , i ¯ ) u m , i ¯ = 1 or x m , i ( 1 y m , i ) u m , i = 1 under i ¯ i , MD m can select (be associated with) some subchannel k of BS i, which means z m , i , k = 1 . Otherwise, the subchannel k of BS i cannot be selected by MD m, which means z m , i , k = 0 .

2.2. Communication Model

In this paper, the system bandwidth W is divided into K subchannels with equal bandwidth, which are indexed by K = 1 , 2 , , K . These subchannels can be shared by different MDs through NOMA. Significantly, each MD can occupy at most one subchannel, the number of MDs selecting each subchannel cannot exceed the upper limit ρ , and the number of MDs associated with any BS that need to upload tasks should be less than or equal to the number of subchannels K [37].
As revealed in [39], the successive interference cancellation (SIC) technology in NOMA technology can effectively reduce the interference between MDs in the same subchannel. The channel gains of MDs sharing the same subchannel of a BS should be sorted in descending order at first, and then the uplink NOMA signals received by this BS can be decoded in this order. We assume that M k S C is the set of MDs selecting subchannel k, and o m , i , k represents the sequence number of channel gain between MD m and BS i on subchannel k. When MD i and MD m access the subchannel k of BS i simultaneously and the channel gain h j , i , k between MD j and BS i on subchannel k is lower than the channel gain h m , i , k between MD m and BS i on subchannel k, o j , i , k < o m , i , k is satisfied. Then, the signal of MD m is decoded but the signal of MD j will be treated as noise. Therefore, when MD m selects subchannel k of BS i, its uplink data rate r m , i , k can be given by
r m , i , k = W log 2 1 + p m h m , i , k / Γ m , i , k + σ 2 / K ,
where Γ m , i , k = j M k s c / { m } : o j , i , k < o m , i , k p j h j , i , k is the interference caused by other MDs (excluding MD m) sharing subchannel k of BS i through NOMA; p m is the transmission power of MD m; σ 2 is the noise power. When MD m is decoded, it is no longer regarded as interference in the subchannel, and the device with the maximum channel gain among the remaining MDs in the current subchannel is decoded in the same way until all MDs of the current subchannel are decoded.

2.3. Caching and Offloading Models

In this paper, we assume that any MD m has a time-sensitive task denoted as L m = d m , c m , τ m max at each timeslot, where d m is the data size of the task of MD m, c m is the number of CPU cycles required to complete a one-bit task, and τ m max is the maximum task processing time of MD m.
Figure 2 illustrates the caching and offloading models. At each timeslot, BSs precache the tasks for processing at the next timeslot. When MD m is associated with BS i, it first checks whether the associated BS has cached the corresponding task. If i I u m , i = 0 , the task of MD m is calculated by itself, e.g., MD 1 in Figure 2. If x m , i y m , i u m , i = 1 , the task of MD m can be directly calculated at its associated BS i, and the results will be fed back from BS i to MD m, e.g., MD 2 in Figure 2. If x m , i y m , i ( 1 y m , i ¯ ) u m , i ¯ = 1 , the task of MD m is offloaded from its associated BS i to another auxiliary BS i ¯ i for computing through a wired link, e.g., MD 3 in Figure 2. If x m , i y m , i y m , i ¯ u m , i ¯ = 1 , the task of MD m can be directly calculated at auxiliary BS i ¯ i , e.g., MD 4 in Figure 2. If x m , i ( 1 y m , i ) u m , i = 1 , the task of MD m will be offloaded to its associated BS i for computing, e.g., MD 5 in Figure 2. If x m , i ( 1 y m , i ) ( 1 y m , i ¯ ) u m , i ¯ = 1 , the task of MD m first needs to be offloaded to its associated BS i, and then it is transmitted from this BS to another auxiliary BS i ¯ i for computing through a wired link, e.g., MD 6 in Figure 2. If x m , i ( 1 y m , i ) y m , i ¯ u m , i ¯ = 1 , the task of MD m can be directly calculated at an auxiliary BS i ¯ i , e.g., MD 4 in Figure 2.

2.3.1. Local Computing

If i I u m , i = 0 is satisfied, the task of MD m should be executed locally, and the processing time and energy consumption are, respectively, given by
τ m loc = c m d m c m d m f m loc f m loc ,
ε m loc = ξ c m d m f m loc 2 ,
where f m loc is the computing capacity of MD m, and ξ is an energy-consumption coefficient depending on the hardware architecture.

2.3.2. Task Transmission

If x m , i ( 1 y m , i ) ( 1 y m , i ¯ ) u m , i ¯ = 1 or x m , i ( 1 y m , i ) u m , i = 1 are satisfied under i ¯ i , the task of MD m should be, respectively, uploaded to BS i ¯ or i for execution through NOMA. Then, the uploading time and energy consumption of MD m, respectively, are given by
τ m trs = i I k K z m , i , k d m / r m , i , k ,
ε m trs = p m τ m trs .
In addition, if x m , i ( 1 y m , i ) ( 1 y m , i ¯ ) u m , i ¯ = 1 or x m , i y m , i ( 1 y m , i ¯ ) u m , i ¯ = 1 is satisfied under i ¯ i , the task of MD m should be transmitted from its associated BS i to an auxiliary BS i ¯ through a wired link, and the corresponding time is given by
τ m bh = d m d m r bh r bh ,
where r bh is the backhualing rate between any two BSs.
In this paper, we mainly concentrate on the energy consumption of MDs but not the energy consumed by BSs. In addition, the downlink transferring time of results is often ignored since they are fairly small [40].

2.3.3. Edge Computing

When MD m executes its task at BS i, the task processing time at this BS can be given by
τ m , i exe = c m d m c m d m f m , i f m , i ,
where f m , i is the computing capacity allocated to MD m by BS i.

2.3.4. Task Processing Time and Energy Consumption

Then, the total time used for processing the task of MD m can be given by
τ m tot = i I ( 1 i I u m , i τ m loc + x m , i 1 y m , i i ¯ I { i } u m , i ¯ 1 y m , i ¯ τ m trs + x m , i 1 y m , i i ¯ I { i } u m , i ¯ 1 y m , i ¯ τ m bh + x m , i 1 y m , i i ¯ I { i } u m , i ¯ τ m , i ¯ exe + x m , i y m , i i ¯ I { i } u m , i ¯ 1 y m , i ¯ τ m bh + x m , i y m , i i ¯ I { i } u m , i ¯ τ m , i ¯ exe + x m , i 1 y m , i u m , i τ m trs + τ m , i exe + x m , i y m , i u m , i τ m , i exe ) .
On the right side of the equality sign in (8), the first item represents the local executing time; the second item is the time used for uploading the task from MD m to the associated BS i, which does not cache this task and further transmits it to auxiliary BS for computing; the third item is the time used for transmitting the task from the associated BS i to another auxiliary BS, where these two BSs do not cache this task; the fourth item is the time used for executing the task of MD m at an auxiliary BS, where the associated BS does not cache this task; the fifth item is the time used for transmitting the task from the associated BS i to another auxiliary BS, where the associated BS caches this task but the auxiliary BS does not; the sixth item is the time used for executing the task of MD m at an auxiliary BS, where the associated BS caches this task; the seventh item includes the time used for transmitting the task from MD m to the associated BS i, which does not cache this task, and the time used for executing the task of MD m at this BS; the eighth item is the time used for executing the task of MD m at the associated BS i, which caches this task.
Then, the total local energy consumption used for processing the task of MD m can be given by
ε m tot = i I ( 1 i I u m , i ε m loc + x m , i 1 y m , i i ¯ I { i } u m , i ¯ 1 y m , i ¯ ε m trs + x m , i 1 y m , i u m , i ε m trs ) ,
on the right side of equality sign in (9), the first item represents the local executing energy consumption; the second item is the energy consumption caused by offloading the task from MD m to its associated BS i, which further transmits this task to auxiliary BS i ¯ i for computing; the third item is the energy consumption caused by transmitting the task from MD m to the associated BS i, which does not cache this task.

3. Problem Formulation

Until now, we have formulated a problem of minimizing local energy consumption at each given period. Specifically, under the constraints of time, computing resources, caching capacity, the number of MDs associated with each BS and the number of MDs associated with each subchannel, we jointly optimized the edge computing resource allocation, subchannel selection, device association, offloading and caching decisions to minimize the energy consumed by MDs in cache-assisted vehicular NOMA–MEC networks. Mathematically, this is formulated as
P 1 : min X , Y , U , Z , F m M ε m tot s . t . C 1 : τ m tot τ m , m M , C 2 : i I x m , i = 1 , m M , C 3 : i I k K z m , i , k 1 , m M , C 4 : m M k K z m , i , k K , i I , C 5 : i I u m , i 1 , m M , C 6 : x m , i 0 , 1 , m M , i I , C 7 : y m , i 0 , 1 , m M , i I , C 8 : z m , i , k 0 , 1 , m M , i I , k K , C 9 : u m , i 0 , 1 , m M , i I , C 10 : m M y m , i d m ϑ i , i I , C 11 : m M i I z m , i , k ρ , k K , C 12 : m M u m , i f m , i f i BS , i I ,
where F = f m , i m M , i I ; the constraint C 1 gives the maximum task processing time of MD m; C 2 and C 6 indicate that any MD m can select only one BS; C 3 and C 8 indicate that any MD m can occupy at most one subchannel; C 4 and C 8 mean that the number of MDs selecting any BS that needs to upload tasks should be less than or equal to the number of subchannels; C 5 and C 9 mean that any MD m can select at most one BS to execute its task; C 7 and C 10 indicate that the data size of tasks cached at BS i does not exceed the caching capacity ϑ i of this BS; C 8 and C 11 show that the number of MDs selecting a subchannel cannot exceed its upper limit; C 7 and C 12 reveal that the total computing capacity allocated to MDs by BS i cannot exceed the computing capacity of this BS.

4. Algorithm Design

As previously mentioned, the optimization problem P1 refers to minimizing local energy consumption within a given period. In view of this, we adopt the DRL algorithm to solve it. DRL is based on MDP, which implements the environment-based output of agent policy in MDP through neural networks, maximizing certain rewards. Considering that the overestimation of some conventional DRL algorithms (e.g., DQN and DDPG), the TD3 algorithm has been widely advocated because it can overcome well the problems of the above algorithms and achieve more stable output [41,42]. The main features of the TD3 algorithm are adding a new neural network and reducing the training frequency of the network based on DDPG.
The problem P1 has both continuous and discrete variables, and the solution space formed by the combination of all variables is very large, which is not a suitable scenario for the DQN algorithm to solve discrete space problems. Therefore, we use the TD3 algorithm, which can solve the continuous solution space problems. Considering that a simple mapping between only the decision of the algorithm and the actual strategy will fail to achieve convergence because of there being too many feasible strategies and the inability to search for the correct one, we develop an effective TD3O algorithm integrated with the AT algorithm to solve the problem P1.

4.1. MDP Used for Describing Problem P1

Considering that the optimization problem P1 needs to be tackled within a given period, in order to apply TD3O to the problem P1, such a period is divided into T timeslots and denoted as T = { 1 , 2 , , T } . Furthermore, the problem of joint computing offloading, task caching and resource allocation is described as a MDP, the state space, action space and reward function are defined as follows.
State space: At each timeslot, the state space contains the information used for decisions made by the network. Here, the state s t at timeslot t can be denoted as s t = { D ¯ t + 1 , Y ¯ t } . The detailed definitions can be found as follows.
  • D ¯ t + 1 = d ¯ m t + 1 m M are the standardized data sizes of tasks of MDs at timeslot t + 1 , where
    d ¯ m t = d m t d min t d max t d min t ,
    d min t is the minimum data size of the tasks of all MDs at timeslot t, and d max is the maximum data size of the tasks of all MDs at timeslot t.
  • Y ¯ t = y ¯ m t m M are the task caching decision factors at BSs at timeslot t, where y ¯ m 0 , 1 .
Action space: At each timeslot, the action space refers to the decisions made by the network according to the state s t . The action a t at timeslot t can be denoted as a t = X ¯ t , Y ¯ t + 1 , Z ¯ t , U ¯ t , F ¯ t . Specifically:
  • X ¯ t = x ¯ m t m M are the association decision factors of MDs at timeslot t, where x ¯ m 0 , 1 .
  • Y ¯ t + 1 = y ¯ m t + 1 m M are the caching decision factors at timeslot t for the next timeslot.
  • Z ¯ t = z ¯ i t i I are the subchannel allocation decision factors of BSs at timeslot t, where z ¯ i 0 , 1 .
  • U ¯ t = u ¯ m t m M are the offloading decision factors of MDs at timeslot t, where u ¯ m 0 , 1 .
  • F ¯ t = f ¯ m t m M are the computing resource allocation factors of MDs at timeslot t, where f ¯ m 0 , 1 .
It is noteworthy that the dimensions of the above-mentioned state and action spaces have been greatly reduced compared to the actual ones. The actual state and action spaces can be achieved by executing an AT algorithm in the following parts.
Reward: Considering that the goal of problem P1 is to minimize local energy consumption and the constraints C 1 and C 10 cannot be strictly satisfied in the DRL-based iteration procedure, the reward w t at timeslot t is given by
w t = ω 1 m M ε m tot t ω 2 ϕ t ω 3 φ t ,
where ϕ ( t ) = m M max ( τ m tot ( t ) τ m , 0 ) is the penalty function added for guaranteeing the constraint C 1 ; φ ( t ) = i I max ( m M y m , i ( t ) d m ( t ) ϑ i ( t ) , 0 ) is the penalty function introduced for guaranteeing the constraint C 10 ; ω 1 is the energy-consumption discount factor; ω 2 and ω 3 are penalty coefficients.
When the network obtains action a t according to the state s t , the state space will obtain the next state s t + 1 according to the action a t . Specifically, the task-caching decisions of BSs can be directly achieved from Y t + 1 in a t . Therefore, the total return of minimizing long-term local energy consumption within T timeslots can be given by
R = t T γ w t ,
where γ is the reward discount factor satisfying γ 0 , 1 .

4.2. TD3O Algorithm

The TD3 algorithm is an actor-critic-based framework that comprises the policy ( μ ) network, critic (Q) network and their corresponding target networks and updates the network parameters using gradient algorithms. It is characterized by using two critic networks and two critic target networks in the design of critic networks. The TD3 algorithm is often divided into two parts consisting of experience collection and training. In the phase of collecting experience, a new action a t can be generated by adding random Gaussian noise into the output of the policy network at the state s t , i.e.,
a t = μ s t , θ μ + σ ¯ 2 .
where θ μ is the parameter of the policy network and σ ¯ 2 is the additive Gaussian noise.
After that, the environment is rewarded with w t and the next state s t + 1 can be achieved according to the state and action s t , a t . To enable the algorithm to obtain better decisions through past experience-assisted training, we put the quadruple s t , a t , w t , s t + 1 into the experience replay buffer as a historical experience. In the training process, a certain number of quadruples are randomly selected from the experience replay buffer for training. Since the TD3 algorithm consists of policy and critic networks, the training part of the network is relatively independent, so it is divided into the following two parts.

4.2.1. Training Policy Network

The training process of the policy network is shown in Figure 3. In the training phase, N quadruples are extracted from the experience replay buffer and denoted as N = 1 , 2 , , N . For any quadruple n N , the policy network outputs a new action a n = μ s n , θ μ according to the state s n . It should be noted that the policy a n is different from a n existing in the experience replay buffer. After s n and a n are inputted into any critic network (e.g., critic Q 1 network), such network outputs q n = Q 1 s n , μ s n , θ μ , θ Q 1 , where θ Q 1 is the parameter of the critic Q 1 network. After achieving all q n , their mathematical expectation is given by
J θ μ = E Q 1 S , μ S , θ μ , θ Q 1 ,
where S = s n | n N . Then, the policy gradient of function J with respect to θ μ can be given by
θ μ J = E A Q 1 S , A , θ Q 1 θ μ μ S , θ μ ,
where A = a n | n N .
Significantly, the calculated gradient requires gradient clipping, which can avoid skipping the optimal solution because the gradient is too large. The calculated policy gradients will be used to update the parameters of the policy networks. We assume that the learning rate of the policy network is β μ , and use the adaptive moment (Adam) estimation commonly used in DRL to obtain the optimal θ μ [43].

4.2.2. Training Critic Network

Figure 4 shows the training process of the critic network. During the critic network training, the policy at the next time is first estimated through the state at the next time by the policy target( μ ) network, i.e., a n = μ ( s n , θ μ ) + σ ^ 2 , where σ ^ 2 is policy noise, that is, trimmed additive Gaussian noise. Then, the action a n and the state s n are used as the input of the critic target ( Q 1 ) network and critic target ( Q 2 ) network, where θ Q 1 and θ Q 2 are their parameters. After that, these two networks output q ˜ n , 1 and q ˜ n , 2 , respectively. Next, the approximation of Q value is q ¯ n = r n + γ q ˜ n achieved using Behrman equation, where q ˜ n = min q ˜ n , 1 , q ˜ n , 2 . At the same time, the action a n and the state s n are used as the input of the critic Q 1 network and critic Q 2 network, where θ Q 1 and θ Q 2 are their parameters. After that, these two networks output q n , 1 and q n , 2 . At last, for all q ¯ n , according to the theorem of mean squared error (MSE), the expectation function of the squared loss between Q 1 S , A , θ Q 1 and Q ¯ is
L 1 θ Q 1 = 0.5 E Q 1 S , A , θ Q 1 Q ¯ 2 ,
and the expectation function of the squared loss between Q 2 S , A , θ Q 2 and Q ¯ is given by
L 2 θ Q 2 = 0.5 E Q 2 S , A , θ Q 2 Q ¯ 2 ,
where Q ¯ = q ¯ n | n N . Then, the gradient of the loss function L 1 θ Q 1 with respect to the parameter θ Q 1 is
θ Q 1 L 1 = E Q 1 S , A , θ Q 1 Q ¯ θ Q 1 Q 1 S , A , θ Q 1 ,
and the gradient of the loss function L 2 θ Q 2 with respect to the parameter θ Q 2 is given by
θ Q 2 L 2 = E Q 2 S , A , θ Q 2 Q ¯ θ Q 2 Q 2 S , A , θ Q 2 .
Similar to calculating the policy gradient, the gradient clipping needs to be performed after calculating the gradients using (19) and (20). In addition, β Q is the learning rate of the critic network, and the parameters of the two critic networks are updated using the Adam algorithm. Certainly, the parameters of critic target networks also need to be updated using the soft update method, i.e.,
θ μ = λ θ μ + 1 λ θ μ ,
θ Q 1 = λ θ Q 1 + 1 λ θ Q 1 ,
θ Q 2 = λ θ Q 2 + 1 λ θ Q 2 ,
where λ is the learning rate of target networks.
It is noteworthy that a lower network updating frequency is adopted in this paper. We assume that the update interval of the critic network is t cti and the update interval between the policy and critic networks is t pti . The critic networks are trained many times to ensure the stability of Q value. After that, the policy network can be updated. The detailed procedure of the TD3O algorithm is summarized in Algorithm 1, where t mep is the maximal number of epochs.
Algorithm 1: TD3-based offloading (TD3O)
1: Initialization: θ Q 1 , θ Q 2 , θ μ , θ Q 1 , θ Q 2 , θ μ , t step = 0 , t epoch = 0 .
2: While  t epoch < t mep
3:   Let t = 0 , state s t and reward R = 0 .
4:   While  t < T
5:    Generate action a t using (14).
6:    Achieve actual action by executing Algorithm 2.
7:    Calculate reward w t using (12) and obtain the state s t + 1 .
8:    If  t step κ
9:      Replace the previous quadruple with s t , a t , w t , s t + 1 .
10:     Else
11:       Put the quadruple s t , a t , w t , s t + 1 into the queue.
12:     EndIf
13:     Update state s t = s t + 1 .
14:     If  t step % t cti = 0 and t step > N
15:       Extract N quadruples for training.
16:       For any sample n, Q 1 and Q 2 networks output q ˜ n , 1 and
17:         q ˜ n , 2 , respectively, and obtain the minimum value q ˜ n .
18:       Calculate L ( θ 1 Q ) and L ( θ 2 Q ) using (17) and (18), respectively.
19:       Calculate Q gradient using (19) and (20), and clip it.
20:       Find θ Q 1 and θ Q 2 using Adam optimizer.
21:       If  t step % t pti = 0
22:        Calculate q through Q 1 .
23:        Calculate policy gradient using (16), and clip it.
24:        Find θ μ using Adam optimizer.
25:       EndIf
26:       Calculate θ Q 1 , θ Q 2 and θ μ using (21)–(23), respectively.
27:     EndIf
28:      R = R + γ w t .
29:      t step = t step + 1 ; t = t + 1 .
30:   EndWhile
31:      t epoch = t epoch + 1 .
32: EndWhile
Algorithm 2: Action transformation (AT)
1: For each MD m M
2:   Achieve MD association matrix X using discretization rule.
3:   Achieve task caching matrix Y using discretization rule.
4:   Achieve task offloading matrix U using discretization rule.
5: EndFor
6: For each BS i I
7:  Returns the set K i of available subchannels and the set M i of
8:   offloading MDs.
9:  If  M i > K i
10:      M i K i associated MDs are randomly selected, disassociated
11:      and execute tasks locally.
12:   EndIf
13:   Achieve subchannel allocation matrix Z using discretization rule.
14: EndFor
15: For each MD m M
16:   If  i I u m , i = 1
17:     If  f ¯ m = 0
18:      Assign small enough computing capacity to MD m to avoid
19:       zero division.
20:     Else
21:      Allocate computing resources to MD m using (28).
22:     EndIf
23:   EndIf
24: EndFor

4.3. AT Algorithm

In order to apply the TD3O algorithm to solve the problem P1, it is necessary to convert the achieved continuous action a t = X ¯ t , Y ¯ t + 1 , Z ¯ t , U ¯ t , F ¯ t into a discrete one [44]. To this end, we consider the following transformations for a t .

4.3.1. The Discretization of Device Association Array

In X ¯ = x ¯ m m M , x ¯ m is the non-integer association index of MD m, which is the continuous action achieved by the TD3 algorithm. Then, it is converted into an integer form, i.e.,
x m , ceil I x ¯ m = 1 , if I x ¯ m 0 , x m , 1 = 1 , otherwise ,
where ceil ( b ) is an upward rounding function with respect to b. Such a transformation can ensure that each MD can be associated with one BS.

4.3.2. The Discretization of Task-Caching Array

In Y ¯ , y ¯ m represents the non-integer caching index of MD m, which is the continuous action achieved by the TD3 algorithm. Since each MD can store its task at all BSs, there exist 2 I storage options for it. Consequently, in order to convert y ¯ m into a discrete form, we first need to perform
y ^ m = floor 2 I y ¯ m , if 2 I y ¯ m 0 , y ^ m = 0 , otherwise ,
where floor ( b ) is a downward rounding function with respect to b. Then, in order to achieve the binary caching index, the decimal y ^ m needs to be converted into a binary number of I 0–1 digits, which is given by bin y ¯ m . In it, bin ( b ) is a function used for calculating the binary number of decimal b. Then, y m , i = bin y ¯ m i , where bin y ¯ m i represents the i-th digit of the binary number bin y ¯ m .

4.3.3. The Discretization of Task Offloading Array

In U ¯ , u ¯ m is the non-integer offloading index of MD m, which is the continuous action achieved by the TD3 algorithm. Considering that each MD can offload its task to at most one BS, u ¯ m is converted into an integer form, i.e.,
u m , ceil I u ¯ m = 1 , if I u ¯ m 0 , u m , i = 0 , i I , otherwise .

4.3.4. The Discretization of Subchannel Allocation Array

In Z ¯ , z ¯ i is the non-integer index of the subchannels allocated by BS i to its associated MDs who need to offload tasks, which is the continuous action achieved by the TD3 algorithm. To achieve the integer form of z ¯ i , we first need to perform
z ^ i = ceil C M i , K i z ¯ i , if C M i , K i z ¯ i 0 , z ^ i = 1 , otherwise ,
where M i is the number of MDs that are associated with BS i and need to offload tasks; K i is the number of available subchannels at BS i; C M i , K i = fac ( K i ) / fac ( M i ) fac K i M i is a function with respect to M i and K i and is used for calculating the number of feasible subchannel allocation policies between M i MDs and K i subchannels at BS i; fac ( b ) is a factorial function with respect to b; M i K i shall be satisfied.
Then, we assume that Z i = 1 , 2 , , C M i , K i is the set of C M i , K i feasible subchannel allocation policies between M i MDs and K i subchannels at BS i. After that, the subchannel allocation policy z ^ i in the set Z i is selected according to the Equation (27). It is noteworthy that C M i , K i feasible subchannel allocation policies are generated in advance. That is to say, in the policy z ^ i , we can easily know the utilized indices of K i subchannels for M i MDs. According to these rules, we can easily find the subchannel allocation index Z .

4.3.5. The Transformation of Computing Resource Allocation Array

In F ¯ = f ¯ m m M , f ¯ m represents the computing resource score of MD m at the target BS that is executing its task. If i I u m , i = 1 is satisfied between MD m and BS i, according to the proportional allocation of computing resources, the computing resources allocated to MD m by BS i can be given by
f m , i = u m , i f i BS f ¯ m / j M u j , i f ¯ j .
Based on the above-mentioned operations, the output action a t = X ¯ , Y ¯ , Z ¯ , U ¯ , F ¯ of the TD3O algorithm can be effectively converted into an actual decision, which is summarized as Algorithm 2.

4.4. HA

To solve the problem P1 in a non-iterative manner, we design an effective heuristic algorithm, which is summarized in Algorithm 3. In such an algorithm, to reduce the uplink transmission time and energy consumption, some MDs are associated with the nearest BSs, and the BSs randomly cache the tasks of their associated MDs until the cache space cannot cache more tasks. Then, the uncached MDs randomly select a BS as the offloading target, and to guarantee time constraints, the BS will evenly distribute the computing resources according to the computation amount of the task. Finally, a part of the MDs are disassociated from BSs without sufficient subchannels and execute tasks by themselves.
Algorithm 3: Heuristic algorithm (HA)
1: Initialization: energy consumption ε ¯ tot = 0 .
2: Each MD selects (is associated with) the nearest BS.
3: For each BS i I
4:    If  M i > K i
5:      M i K i associated MDs are randomly selected, disassociated
6:      and execute tasks locally.
7:    EndIf
8:    Randomly select the tasks of MDs associated with BS i for caching
9:     until the caching space is full.
10: EndFor
11: For  t T
12:    Randomly select a target BS for each MD without cached task.
13:    Randomly allocate subchannels to MDs associated with each BS.
14:    If subchannels are insufficient
15:     Extra MDs are randomly selected to execute tasks locally.
16:    EndIf
17:    Proportionally allocate computing resources to MDs associated with
18:     each BS according to the CPU cycles required by tasks.
19:    Calculate the total local energy consumption ε ¯ .
20:     ε ¯ tot = ε ¯ tot + ε ¯ .
21: EndFor

5. Algorithm Analysis

5.1. Computation Complexity Analysis

In this section, the computation complexity of proposed algorithms are analysed as follows.
Proposition 1.
The computation complexity of Algorithm 2 is O M I K in the worst case.
Proof. 
In Algorithm 2, the computation complexity of steps 1–5 is O M , the computation complexity of steps 6–14 is O M I K in the worst case, and the computation complexity of steps 15–24 is O M I . In general, the computation complexity of Algorithm 2 is O M I K in the worst case. □
Proposition 2.
The computation complexity of Algorithm 1 is O ( max ( l = 0 L Q ψ l Q ψ l + 1 Q , l = 0 L μ ψ l μ ψ l + 1 μ ) ) at each timeslot, where L μ is the number of layers of the policy network, L Q is the number of layers of the critic network, ψ l μ is the number of neurons at the l-th layer of the policy network and ψ l Q is the number of neurons at l-th layer of the critic network.
Proof. 
In Algorithm 1, the computation complexity is mainly related to the action transformation, the calculation of the reward and task processing time and the structure of the neural network. As previously mentioned, the computation complexity of the action transformation should be O M I K in the worst case. As seen from Formulas (8) and (12), the computation complexity of the calculation of the reward and task processing time is O M I K .
In Algorithm 1, there exist four critic networks and two policy networks. We assume that the structure of the policy network and its target network is the same, and the structure of the two critic networks and its target network is the same. Then, we can easily deduce that the computation complexity of establishing policy networks is O l = 0 L μ ψ l μ ψ l + 1 μ and the computation complexity of establishing critic networks is O l = 0 L Q ψ l Q ψ l + 1 Q . Therefore, the computation complexity of establishing neural networks is O ( max ( l = 0 L Q ψ l Q ψ l + 1 Q , l = 0 L μ ψ l μ ψ l + 1 μ ) ) .
Since the computation complexity of establishing neural networks is much higher than that of the other operations in Algorithm 2. In general, the computation complexity of Algorithm 2 is O max l = 0 L Q ψ l Q ψ l + 1 Q , l = 0 L μ ψ l μ ψ l + 1 μ at each timeslot. □
Proposition 3.
The computation complexity of Algorithm 3 is O M I at each timeslot.
Proof. 
In Algorithm 3, the computation complexity of step 2 is O M I , the computation complexity of steps 3–10 is O I , the computation complexity of steps 12–16 is O M , the computation complexity of steps 17–19 is O M I . In general, the computation complexity of Algorithm 3 is O M I at each timeslot. □

5.2. Convergence Analysis

Since Algorithm 2 is a part of Algorithms 1 and 3 and is non-iterative, we just need to concentrate on the convergence of Algorithm 1. In detail, this is established as follows.
Theorem 1.
Algorithm 1 can be guaranteed to converge after finite iterations.
Proof. 
In Algorithm 1, the neural networks are updated by the gradient descent method used in the Adam optimizer. This utilizes the gradient information of the functions J ( θ μ ) , L 1 ( θ Q 1 ) and L 2 ( θ Q 2 ) to guide the updating directions of the parameters θ μ , θ Q 1 and θ Q 2 , so that the objective functions can reach the optimal or suboptimal values. When these values tend to be stable, the parameters θ μ , θ Q 1 and θ Q 2 also tend to be stable. At this time, Algorithm 1 is deemed convergent. □

6. Performance Evaluation

In order to verify the performance of the designed algorithms, we introduce the following algorithms for comparison.
DDPG-based offloading (DDPGO): DDPG is a classical DRL algorithm [45]. Compared with the TD3 algorithm, the DDPG algorithm reduces the critic network and the critic target network. In addition, both the critic network and policy network are updated at each timeslot in the DDPG algorithm. In this paper, the DDPG algorithm used to solve the problem P1 is named the DDPG-based offloading (DDPGO) algorithm. The difference between the two algorithms is that the state input and action output are the same as the mode used by TD3O, and it also uses the AT algorithm to convert continuous actions.
Completely offloading (CO): In the CO algorithm, the task of each MD is offloaded to the nearest BS for computing. Such BS proportionally allocates the computing capacity to its associated MDs according to the CPU cycles required by the tasks of these MDs.
Completely local executing (CLE): In the CLE algorithm, the tasks of all MDs can be executed by themselves.
In this paper, we consider that each BS is deployed in a non-overlapping area with a radius of 400 m and a power spectral density of −174 dBm/Hz. In addition, I = 3 , f m loc = 1 GHz , f i = 8 GHz , W = 40 MHz , K = 4 , d m = 2 5 MB , c m = 50 cycles / bit , ξ = 10 27 , τ m = 10 s , ρ = 2 , r bh = 1 Gbps , p m = 23 dBm , κ = 80,000 , N = 128 , γ = 0.94 and λ = 0.04 . In the DRL algorithm, we consider that both the policy network and the critic network are composed of three-layer fully connected neural networks, where the numbers of neurons in three-layer neural networks in the policy network are 300, 200 and 128, respectively, and the corresponding target network has the same structure with this policy network; the number of neurons in three-layer neural networks in the critic network are 300, 128 and 32, respectively, and the corresponding target network has the same structure as this critic network. Significantly, the first-layer fully connected neural network of the policy network and the critic network utilizes the rectified linear unit 6 (RELU6), which suppresses the maximum value as the activation function, while other layers use RELU as the activation function.
Figure 5 shows the convergence of the TD3O and DDPGO algorithms. As shown in Figure 5, DDPGO may have a higher convergence rate than TD3O, but the former may have worse convergence stability than the latter. The reason for this may be that the critic network and the policy network are updated synchronously in DDPGO. In DDPGO, the network parameters are updated in each training phase, which speeds up the convergence. Synchronously, the policy network parameters are updated in the training, which results in the instability of the long-term reward value and training bias. As we know, TD3O is composed of two sets of critic networks. Consequently, it could be trained in a relatively stable Q value so that the algorithm can converge stably. In the simulation, it is also easy to find that TD3O could achieve a more stable and better solution to the problem P1 than DDPGO in general.
Figure 6 shows the impact of the training interval t pti on the convergence of the TD3O algorithm. As we know, under the same number of iterations, a larger t pti can effectively reduce the overall training time of the network. However, it will reduce the total learning times of the policy network and its target network. As illustrated in Figure 6, the convergence rate of TD3O may decrease with t pti in general.
Figure 7 shows the impacts of learning rates β Q and β μ on the convergence of the TD3O algorithm. As we know, when the learning rate β Q of the critic network increases, the parameters of such network will be updated at a larger scale, which speeds up the convergence of TD3O. However, it may lead to the failure of stable evaluation of environmental information, which weakens the convergence stability of TD3O. As illustrated in Figure 7, when β Q = 0.001 , the convergence rate of TD3O is relatively high, but the achieved long-term reward dramatically fluctuates at this moment. On the other hand, the learning rate β μ of the policy network can affect the optimization capability of TD3O. Specifically, a lower β μ means a smaller amplitude of updating the policy network, which is better for finding better solutions. As seen from Figure 7, TD3O can achieve a better long-term reward when β Q = 0.0001 and β μ = 0.0001 .
Figure 8 shows the impact of the number of MDs on the long-term local energy consumption ε MD , where ε MD is the sum of the total local energy consumption in T timeslots. In general, the ε MD increases with the number of MDs since a greater energy consumption is used when tackling more tasks of more MDs. Since CLE executes tasks in maximal computation capacity, it could achieve the highest ε MD among all algorithms. In CO, MDs are associated with the nearest BSs, which may result in a relatively imbalanced load distribution. Then, some overloaded BSs cannot provide good services for their associated MDs because of limited resources, which may result in high ε MD . Consequently, CO could achieve higher ε MD than other algorithms excluding CLE. As illustrated in Figure 8, TD3O could achieve lower ε MD than DDPGO since the former can effectively mitigate the overestimation existing in the latter. Although HA lets MDs be associated with the nearest BSs, some MDs associated with overloaded BSs will disassociate and execute tasks locally. Such an operation may result in relatively low ε MD .
Figure 9 shows the impacts of the number of MDs on the long-term reward (R). As illustrated in Figure 9, R may decrease with the number of MDs since more MDs result in a higher energy consumption. Since both TD3O and DDPGO try to maximize the reward but in other algorithms this is not the case, the former could achieve higher R than other algorithms in general. Since TD3O could achieve lower ε MD than DDPGO, the former could achieve a higher R than the latter. In view of the unstable convergence of DDPGO, its reward may dramatically fluctuate. Since CLE could achieve the highest ε MD among all algorithms, it could achieve the lowest R in general. In addition, CO could achieve a lower R than HA since the former consumes more energy than the latter.
Figure 10 shows the impacts of the size of caching space of each BS on the relative long-term energy consumption η ε denoted as the ratio of ε MD achieved by an offloading algorithm to the one attained by CLE. Evidently, a smaller η ε means a higher energy gain caused by offloading tasks. As illustrated in Figure 10, in addition to DDPGO and CO, η ε in other algorithms decreases with the size of the caching space of each BS in general. The reason for this may be that a larger caching space can hold more tasks to reduce the transmission energy consumption. However, as revealed in Figure 5 and Figure 9, the unstable convergence of DDPGO may result in a dramatically fluctuating performance. Therefore, η ε in DDPGO may evidently be fluctuating. In addition, η ε in CO may not change with the size of caching space of each BS since it does not utilize caching space. By minimizing the ε MD , TD3O and DDPGO could achieve a lower η ε than other algorithms in general. In addition, TD3O could achieve a lower η ε than DDPGO since the former mitigates the overestimation existing in DDPGO. As seen from Figure 10, CO could achieve the highest η ε among all algorithms since it has no sufficient resources to provide for MDs associated with overloaded BSs.
Figure 11 shows the impacts of the size of the caching space of each BS on the relative long-term reward η R denoted as the ratio of the long-term reward achieved by an offloading algorithm to the one attained by CLE. Evidently, a smaller η R means a higher reward gain caused by offloading tasks. As illustrated in Figure 11, η R in TD3O and HA decreases with the size of the caching space of each BS in general. The reason for this may be that a larger caching space can hold more tasks to reduce the transmission energy consumption, and then bring a higher reward. However, due to the unstable convergence of DDPGO, η R in DDPGO may be fluctuating. Moreover, η R in CO may not change with the size of caching space of each BS since it does not utilize caching space. By minimizing the ε MD and thus increasing the reward, TD3O and DDPGO could achieve a lower η R than other algorithms in general. In addition, TD3O could achieve a lower η R than DDPGO since the former mitigates the overestimation existing in DDPGO. As can be seen from Figure 11, CO could achieve the highest η R because of the high energy consumed by MDs associated with overloaded BSs.
As seen from the above-mentioned simulation figures, although HA is a non-iterative algorithm, it could sometimes achieve better performance than DDPGO. In addition, it may always achieve fairly better performance than CO and CLE.

7. Conclusions

In this paper, the problem of minimizing the local energy consumption is concentrated in the cache-assisted vehicular NOMA–MEC networks under time and resource constraints, which refers to the joint optimization of the computing resource allocation, subchannel selection, device association, offloading and caching decisions. To solve the formulated problem, we developed an effective TD3O algorithm that was integrated with the AT algorithm and designed HA simultaneously. As for the designed algorithms, we have given some analyses of the convergence and computation complexity. Simulation results show that TD3O could achieve lower local energy consumption than several benchmark algorithms, and HA could achieve lower local energy consumption than the CO and CLE algorithms. Future work can include power allocation and secure communications, such as optimizing the transmission power of MDs for task offloading, and how to encrypt part of the task data in the network at a low cost to achieve secure communication, etc.

Author Contributions

Conceptualization, T.Z. and M.X.; methodology, T.Z. and M.X.; software, M.X.; investigation, X.N., X.L. and C.L.; writing—original draft preparation, T.Z. and M.X.; writing—review and editing, D.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under grant Nos. 62261020, 62171119, 62062034, 62361026, 61961020, 62001201 and 61963017; the National Key Research and Development Program of China under grant No. 2020YFB1807201; the Jiangxi Provincial Natural Science Foundation under grant Nos. 20232ACB212005, 20224BAB202001, 20232BAB202019, 20212BAB202004 and 20212BAB212001; the key research and development plan of Jiangsu Province under grant No. BE2021013-3.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yao, Y.; Shu, F.; Li, Z.; Cheng, X.; Wu, L. Secure transmission scheme based on joint radar and communication in mobile vehicular networks. IEEE Trans. Intell. Transport. Syst. 2023, 24, 10027–10037. [Google Scholar] [CrossRef]
  2. Yao, Y.; Zhao, J.; Li, Z.; Cheng, X.; Wu, L. Jamming and eavesdropping defense scheme based on deep reinforcement learning in autonomous vehicle networks. IEEE Trans. Inf. Forens. Secur. 2023, 18, 1211–1224. [Google Scholar] [CrossRef]
  3. Wu, Q.; Shi, S.; Wan, Z.; Fan, Q.; Fan, P.; Zhang, C. Towards V2I age-aware fairness access: A dqn based intelligent vehicular node training and test method. Chin. J. Electron. 2023, 32, 1230–1244. [Google Scholar]
  4. Khan, S.; Luo, F.; Zhang, Z.; Ullah, F.; Amin, F.; Qadri, S.; Heyat, M.; Ruby, R.; Wang, L.; Ullah, S.; et al. A survey on X.509 public-key infrastructure, certificate revocation, and their modern implementation on blockchain and ledger technologies. IEEE Commun. Surv. Tutor. 2023; early access. [Google Scholar] [CrossRef]
  5. Khan, S.; Luo, F.; Zhang, Z.; Rahim, M.; Ahmad, M.; Wu, K. Survey on issues and recent advances in vehicular public-key infrastructure (VPKI). IEEE Commun. Surv. Tutor. 2023, 24, 1574–1601. [Google Scholar] [CrossRef]
  6. Wang, D.; Tian, X.; Cui, H.; Liu, Z. Reinforcement learning-based joint task offloading and migration schemes optimization in mobility-aware MEC network. China Commun. 2020, 17, 31–44. [Google Scholar] [CrossRef]
  7. Spinelli, F.; Mancuso, V. Toward enabled industrial verticals in 5G: A survey on MEC-based approaches to provisioning and flexibility. IEEE Commun. Surv. Tutor. 2021, 23, 596–630. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Di, B.; Zheng, Z.; Lin, J.; Song, L. Distributed multi-cloud multi-access edge computing by multi-agent reinforcement learning. IEEE Trans. Wirel. Commun. 2021, 20, 2565–2578. [Google Scholar] [CrossRef]
  9. Zhou, T.; Yue, Y.; Qin, D.; Nie, X.; Li, X.; Li, C. Joint device association, resource allocation, and computation offloading in ultradense multidevice and multitask IoT networks. IEEE Internet Things J. 2022, 9, 18695–18709. [Google Scholar] [CrossRef]
  10. Zhang, W.; Zhang, G.; Mao, S. Joint parallel offloading and load balancing for cooperative-MEC systems with delay constraints. IEEE Trans. Veh. Technol. 2022, 71, 4249–4263. [Google Scholar] [CrossRef]
  11. Malik, R.; Vu, M. Energy-efficient joint wireless charging and computation offloading in MEC systems. IEEE J. Sel. Top. Signal Proces. 2021, 15, 1110–1126. [Google Scholar] [CrossRef]
  12. Hu, H.; Song, W.; Wang, Q.; Hu, R.Q.; Zhu, H. Energy efficiency and delay tradeoff in an MEC-enabled mobile IoT network. IEEE Internet Things J. 2022, 9, 15942–15956. [Google Scholar] [CrossRef]
  13. Wu, Q.; Wang, S.; Ge, H.; Fan, P.; Fan, Q.; Letaief, K. Delay-sensitive task offloading in vehicular fog computing-assisted platoons. IEEE Trans. Netw. Service Manag. 2023; early access. [Google Scholar] [CrossRef]
  14. Bai, Y.; Zhao, H.; Zhang, X.; Chang, Z.; Jäntti, R.; Yang, K. Towards autonomous multiuav wireless network: A survey of reinforcement learning-based approaches. IEEE Commun. Surv. Tutor. 2023; early access. [Google Scholar] [CrossRef]
  15. Liu, Y.; Liu, J.; Argyriou, A.; Wang, L.; Xu, Z. Rendering-aware VR video caching over multi-cell MEC networks. IEEE Trans. Veh. Technol. 2021, 70, 2728–2742. [Google Scholar] [CrossRef]
  16. Lekharu, A.; Jain, M.; Sur, A.; Sarkar, A. Deep learning model for content aware caching at MEC servers. IEEE Trans. Netw. Service Manag. 2022, 19, 1413–1425. [Google Scholar] [CrossRef]
  17. Huang, X.; He, L.; Wang, L.; Li, F. Towards 5G: Joint optimization of video segment caching, transcoding and resource allocation for adaptive video streaming in a multi-access edge computing network. IEEE Trans. Veh. Technol. 2021, 70, 10909–10924. [Google Scholar] [CrossRef]
  18. Zheng, G.; Xu, C.; Wen, M.; Zhao, X. Service caching based aerial cooperative computing and resource allocation in multi-UAV enabled MEC systems. IEEE Trans. Veh. Technol. 2022, 71, 10934–10947. [Google Scholar] [CrossRef]
  19. Wu, Q.; Wang, X.; Fan, Q.; Fan, P.; Zhang, C.; Li, Z. High stable and accurate vehicle selection scheme based on federated edge learning in vehicular networks. China Commun. 2023, 20, 1–17. [Google Scholar] [CrossRef]
  20. Liu, B.; Liu, C.; Peng, M. Resource allocation for energy-efficient MEC in NOMA-enabled massive IoT networks. IEEE J. Sel. Areas Commun. 2021, 39, 1015–1027. [Google Scholar] [CrossRef]
  21. Song, Z.; Liu, Y.; Sun, X. Joint task offloading and resource allocation for NOMA-enabled multi-access mobile edge computing. IEEE Trans. Commun. 2021, 69, 1548–1564. [Google Scholar] [CrossRef]
  22. Ding, Z.; Xu, D.; Schober, R.; Poor, H.V. Hybrid NOMA offloading in multi-user MEC networks. IEEE Trans. Wirel. Commun. 2022, 21, 5377–5391. [Google Scholar] [CrossRef]
  23. Farha, Y.A.; Ismail, M.H. Design and optimization of a UAV-enabled non-orthogonal multiple access edge computing IoT system. IEEE Access 2022, 10, 117385–117398. [Google Scholar] [CrossRef]
  24. Wang, K.; Li, H.; Ding, Z.; Xiao, P. Reinforcement learning based latency minimization in secure NOMA-MEC systems with hybrid SIC. IEEE Trans. Wirel. Commun. 2023, 22, 408–422. [Google Scholar] [CrossRef]
  25. Xu, C.; Zheng, G.; Zhao, X. Energy-minimization task offloading and resource allocation for mobile edge computing in NOMA heterogeneous networks. IEEE Trans. Veh. Technol. 2020, 69, 16001–16016. [Google Scholar] [CrossRef]
  26. Ylmaz, S.S.; Zbek, B. Massive MIMO-NOMA based MEC in task offloading for delay minimization. IEEE Access 2023, 11, 162–170. [Google Scholar] [CrossRef]
  27. Tuong, V.D.; Truong, T.P.; Nguyen, T.-V.; Noh, W.; Cho, S. Partial computation offloading in NOMA-assisted mobile-edge computing systems using deep reinforcement learning. IEEE Internet Things J. 2021, 8, 13196–13208. [Google Scholar] [CrossRef]
  28. Feng, H.; Guo, S.; Yang, L.; Yang, Y. Collaborative data caching and computation offloading for multi-service mobile edge computing. IEEE Trans. Veh. Technol. 2021, 70, 9408–9422. [Google Scholar] [CrossRef]
  29. Ren, D.; Gui, X.; Zhang, K. Adaptive request scheduling and service caching for MEC-assisted IoT networks: An online learning approach. IEEE Internet Things J. 2022, 9, 17372–17386. [Google Scholar] [CrossRef]
  30. Yang, S.; Liu, J.; Zhang, F.; Li, F.; Chen, X.; Fu, X. Caching-enabled computation offloading in multi-region MEC network via deep reinforcement learning. IEEE Internet Things J. 2022, 9, 21086–21098. [Google Scholar] [CrossRef]
  31. Zhou, T.; Yue, Y.; Qin, D.; Nie, X.; Li, X.; Li, C. Mobile device association and resource allocation in HCNs with mobile edge computing and caching. IEEE Syst. J. 2023, 17, 976–987. [Google Scholar] [CrossRef]
  32. Wu, Q.; Zhao, Y.; Fan, Q.; Fan, P.; Wang, J.; Zhang, C. Mobility-aware cooperative caching in vehicular edge computing based on asynchronous federated and deep reinforcement learning. IEEE J. Sel. Top. Signal Process. 2023, 17, 66–81. [Google Scholar] [CrossRef]
  33. Zhou, H.; Wang, Z.; Zheng, H.; He, S.; Dong, M. Cost minimization-oriented computation offloading and service caching in mobile cloud-edge computing: An A3C-based approach. IEEE Trans. Netw. Sci. Eng. 2023, 10, 1326–1338. [Google Scholar] [CrossRef]
  34. Zhang, W.; Zeadally, S.; Zhou, H.; Zhang, H.; Wang, N.; Leung, V. Joint service quality control and resource allocation for service reliability maximization in edge computing. IEEE Trans. Commun. 2023, 71, 935–948. [Google Scholar] [CrossRef]
  35. Li, X.; Zhang, H.; Zhou, H.; Wang, N.; Long, K.; Al-Rubaye, S.; Karagiannidis, G. Multi-agent drl for resource allocation and cache design in terrestrial-satellite networks. IEEE Trans. Wirel. Commun. 2023, 22, 5031–5042. [Google Scholar] [CrossRef]
  36. Yang, Z.; Liu, Y.; Chen, Y.; Al-Dhahir, N. Cache-aided NOMA mobile edge computing: A reinforcement learning approach. IEEE Trans. Wirel. Commun. 2020, 19, 6899–6915. [Google Scholar] [CrossRef]
  37. Li, S.; Li, B.; Zhao, W. Joint optimization of caching and computation in multi-server NOMA-MEC system via reinforcement learning. IEEE Access 2020, 8, 112762–112771. [Google Scholar] [CrossRef]
  38. Huynh, L.N.T.; Pham, Q.-V.; Nguyen, T.D.T.; Hossain, M.D.; Shin, Y.-R.; Huh, E.-N. Joint Computational offloading and data-content caching in NOMA-MEC networks. IEEE Access 2021, 9, 12943–12954. [Google Scholar] [CrossRef]
  39. Yang, Z.; Ding, Z.; Fan, P.; Al-Dhahir, N. A general power allocation scheme to guarantee quality of service in downlink and uplink NOMA systems. IEEE Trans. Wirel. Commun. 2016, 15, 7244–7257. [Google Scholar] [CrossRef]
  40. Cheng, Y.; Liang, C.; Chen, Q.; Yu, F.R. Energy-efficient D2D-assisted computation offloading in NOMA-enabled cognitive networks. IEEE Trans. Veh. Technol. 2020, 70, 13441–13446. [Google Scholar] [CrossRef]
  41. Li, C.; Wang, H.; Song, R. Mobility-aware offloading and resource allocation in NOMA-MEC systems via DC. IEEE Commun. Lett. 2022, 26, 1091–1095. [Google Scholar] [CrossRef]
  42. Huang, H.; Ye, Q.; Zhou, Y. 6G-empowered offloading for realtime applications in multi-access edge computing. IEEE Trans. Netw. Sci. Eng. 2023, 10, 1311–1325. [Google Scholar] [CrossRef]
  43. Chen, J.; Xing, H.; Xiao, Z.; Xu, L.; Tao, T. A DRL agent for jointly optimizing computation offloading and resource allocation in MEC. IEEE Internet Things J. 2021, 8, 17508–17524. [Google Scholar] [CrossRef]
  44. Dai, Y.; Zhang, K.; Maharjan, S.; Zhang, Y. Edge intelligence for energy-efficient computation offloading and resource allocation in 5G beyond. IEEE Trans. Veh. Technol. 2020, 69, 12175–12186. [Google Scholar] [CrossRef]
  45. Long, D.; Wu, Q.; Fan, Q.; Fan, P.; Li, Z.; Fan, J. A power allocation scheme for MIMO-NOMA and D2D vehicular edge computing based on decentralized DRL. Sensors 2023, 23, 3449. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Cache-assisted vehicular NOMA–MEC networks.
Figure 1. Cache-assisted vehicular NOMA–MEC networks.
Sensors 23 09064 g001
Figure 2. Caching and offloading models.
Figure 2. Caching and offloading models.
Sensors 23 09064 g002
Figure 3. Training policy network.
Figure 3. Training policy network.
Sensors 23 09064 g003
Figure 4. Training critic network.
Figure 4. Training critic network.
Sensors 23 09064 g004
Figure 5. The convergence of DDPGO and TD3O algorithms.
Figure 5. The convergence of DDPGO and TD3O algorithms.
Sensors 23 09064 g005
Figure 6. The impact of training interval t pti on the convergence of TD3O algorithm.
Figure 6. The impact of training interval t pti on the convergence of TD3O algorithm.
Sensors 23 09064 g006
Figure 7. The impacts of learning rates β Q and β μ on the convergence of TD3O algorithm.
Figure 7. The impacts of learning rates β Q and β μ on the convergence of TD3O algorithm.
Sensors 23 09064 g007
Figure 8. The impacts of the number of MDs on the long-term local energy consumption ε MD .
Figure 8. The impacts of the number of MDs on the long-term local energy consumption ε MD .
Sensors 23 09064 g008
Figure 9. The impacts of the number of MDs on the long-term reward R.
Figure 9. The impacts of the number of MDs on the long-term reward R.
Sensors 23 09064 g009
Figure 10. The impacts of the size of caching space of each BS on the relative long-term energy consumption η ε .
Figure 10. The impacts of the size of caching space of each BS on the relative long-term energy consumption η ε .
Sensors 23 09064 g010
Figure 11. The impacts of the size of caching space of each BS on the relative long-term reward η R .
Figure 11. The impacts of the size of caching space of each BS on the relative long-term reward η R .
Sensors 23 09064 g011
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, T.; Xu, M.; Qin, D.; Nie, X.; Li, X.; Li, C. Computing Offloading Based on TD3 Algorithm in Cache-Assisted Vehicular NOMA–MEC Networks. Sensors 2023, 23, 9064. https://doi.org/10.3390/s23229064

AMA Style

Zhou T, Xu M, Qin D, Nie X, Li X, Li C. Computing Offloading Based on TD3 Algorithm in Cache-Assisted Vehicular NOMA–MEC Networks. Sensors. 2023; 23(22):9064. https://doi.org/10.3390/s23229064

Chicago/Turabian Style

Zhou, Tianqing, Ming Xu, Dong Qin, Xuefang Nie, Xuan Li, and Chunguo Li. 2023. "Computing Offloading Based on TD3 Algorithm in Cache-Assisted Vehicular NOMA–MEC Networks" Sensors 23, no. 22: 9064. https://doi.org/10.3390/s23229064

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop