MADDPG-Based Offloading Strategy for Timing-Dependent Tasks in Edge Computing
Abstract
:1. Introduction
- Aiming at the problem of long offloading delay due to concurrent application of services by multiple users and timing dependency constraints among tasks, this paper analyzes timing dependencies based on the depth-first traversal (DFS) algorithm and integrates gated recursive units (GRUs) into a multi-agent reinforcement learning algorithm to memorize effective information about historical offloading decisions and task attributes and predicts computational resources required for subsequent tasks. This method guarantees the supply of computational resources for user demand and reduces the number of iterations of the multi-agent algorithm learning, thus reducing the offloading delay and improving the timeliness of the offloading strategy.
- Aiming at the problem of low resource utilization due to resource conflict caused by the concurrent offloading of tasks by multiple users, an edge-computing, time-dependent task offloading algorithm based on user behaviour prediction is proposed. In a complex environment with large differences in resource conflicts and task time-dependent constraints, the training process of the MADDPG algorithm is improved to avoid a situation in which the local decision-making of single-agent algorithms is difficult to comprehensively optimize the global benefits, reduce the offloading delay, and improve the global resource utilization.
- Simulation results show that the proposed algorithm performs better than the DQN and DDQN single-agent algorithms in terms of both delay reduction and resource utilization improvement. In addition, the algorithm reduces the response latency by 6.7% and improves the resource utilization by 30.6% compared to the suboptimal offloading algorithm based only on MADDPG and reduces nearly 500 training rounds during the learning process, which effectively improves the timeliness of the offloading strategy.
2. Related Works
2.1. Timing-Dependent Task Offloading in MEC
2.2. Edge-Computing Task Offloading Based on Resource Prediction
3. System Model
3.1. Time-Dependent Task Model
3.2. Task-Offloading Delay Evaluation Model
3.3. Computing Resource Forecasting Model
3.4. Problem Formulation
4. User-Prediction-Based Timing-Dependent Task Offloading Algorithm
4.1. Timing-Dependent Task Preprocessing
4.2. Timing-Dependent Task Offloading Algorithm
Algorithm 1: MADDPG-based timing-dependent task offloading algorithm | |
Input: Task serial and parallel scheduling list SL. | |
Output: Independent offloading strategy πd for each agent d. | |
1: | Initialize the environment, including the resources, quantity, and location of MEC nodes and users |
2: | Agent d observes the state of the environment ot and selects the action at |
3: | Obtain the environment state feedback reward rt |
4: | for episode = 2 to max_episode do |
5: | Reset the MEC network |
6: | Get the task scheduling sequence |
7: | while the episode do not end do |
8: | for agent d in the set of agents do |
9: | Observe the ot+1, rt and select the action at+1 = μd(ot+1, rt; θd) |
10: | end for |
11: | Executing at+1:(ot+1, r) = Actions[ad] in parallel |
12: | for agent d in the set of agents do |
13: | Record the experience (, , ) of agent d and put it into the experience replay pool |
14: | end for |
15: | for agent d in the set of agents do |
16: | Update policy network (agent d) |
17: | Update value network (agent d) |
18: | end for |
19: | θ’d←τθd + (1–τ)θ’d |
20: | end while |
21: | end for |
22: | Each agent autonomously decides the task offloading mode and generates the offloading policy πd |
23: | Return π |
5. Experimental Results and Discussion
5.1. Parameter Settings
- (1)
- Remote offloading algorithm: all tasks in the DAG are offloaded to the MEC server on the ground for execution.
- (2)
- Greedy algorithm: each task in the DAG is greedily assigned to either local or remote execution based on the algorithm’s estimation of the local processor’s or the MEC node’s completion time.
- (3)
- Deep Q-network-based dependent task offloading algorithm (DQN_DTO): using deep Q-networks, the agent automatically learns how to choose the optimal task offloading decision under different environment states by learning the reward signals in the environment.
- (4)
- DDQN-based dependent task-offloading algorithm (DDQN _DTO): agents are trained by alternating between these two neural networks to dynamically adjust and optimize the task offloading strategy to achieve a more accurate benefit trade-off.
- (5)
- MADDPG-based dependent task-offloading algorithm (MADDPG_DTO): By continuously training the MADDPG algorithm, multiple agents collaborate with each other to jointly weigh the offloading benefits and generate offloading strategies.
5.2. Parameter Settings
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Appendix A
Symbol | Description |
---|---|
Gd | the application program of the dth terminal device |
Vd | the set of subtasks in the graph |
Ei | the dependency between tasks |
ai | the size of the input data of task i |
ci | the number of CPU cycles required to compute each bit of data for the task |
the maximum tolerable delay for computing the task | |
fl | the UE local computation rate |
fn | the calculation rate of the MEC server |
, | the communication channel bandwidth of the uplink |
the communication channel bandwidth of the downlink | |
Pd,n | the uplink/downlink channel transmission power |
η2 | the channel noise during the transmission process |
the uplink data transmission rate | |
the downlink data transmission rate | |
the uplink data transmission delay | |
the downlink data transmission delay | |
the calculation delay of processing tasks on the MEC server | |
xt | the input vector |
ht | the state memory variable |
ht−1 | the state memory variable at the previous moment |
the state of the current candidate set | |
σ | the sigmoid activation function |
W | the weight parameter for the multiplication of the connection matrix |
F, Q, Qi | the set of all serial tasks, the set of all parallel tasks, the ith subset of the parallel set |
Ti,j | the ith subtask latency in the jth parallel set |
ti | the execution latency of the ith serial task |
Ttotal | the total offload delay |
Utotal | the resource utilization |
w1, w2 | the weights in the definition of the reward function |
the reciprocal of the processing delay of the jth task of the dth application | |
the resource utilization when offloading task j to node k | |
the maximum available computing resource on node k | |
the computing resource required for task j |
References
- Aikyn, N.; Zhanegizov, A.; Aidarov, T.; Bui, D.; Tu, N.A. Efficient facial expression recognition framework based on edge computing. J. Supercomput. 2024, 80, 1935–1972. [Google Scholar] [CrossRef]
- Gong, T.; Zhu, L.; Yu, F.R.; Tang, T. Train-to-Edge Cooperative Intelligence for Obstacle Intrusion Detection in Rail Transit. IEEE Trans. Veh. Technol. 2024, 1–13. [Google Scholar] [CrossRef]
- Wang, M.; Zhang, Y.; He, X.; Yu, S. Joint scheduling and offloading of computational tasks with time dependency under edge computing networks. Simul. Model. Pract. Theory 2023, 129, 102824. [Google Scholar] [CrossRef]
- Duan, S.; Wang, D.; Ren, J.; Lyu, F.; Zhang, Y.; Wu, H.; Shen, X. Distributed artificial intelligence empowered by end-edge-cloud computing: A survey. IEEE Commun. Surv. Tutor. 2022, 25, 591–624. [Google Scholar] [CrossRef]
- Mao, Y.; You, C.; Zhang, J.; Huang, K.; Letaief, K.B. A survey on mobile edge computing: The communication perspective. IEEE Commun. Surv. Tutor. 2017, 19, 2322–2358. [Google Scholar] [CrossRef]
- Yang, J.; Yuan, Q.; Chen, S.; He, H.; Jiang, X.; Tan, X. Cooperative task offloading for mobile edge computing based on multi-agent deep reinforcement learning. IEEE Trans. Netw. Serv. Manag. 2023, 20, 3205–3219. [Google Scholar] [CrossRef]
- An, X.; Fan, R.; Hu, H.; Zhang, N.; Atapattu, S.; Tsiftsis, T.A. Joint task offloading and resource allocation for IoT edge computing with sequential task dependency. IEEE Internet Things J. 2022, 9, 16546–16561. [Google Scholar] [CrossRef]
- Wang, J.; Hu, J.; Min, G.; Zomaya, A.Y.; Georgalas, N. Fast adaptive task offloading in edge computing based on meta reinforcement learning. IEEE Trans. Parallel Distrib. Syst. 2020, 32, 242–253. [Google Scholar] [CrossRef]
- Gao, S.; Wang, Y.; Feng, N.; Wei, Z.; Zhao, J. Deep reinforcement learning-based video offloading and resource allocation in noma-enabled networks. Future Internet 2023, 15, 184. [Google Scholar] [CrossRef]
- Ju, Y.; Chen, Y.; Cao, Z.; Liu, L.; Pei, Q.; Xiao, M.; Ota, K.; Dong, M.; Leung, V.C. Joint secure offloading and resource allocation for vehicular edge computing network: A multi-agent deep reinforcement learning approach. IEEE Trans. Intell. Transp. Syst. 2023, 24, 5555–5569. [Google Scholar] [CrossRef]
- Guo, Y.; Xu, X.; Xiao, F. MADRLOM: A Computation Offloading Mechanism for Software-Defined Cloud-Edge Computing Power Network. Comput. Netw. 2024, 245, 110352. [Google Scholar] [CrossRef]
- Wang, Q.; Wang, Q.; Zhao, H.; Zhang, H.; Zhu, H.; Wang, X. Device-Specific QoE Enhancement Through Joint Communication and Computation Resource Scheduling in Edge-Assisted IoT Systems. IEEE Internet Things J. 2024, 11, 13257–13270. [Google Scholar] [CrossRef]
- Maray, M.; Mustafa, E.; Shuja, J.; Bilal, M. Dependent task offloading with deadline-aware scheduling in mobile edge networks. Internet Things 2023, 23, 100868. [Google Scholar] [CrossRef]
- Tang, T.; Li, C.; Liu, F. Collaborative cloud-edge-end task offloading with task dependency based on deep reinforcement learning. Comput. Commun. 2023, 209, 78–90. [Google Scholar] [CrossRef]
- Nguyen, L.X.; Tun, Y.K.; Dang, T.N.; Park, Y.M.; Han, Z.; Hong, C.S. Dependency tasks offloading and communication resource allocation in collaborative UAVs networks: A meta-heuristic approach. IEEE Internet Things J. 2023, 10, 9062–9076. [Google Scholar] [CrossRef]
- Liu, J.; Ren, J.; Zhang, Y.; Peng, X.; Zhang, Y.; Yang, Y. Efficient dependent task offloading for multiple applications in MEC-cloud system. IEEE Trans. Mob. Comput. 2021, 22, 2147–2162. [Google Scholar] [CrossRef]
- Liu, S.; Yu, Y.; Lian, X.; Feng, Y.; She, C.; Yeoh, P.L.; Guo, L.; Vucetic, B.; Li, Y. Dependent task scheduling and offloading for minimizing deadline violation ratio in mobile edge computing networks. IEEE J. Sel. Areas Commun. 2023, 41, 538–554. [Google Scholar] [CrossRef]
- Al-hammadi, I.; Li, M.; Islam, S.M.; Al-Mosharea, E. A hierarchical hybrid subtask scheduling algorithm in UAV-assisted MEC emergency network. IEEE Internet Things J. 2021, 9, 12737–12753. [Google Scholar]
- Zhang, X.; Lin, T.; Lin, C.K.; Chen, Z.; Cheng, H. Computational Task Offloading Algorithm Based on Deep Reinforcement Learning and Multi-Task Dependency. Theor. Comput. Sci. 2024, 993, 114462. [Google Scholar] [CrossRef]
- Hosny, K.M.; Awad, A.I.; Khashaba, M.M.; Fouda, M.M.; Guizani, M.; Mohamed, E.R. Enhanced multi-objective gorilla troops optimizer for real-time multi-user dependent tasks offloading in edge-cloud computing. J. Netw. Comput. Appl. 2023, 218, 103702. [Google Scholar] [CrossRef]
- Zeng, C.; Wang, X.; Zeng, R.; Li, Y.; Shi, J.; Huang, M. Joint optimization of multi-dimensional resource allocation and task offloading for QoE enhancement in Cloud-Edge-End collaboration. Future Gener. Comput. Syst. 2024, 155, 121–131. [Google Scholar] [CrossRef]
- Sun, Y.; Li, H.; Wei, T.; Zhang, Y.; Wang, Z.; Wu, W.; Fang, C. Dependency-aware flexible computation offloading and task scheduling for multi-access edge computing networks. In Proceedings of the 2021 24th International Symposium on Wireless Personal Multimedia Communications (WPMC), Okayama, Japan, 14–16 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar]
- Shu, C.; Zhao, Z.; Han, Y.; Min, G.; Duan, H. Multi-user offloading for edge computing networks: A dependency-aware and latency-optimal approach. IEEE Internet Things J. 2019, 7, 1678–1689. [Google Scholar] [CrossRef]
- Hou, W.; Wen, H.; Song, H.; Lei, W.; Zhang, W. Multiagent deep reinforcement learning for task offloading and resource allocation in cybertwin-based networks. IEEE Internet Things J. 2021, 8, 16256–16268. [Google Scholar] [CrossRef]
- Kang, H.; Chang, X.; Mišić, J.; Mišić, V.B.; Fan, J.; Liu, Y. Cooperative UAV resource allocation and task offloading in hierarchical aerial computing systems: A MAPPO based approach. IEEE Internet Things J. 2023, 10, 10497–10509. [Google Scholar] [CrossRef]
- Song, S.; Ma, S.; Zhao, J.; Yang, F.; Zhai, L. Cost-efficient multi-service task offloading scheduling for mobile edge computing. Appl. Intell. 2022, 52, 4028–4040. [Google Scholar] [CrossRef]
- Chen, S.; Wang, X.; Sun, Y. TODO: Task offloading decision optimizer for the efficient provision of offloading schemes. Pervasive Mob. Comput. 2024, 99, 101892. [Google Scholar] [CrossRef]
- Zaman, S.K.U.; Jehangiri, A.I.; Maqsood, T.; Haq, N.U.; Umar, A.I.; Shuja, J.; Ahmad, Z.; Dhaou, I.B.; Alsharekh, M.F. LiMPO: Lightweight mobility prediction and offloading framework using machine learning for mobile edge computing. Clust. Comput. 2023, 26, 99–117. [Google Scholar] [CrossRef]
- Xu, X.L.; Fang, Z.J.; Qi, L.; Dou, W.; He, Q.; Duan, Y. A deep reinforcement learning-based distributed service off loading method for edge computing empowered internet of vehicles. Chin. J. Comput. 2021, 44, 2382–2405. [Google Scholar]
- Zhang, X.; Wu, W.; Wang, J.; Liu, S. BiLSTM-based Federated Learning Computation Offloading and Resource Allocation Algorithm in MEC. ACM Trans. Sens. Netw. 2023, 19, 1–20. [Google Scholar] [CrossRef]
- Meng, L.; Wang, Y.; Wang, H.; Tong, X.; Sun, Z.; Cai, Z. Task offloading optimization mechanism based on deep neural network in edge-cloud environment. J. Cloud Comput. 2023, 12, 76. [Google Scholar] [CrossRef]
- Ai, Z.; Zhang, W.; Li, M.; Li, P.; Shi, L. A smart collaborative framework for dynamic multi-task offloading in IIoT-MEC networks. Peer—Peer Netw. Appl. 2023, 16, 749–764. [Google Scholar] [CrossRef]
- Zhang, J.; Wang, J.; Yuan, Z.; Zhang, W.; Liu, L. Offloading demand prediction-driven latency-aware resource reservation in edge networks. IEEE Internet Things J. 2023, 10, 13826–13836. [Google Scholar] [CrossRef]
- Yao, W. The application of artificial intelligence in the internet of things. In Proceedings of the 2019 International Conference on Information Technology and Computer Application (ITCA), Guangzhou, China, 20–22 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 141–144. [Google Scholar]
- Yan, J.; Bi, S.; Zhang, Y.J.A. Offloading and resource allocation with general task graph in mobile edge computing: A deep reinforcement learning approach. IEEE Trans. Wirel. Commun. 2020, 19, 5404–5419. [Google Scholar] [CrossRef]
- Miao, Y.; Wu, G.; Li, M.; Ghoneim, A.; Al-Rakhami, M.; Hossain, M.S. Intelligent task prediction and computation offloading based on mobile-edge cloud computing. Future Gener. Comput. Syst. 2020, 102, 925–931. [Google Scholar] [CrossRef]
- Qin, W.; Chen, H.; Wang, L.; Xia, Y.; Nascita, A.; Pescapè, A. MCOTM: Mobility-aware computation offloading and task migration for edge computing in industrial IoT. Future Gener. Comput. Syst. 2024, 151, 232–241. [Google Scholar] [CrossRef]
- Shewalkar, A.; Nyavanandi, D.; Ludwig, S.A. Ludwig S A. Performance evaluation of deep neural networks applied to speech recognition: RNN, LSTM and GRU. J. Artif. Intell. Soft Comput. Res. 2019, 9, 235–245. [Google Scholar] [CrossRef]
- Zhou, X.; Bai, G.; Tao, J.; Xu, B. An improved method to search all minimal paths in networks. IEEE Trans. Reliab. 2023, 72, 1420–1431. [Google Scholar] [CrossRef]
- Zhou, H.; Wang, Z.; Min, G.; Zhang, H. UAV-aided computation offloading in mobile-edge computing networks: A Stackelberg game approach. IEEE Internet Things J. 2022, 10, 6622–6633. [Google Scholar] [CrossRef]
- Lowe, R.; Wu, Y.I.; Tamar, A.; Harb, J.; Pieter Abbeel, O.; Mordatch, I. Multi-agent actor-critic for mixed cooperative-competitive environments. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. Advances in neural information processing systems. [Google Scholar]
- Li, S.; Wu, Y.; Cui, X.; Dong, H.; Fang, F.; Russell, S. Robust multi-agent reinforcement learning via minimax deep deterministic policy gradient. In Proceedings of the AAAI conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; AAAI: Palo Alto, CA, USA, 2019; pp. 4213–4220. [Google Scholar]
- Ra, M.R.; Sheth, A.; Mummert, L.; Pillai, P.; Wetherall, D.; Govindan, R. Odessa: Enabling interactive perception applications on mobile devices. In Proceedings of the 9th International Conference on Mobile Systems, Applications, and Services, Bethesda MD, USA, 28 June–11 July 2011; Association for Computing Machinery: New York, NY, USA, 2011; pp. 43–56. [Google Scholar]
- Wang, J.; Hu, J.; Min, G.; Zhan, W.; Zomaya, A.Y.; Georgalas, N. Dependent task offloading for edge computing based on deep reinforcement learning. IEEE Trans. Comput. 2021, 71, 2449–2461. [Google Scholar] [CrossRef]
- Zhang, T.; Xu, Y.; Loo, J.; Yang, D.; Xiao, L. Joint computation and communication design for UAV-assisted mobile edge computing in IoT. IEEE Trans. Ind. Inform. 2019, 16, 5505–5516. [Google Scholar] [CrossRef]
Parameters | Value |
---|---|
Location and service radius of edge nodes | (1,2,0) (2,1,0) (2,2,0.1) 1.5 km |
Location and sensing radius of the agent | (1,1,0) (3,1,0) (2,3,0) 1.5 km |
Terminal CPU frequency | 2 × 1010 cycles/s |
CPU frequency of the edge node | 8 × 1011 cycles/s |
Task computing resource allocation ratio | 0.27 × task size bit |
Total computing resources of the terminal | 10 MB |
Total computing resources of the edge nodes | 35 MB |
Uplink/downlink transmission power of edge nodes | 35 W |
Transmission bandwidth of edge nodes | 10 MHz |
Task data size | [640, 6400] KB |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, Y.; Huang, Z.; Wei, Z.; Zhao, J. MADDPG-Based Offloading Strategy for Timing-Dependent Tasks in Edge Computing. Future Internet 2024, 16, 181. https://doi.org/10.3390/fi16060181
Wang Y, Huang Z, Wei Z, Zhao J. MADDPG-Based Offloading Strategy for Timing-Dependent Tasks in Edge Computing. Future Internet. 2024; 16(6):181. https://doi.org/10.3390/fi16060181
Chicago/Turabian StyleWang, Yuchen, Zishan Huang, Zhongcheng Wei, and Jijun Zhao. 2024. "MADDPG-Based Offloading Strategy for Timing-Dependent Tasks in Edge Computing" Future Internet 16, no. 6: 181. https://doi.org/10.3390/fi16060181
APA StyleWang, Y., Huang, Z., Wei, Z., & Zhao, J. (2024). MADDPG-Based Offloading Strategy for Timing-Dependent Tasks in Edge Computing. Future Internet, 16(6), 181. https://doi.org/10.3390/fi16060181