Next Article in Journal
The Role of Human Capital and Energy Transition in Driving Economic Growth in Sub-Saharan Africa
Previous Article in Journal
Research on Synergy Measurement and Digital Finance Driving Mechanism of Enterprise Digital Transformation and Greening Upgrade: An Empirical Analysis Based on the Complex System Coordination Degree Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Delay-Aware Distributed Cloud–Edge Scheduling Framework and Algorithm in Dynamic Network Environments

1
State Key Laboratory of Thermal Energy and Power on Ships, Wuhan 430000, China
2
Wuhan Second Ship Design and Research Institute, Wuhan 430000, China
3
College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Sustainability 2025, 17(11), 4887; https://doi.org/10.3390/su17114887
Submission received: 6 April 2025 / Revised: 17 May 2025 / Accepted: 23 May 2025 / Published: 26 May 2025

Abstract

This paper proposes a distributed cloud–edge collaborative scheduling method to address the oversight of network transmission delay in traditional task scheduling, a critical factor that frequently leads to degraded execution efficiency. A holistic framework is introduced that dynamically models transmission delays, designs a decentralized scheduling algorithm, and optimizes resource competition through a two-dimensional matching mechanism. The framework integrates real-time network status monitoring to adjust task allocation, enabling edge nodes to independently optimize local queues and avoid single-point failures. A delay-aware scheduling algorithm is developed to balance task computing requirements and network latency, transforming three-dimensional resource matching into a two-dimensional problem to resolve conflicts in shared resource allocation. Simulation results verify that the method significantly reduces task execution time and queue backlogs compared with benchmark algorithms, demonstrating improved adaptability in dynamic network environments. This study offers a novel approach to enhancing resource utilization and system efficiency in distributed cloud–edge systems.

1. Introduction

With the rapid advancement of mobile Internet and IoT technologies, traditional centralized cloud computing architectures have struggled to meet the low-latency demands for real-time services, often causing core network congestion and privacy vulnerabilities [1]. This challenge has catalyzed the emergence of cloud–edge collaborative computing architectures that synergize cloud-scale data processing with edge nodes’ localized responsiveness, thus forming a “cloud–edge–end” triad that optimizes resource utilization and service quality [2,3]. As illustrated in Figure 1, the Network Collaborative Open Control Architecture (NCOCA) epitomizes this paradigm through distributed autonomous decision-making and entity-level coordination, which contrasts sharply with legacy centralized systems prone to single-point failures and scalability constraints [4]. By decentralizing management functions to network entities, NCOCA enhances operational efficiency while enabling mission-critical applications such as millisecond-response power grid inspections and the distributed vehicular network processing of dynamic road conditions [5].
However, existing architectures still face serious challenges in coordinating heterogeneous device capabilities and meeting real-time mission requirements. Heterogeneous devices, varying widely in computational, storage, and network capabilities (e.g., grid sensors, IoT terminals, industrial park energy devices, fog computing nodes, and industrial IoT equipment), present significant challenges in resource integration, interoperability, and dynamic access stability. For example, it is necessary to resolve data transmission and task offloading strategies among cameras, edge controllers, and cloud servers due to the computational disparities among these devices [6]. Ren et al. [7] and Deng et al. [8] highlight how the fragmented capabilities of IoT terminals, photovoltaic nodes, and energy storage devices give rise to cross-domain collaboration and dynamic resource scheduling issues. Liu et al. [9] proposed an edge-cloud collaborative information interaction mechanism and distributed iterative optimization algorithm for industrial park integrated energy systems, which can effectively improve multi-energy synergy efficiency. In smart factories and energy systems, fog computing nodes and industrial sensors face reliability challenges during dynamic access owing to the complexities of heterogeneous resource management [10,11]. Dynamic tasks further complicate this landscape with strict real-time requirements—the use of real-time traffic processing in telematics, UAV target tracking, and analogous tasks demand millisecond-level responses to handle load fluctuations and tight time windows—thereby avoiding decision-making errors [7,8,12]. Smart logistics scenarios experience amplified scheduling complexity stemming from environmental uncertainties such as abrupt path changes or equipment failures [13,14]. Additionally, grid video surveillance and fog computing tasks are constrained by the need for latency-sensitive real-time data processing [10,15]. The main contributions of this paper to address the “fragmented” capabilities of heterogeneous devices and the “real-time” requirements of dynamic tasks are as follows:
(1)
Delay-driven dynamic scheduling mechanism: the real-time monitoring of network status (channel quality, node load), the dynamic adjustment of task allocation strategy, and the resolution of resource competition conflicts;
(2)
Decentralized collaborative decision-making framework: design distributed algorithms that allow edge nodes to autonomously optimize local queues, avoiding the single-point failure inherent in centralized architectures;
(3)
Lightweight 2D matching mechanism: Optimize resource allocation for high-priority tasks by incorporating latency and queue backlog perception and downgrading 3D resource matching to a 2D problem. Through comparisons with traditional algorithms (PDPRA, CSA, simple neural networks) in NS-3 simulations, the proposed approach demonstrates excellent adaptability and robustness in large-scale dynamic network environments.
The rest of this paper is as follows: Section 2 provides a comprehensive overview of related works. Section 3 elaborates on the distributed cloud–edge collaborative control modeling with consideration of network transmission delay. Section 4 introduces the cloud–edge collaborative data processing optimization algorithm based on processing delay and backlog perception matching. In Section 5, the effectiveness of the proposed distributed cloud–edge collaborative scheduling algorithm (with network transmission delay constraints) is validated through simulation experiments and compared against three other algorithms. Section 6 concludes the entire work. The overall research content framework is shown in Figure 2.

2. Related Work

The development of cloud–edge collaborative architectures aims to optimize latency and energy consumption through distributed resource management, while existing research focuses on task offloading strategies, resource allocation optimization, and algorithm design to balance multiple objectives such as latency, energy consumption, and quality of service (QoS).
Earlier studies assumed task independence or only considered simple dependencies, a limitation that makes models insufficient for complex workflows. Karimi et al. [16] proposed a deep reinforcement learning (DRL)-based task offloading framework to jointly optimize the resource allocation between edge and cloud nodes, thereby reducing response time and improving task acceptance rates. However, this framework fails to account for the impact of network bandwidth fluctuations. Li et al. [17] designed an edge–cloud cooperative algorithm based on Lyapunov stability theory to balance queue delay and energy consumption, but it overlooked inter-task dependencies, leading to limited scheduling efficiency for complex workflows.
To address the task dependency problem, Tang et al. [18] introduced a directed acyclic graph (DAG) to describe task dependency and proposed a deep reinforcement learning algorithm based on task prioritization (TPDRTO), improving the scheduling efficiency of the dependent tasks. Notably, its high deep neural network training complexity in large-scale scenarios and real-time performance are insufficient. Wang et al. [19] combined blockchain technology and incentive mechanism to facilitate edge node collaboration via smart contracts, but this approach increased system communication overhead, potentially degrading overall performance. Chen et al. [20] proposed the concept of computational task contouring for complex workflows in the medical and industrial domains, designed a DRL task scheduling algorithm to optimize resource usage and computational cost, and verified that it outperformed the baseline approach through NS-3 simulation. However, the algorithm did not address the coupling effect of bandwidth allocation. Collectively, these studies highlight that modeling task dependencies and achieving efficient scheduling remain outstanding challenges.
Heuristic and meta-heuristic algorithms are widely used in energy and performance optimization. Hussain et al. [21,22] proposed Energy and Performance Efficient Task Scheduling Algorithms (EPETS), which dynamically adjust resource allocation to reduce the overuse of high-speed resources via task slack time. However, this approach is only applicable to standalone tasks. The multi-quantum-inspired genetic algorithm (MQGA) proposed by the authors’ team addresses medical workflow scheduling to optimize Makespan and energy consumption, but it suffers from limitations imposed by quantum computing hardware environments and lacks universal applicability.
Data-driven heuristics are gradually becoming mainstream. Bao et al. [23] proposed a gated recurrent unit (GRU)-based edge–cloud cooperative prediction model (TP-ECC), which dynamically adjusts resource allocation by integrating traffic prediction to enhance system foresight, but it is confined to single-edge-node scenarios. Liu et al. [24] combined CNN, LSTM, and attention mechanism to achieve efficient decision-making in dynamic environments with 30–50% computational efficiency improvements, but this approach relied on large amounts of labeled data and exhibited limited generalization ability. The MoDEMS system used the sequential greedy algorithm to generate migration plans, applying a Markovian mobility prediction model to reduce the user latency, but it overlooked global suboptimality arising from resource competition among multiple users [25]. To address the Pareto-optimality challenge in multi-objective optimization, Pang et al. [26] proposed a dynamic weight adjustment algorithm, which balances latency and energy consumption optimization objectives in real time via reinforcement learning and mitigates the insufficient adaptability of traditional static weight allocation strategies in dynamic telematics environments. However, it fails to account for the compatibility issues of heterogeneous devices.
Exact optimization methods based on mathematical planning, such as Integer Linear Programming (ILP) or Mixed-Integer Programming (MIP), theoretically approximate the optimal solution, but the computational complexity grows exponentially with the task size. The DEWS algorithm produced by Hussain et al. [27] constructs an optimization model, which is solved by the Variable Neighborhood Descent (VND) algorithm, to minimize the energy cost under deadline constraints, yet its time complexity O ( n 3 ) renders it ill-suited for large-scale real-time scenarios. The subsequent EPEE algorithm integrates electricity price fluctuation and data transmission time to achieve energy efficiency and cost co-optimization, but it still assumes static data center resource allocation and neglects real-time load variations [28].
The time-varying nature of edge node resources and network bandwidth in dynamic environments thus causes traditional static scheduling strategies to fail. Although algorithms such as DRL try to handle dynamic states, real-time performance and convergence in large-scale scenarios still need to be optimized. For example, CTOCLA introduces an attention mechanism to handle task weights, but fails to model the additional delay induced by cross-domain data transmission security and privacy protection [24]. MoDEMS exhibits limited dynamic adaptability, as abrupt changes in user mobility patterns increase migration plan failure probability by approximately 20% [25]. To tackle dynamic resource fluctuations, Jiang et al. [29] proposed a deep reinforcement learning (DRL)-based cloud–edge cooperative multilevel dynamic reconfiguration method for the real-time optimization of urban distribution networks (UDNs). The method is divided into three levels (feeder, transformer, and substation), combining a multi-intelligent system (MAS) with a cloud–edge architecture, and adopting a “centralized training-decentralized execution” (CTDE) model. Meanwhile, Li et al. [30] proposed a cloud–edge cooperative approach based on MADDPG for the integrated electric–gas system (IEGS) scheduling problem. The electricity and gas systems are computed independently at the edge, and the cloud optimizes globally based on the boundary information. However, it overlooks the parallelism limitation stemming from the cloud’s reliance on edge parameter transfers.
There are three systemic deficiencies. Firstly, there is a lack of dynamic delay modeling, ignoring the impact of network fluctuations and interference on task offloading. This results in delay estimation bias. Secondly, there is insufficient decentralized scheduling, and the centralized architecture suffers from lagging state synchronization and privacy risks. Thirdly, there is an imbalance in multi-objective optimization, lacking a joint optimization framework for multiple objectives such as latency and energy consumption, making it difficult to meet the demands of complex scenarios. This paper systematically addresses these gaps through dynamic latency modeling, distributed collaborative decision-making, and lightweight resource matching:
(1)
Dynamic adaptability: avoiding static model bias by the real-time sensing of network state.
(2)
Privacy and reliability: decentralized architecture circumvents a single point of failure.
(3)
Multi-objective synergy: balancing delay, energy consumption, and resource fragmentation through a two-dimensional matching mechanism.

3. Distributed Cloud–Edge Collaborative Control Modeling Considering Network Transmission Delay

A distributed cloud–edge collaborative data processing architecture is a system that combines computational power from both cloud and edge devices while accounting for network transmission latency. This architecture optimizes data processing to improve efficiency, reduce latency, and alleviate reliance on network bandwidth. Data generated or collected at edge devices can be transmitted to either the cloud or other edge nodes for processing. To mitigate the impact of transmission latency on performance, designers employ strategies such as edge-side preprocessing and adaptive task partitioning.
To address the dynamic interaction between network transmission and task processing, this section introduces a distributed cloud–edge collaborative control model. We first model the data backlog dynamics at terminals (Section 3.1), subsequently presenting a transmission model that integrates channel characteristics and resource allocation (Section 3.2). Finally, we systematically derive a comprehensive latency model (Section 3.3) that combines computational delays at servers with network transmission delays across the cloud–edge continuum, serving as the foundation for the optimization algorithm in Section 4. The nomenclature used throughout the paper is systematically presented and explained in Table 1.

3.1. Production Equipment Data Backlog Model

Assume that there is I data acquisition terminal, represented by D = { d 1 , , d i , , d I } . T data processing iterations are considered, expressed as T = 1 , , t , , T . In each iteration, according to the available communication and computing resources, the terminal will offload the locally cached data to the edge layer or cloud layer for collaborative data processing. The locally cached data of the terminal can be simulated using the cache queue, and the queue backlog updates can be represented by the following formula:
Q i ( t + 1 ) = Q i ( t ) + A i ( t ) U i ( t )
where A i ( t ) and U i ( t ) are the amounts of data collected and transmitted in the t iteration, representing the input and output of the queue, respectively.
Equation (1) describes the dynamic balance of terminal cache: the current backlog equals the previous backlog plus new data minus transmitted data. This captures the accumulation of data that requires offloading, directly influencing subsequent transmission and processing decisions.

3.2. Intelligent Operation and Maintenance Data Transmission Model

Assume that there are a total of K servers that form the set C = c 0 , c 1 , , c k , , c K , where c 0 represents the cloud server and c 1 , , c k , , c K represents the edge server. Considering the end–edge channel, the end–cloud channel, and the cloud–edge load profile, the available communication and computation resources are partitioned into J resource blocks, denoted as F = { f 1 , , f j , , f j } . In the tth iteration, the signal-to-noise ratio of the terminal d i transmitting the captured data to the server c k through the resource block f j is expressed as follows:
γ i , j , k t = P i ( t ) g i , j , k ( t ) δ 2
We define the matching indicator variable between the terminal and the resource block as x i , j ( t ) 0 , 1 . x i , j ( t ) = 1 shows that the terminal d i matches with the resource block f j at the tth iteration; otherwise, x i , j ( t ) = 0 . We define the matching indicator variable between the terminal and the server as y i , j ( t ) 0 , 1 . y i , j ( t ) = 1 shows that the terminal d i offloads the data to the server f j at the tth iteration; otherwise, y i , j ( t ) = 0 . Based on Equation (2), d i the transmission rate for the transfer of data to the server c K from the terminal f j is calculated as follows:
R i , j , k ( t ) = B i , j , k   log 2 1 + γ i , j , k ( t )
The amount of data U i ( t ) transmitted in the tth iteration is expressed as follows:
U i ( t ) = min Q i ( t ) + A i ( t ) , τ j = 1 J k = 1 K x i , j ( t ) y i , k ( t ) R i , j , k ( t )
where τ denotes the iterative transmission length.
The data transmission delay for transferring data from f j to c k is expressed as follows:
T i , j , k , t r ( t ) = U i ( t ) R i , j , k ( t )
Consider T data processing iterations, denoted as T = 1 , , t , , T . In each iteration, the terminal offloads the locally cached collection data to the edge layer or the cloud layer for collaborative data processing, depending on the available communication and computation resources. The locally cached data is modeled by the cache queue shown in Equation (1).
Equations (2)–(5) link channel quality (SNR) to transmission efficiency: higher SNR (better channel) enables higher rates and lower delays. This forms the basis of resource block allocation, as terminals prefer resource blocks with higher γ i , j , k t to minimize transmission latency.

3.3. Cloud–Edge Data Co-Computation and Data Processing Latency Modeling

The edge server or cloud server can process the data from the endpoints only after it has processed its own computational load. The latency to process the collected data from d i is expressed as follows:
T i , k , c ( t ) = T i , k , w ( t ) + α i U i ( t ) j = 1 J x i , j ( t ) l j ( t )
where T i , k , w is the c K delay when waiting to complete its computational load, α i is the computational resources required to process the unit d i data, i.e., computational complexity, and l i ( t ) is the computational resources provided by the resource block f j .
The total latency incurred during task offloading consists of two main components. The first part is the CPU cycle consumed by the execution of the task on the server and this is termed execution latency. The other part is the time taken by the task to transfer between different servers during the offloading process. The execution latency further comprises the task execution latency on the server can be subdivided into the processing latency of the task and the queuing latency waiting for the previous task to finish processing. The task processing latency is usually related to the server performance and the computational resources required by the task. Let the total offloading latency of the system be T, where the execution latency is T e x e c i , the queuing latency is T q u e u i , the processing latency is T h a n l e i , the offloading latency for transmitting the task to other servers is T u p i , and the transmission latency for the task to be transmitted from one edge server to another edge server is T t r a n s i . The execution delay T e x e c i of task T a s k i , k is expressed in Equation (7).
T e x e c i = D i , k C j
Let K be the set of offloaded yet incomplete tasks on the ECS; the queuing latency T a s k i , k of T q u e u i is shown in Equation (8).
T q u e u e i = 0 , j = 1 n + 1 u i j = 0 i k j T e x e c i , o t h e r
The offload latency of a task can be decomposed into six parts, including the offload latency, the transmission latency between edge servers, the offload latency from the edge servers to the macro base station, the offload latency from the edge servers to the central cloud server, and the offload latency from the macro base station to the central cloud server. The offload delay of the task transmission to another server depends on the size of the task and the transmission power of the terminal. The offload delay of the task can be formulated using Equation (9).
T u p i = Z i , k θ ν 2 E
The binary variable y i j indicates whether the transmission uses the cross-edge server mode, and y i j = 1 indicates that the transmission is performed in cross-edge server mode. T a s k i , k , the transmission delay T t r a n s i between edge servers, is determined using Equation (10).
T t r a n s i = i = 1 n 1 z i , j y i j R E 2 E j
When a task cannot be offloaded on the local terminal, the cloud server will assign a more reasonable transmission mode to the task—that is, the wireless transmission mode—and set the transmission delay to T V 2 V i , as shown in Equation (11).
T ν 2 ν i = z i , k θ ν 2 ν ( 1 h i ) ρ
h i indicates the task transmission mode selected by the cloud server, ρ represents the task transmission times, and θ ν 2 ν indicates the task transmission rate of the terminal.
Due to the limited storage and computing resources of the edge server, when the edge server cannot handle task offloading, it will offload the task to the central cloud server. However, due to the long distance between the edge server and the central cloud server, frequent tasks will generate large energy consumption, which reduces the utilization rate of system resources. In this architecture, macro base stations with large computing resources and storage resources are placed reasonably near to edge server farms to greatly reduce the probability of requests to the central cloud server, thereby reducing delay and system energy consumption and effectively improving resource utilization. Then, the transmission delay of T a s k i , k from the edge server to the macro base station is T E 2 M i , as shown in Equation (12).
T E 2 M i = Z i , k u i j f i j 6 R M 2 E j , j = n + 1
where A i ( t ) and U i ( t ) are the amounts of data collected and transmitted in the t iteration, representing the input and output of the queue, respectively.
T E 2 C i = Z i , k u i j R C 2 E j , j = n + 1
We set the transmission delay of T a s k i , k offloaded data to the central cloud server through the macro base station T M 2 C i , as shown in Equation (14).
T E 2 C i = Z i , k u i j ( 1 f i j ) R C 2 E j , j = n + 1
The tasks generated by the terminal can be executed on four types of servers: terminal-local, edge server, macro-base station, and central cloud server. The total offloading delay of T a s k i , k is T i j , as shown in Equation (15).
T i j = T c a c h e i + T e x e c i , j = 1 n + 1 u i j = 0 T c a c h e i + T e x e c i + T q u e u e i + T u p i + ( 1 h i ) T t r a n s i + h i T v 2 v i , j = 1 n + 1 u i j = 1 T c a c h e i + T e x e c i + T q u e u e i + T u p i + ( 1 h I ) T t r a n s i h i T v 2 v i + f i j T E 2 C i + ( 1 f i j ) T E 2 C i
The total offloading delay of the system is shown in Equation (16).
T = i = 1 d T i j

4. Cloud–Edge Collaborative Data Processing Optimization Algorithm Based on Processing Delay and Backlog Sensing Matching

In this section, a cloud–edge collaborative data processing optimization algorithm based on processing delay and backlog perception matching is proposed to solve the above optimization problems. The algorithm’s principles and implementation process are described as follows.

4.1. Cloud–Edge Collaborative Data Processing Optimization Problem Construction

The goal of cloud–edge collaborative data processing is to achieve optimal matching among terminals, resource blocks, and servers while minimizing latency and reducing endpoint data queue backlogs. The optimization problem is formulated in Equation (17).
          min x i , j ( t ) , y i , k ( t ) 1 T t = 1 T i = 1 l V Q i t + j = 1 J k = 1 K x i , j ( t ) y i , k ( t ) T i , j , k , s t s . t . C 1 : i = 1 l x i , j t 1 , d i D , f j F , t T             C 2 : j = 1 J x i , j ( t ) 1 , d i D , f j F , t T             C 3 : k = 1 K y i , k t 1 , d i D , c k C , t T             C 4 : i = 1 l y i , k t q k , d i D , c k C , t T
where V is the weight used to weigh the optimization of processing latency and data queue backlog, and q k is the quota of the server. When V tends towards 1, the algorithm gives priority to minimizing latency, which is suitable for real-time tasks; when V tends to 0, the algorithm gives priority to alleviating queue backlog, which is suitable for data throughput-sensitive scenarios. C 1 and C 2 are constraints on terminals matching resource blocks, indicating that each terminal can match at most one resource block and that each resource block can be used by at most one terminal to avoid transmission conflicts caused by multi-channel competition. C 3 and C 4 are constraints on terminals matching servers, indicating that each endpoint can be configured with at most one server to limit endpoint single-path offloads and server load ceilings to prevent steep latency increases in edge/cloud nodes due to task overload, and each server can serve a maximum of q k terminals.
The optimization problem is essentially a three-dimensional resource-matching problem (involving terminals, resource blocks, and servers), which is transformed into a two-dimensional matching problem by combining resource blocks and servers into combinatorial options. However, due to the inclusion of binary decision variables x i , j ( t ) , y i , k ( t ) and coupling constraints, the problem still falls into the category of NP-hard Integer Linear Programming (ILP), and traditional exact algorithms are unable to solve large-scale instances in polynomial time. The multi-objective optimization property (weighted summation processing delay with queue backlog) further increases the solution complexity and requires us to rely on heuristic algorithms for approximate optimization. The detailed proof procedure for the optimization problem is provided in Appendix A.

4.2. Delay Queue-Aware Dynamic Resource Matching

The principle of the algorithm is illustrated in Figure 3. By integrating resource blocks and servers into new matching units, the complexity of three-dimensional matching among terminals, resource blocks, and servers is reduced, transforming the issue into a two-dimensional matching problem between terminals and resource block–server combination units. Using this approach, the cloud–edge collaborative data processing matching preference of terminals is formulated by incorporating the processing delays of cloud/edge servers and the backlogs of terminal data queues. According to the matching preference list, matching is achieved among terminals, resource blocks, and servers. Finally, terminals selectively offload data to edge or cloud servers for collaborative processing based on the matching outcomes.

4.3. Implementation Process

The implementation process of the algorithm is shown below and it includes three steps: processing delay and backlog perception; cloud–edge data processing matching preference construction based on processing delay and backlog perception; and cloud–edge collaborative data processing matching iterative optimization considering resource competition. The details are elaborated as follows.
  • Communication network service orchestration transmission delay model
The edge layer combines information such as the amount of data collected by the terminal, end-side channel status, end-cloud channel status, cloud-available computing resources, and edge–cloud load to determine the processing delay and backlog. Specifically, for the resource block–server combination option s j , k , Equation (6) is used to evaluate the processing delay of the server c k . For the terminal d i , the backlog of its data queue is evaluated according to Equation (1).
2.
Cloud–edge data processing matching preference construction based on processing delay and backlog perception
Based on processing delay and backlog perception, cloud–edge data processing matching preferences are constructed. We define the terminal d i preference value for the resource block server combination option s j , k to be θ i , j , k ( t ) , denoted as follows:
θ i , j , k ( t ) = T i , j , k , s ( t ) V Q i ( t ) η i , j ( t ) η i , k ( t )
where η i , j ( t ) is the matching cost between d i and resource block f j , and η i , j ( t ) is the matching cost between d i and server c k . This which is used to solve the resource competition problem between multiple terminals and the same combination option.
The algorithm has a time complexity of O ( I J K log ( J K ) ) ( I is the number of terminals, J is the number of resource blocks, and K is the number of servers), and achieves polynomial-level complexity by transforming 3D matching into 2D matching and by employing distributed iteration to optimize the cost of matching and finite subconvergence; the space complexity is O ( I J K ) , which can be optimized by sparse storage. The algorithm solves the exponential complexity problem of traditional 3D matching by balancing the computational efficiency and optimization effect in the dynamic network with the help of a distributed architecture of autonomous decision-making by edge nodes. The detailed procedure of the proposed algorithm is presented in Algorithm 1.
Algorithm 1: Delay-Aware Dynamic Resource Matching Algorithm
Input: Terminal set D , Resource blocks F , Servers C , Iterations T , Priority weights V , Terminal priority weight K i
Output: total delay T total ,   total   queue   backlog   Q total ,   Matching   indicator   variables   x i , j ( t ) ,   y i , k ( t )
1:
Initialize   x i , j ( t ) = 0 ,   y i , k ( t ) = 0   for   all   i , j , k , t
2:
Initialize   matching   cos ts   η i , j ( t ) = 0 ,   η i , k ( t ) = 0
3:
For  t from 1 to T  do 
4:
      Processing Delay and Backlog Sensing
5:
       For   each   terminal   d i D  do
6:
             For   each   combination   option   s j , k = ( f j , c k )  do
7:
                   Calculate   transmission   delay   T i , j , k , tr ( t ) = U i ( t ) R i , j , k ( t )  
8:
                   Calculate   computation   delay   T i , k , c ( t ) = waiting   delay + α i U i ( t ) l j ( t )
9:
                   Total   delay   T i , j , k , s ( t ) = T i , j , k , tr ( t ) + T i , k , c ( t )
10:
            End for
11:
      End for
12:
       For   each   terminal   d i D  do
13:
            For   each   combination   option   s j , k S
14:
                  Preference value θ i , j , k ( t ) = T i , j , k , s ( t ) V Q i ( t ) η i , j ( t ) η i , k ( t )
15:
            End for
16:
             Sort   combination   options   in   descending   order   of   θ i , j , k ( t )   to   generate   preference   list   L i ( t )
17:
      End for
18:
      While there exist unmatched terminals do
19:
            For   each   unmatched   terminal   d i D  do
20:
                   Select   the   first   untried   combination   option   s j , k   in   L i ( t )
21:
                   Send   a   matching   request   to   s j , k
22:
            End for
23:
            For   each   combination   option   s j , k  do
24:
                   Count   the   number   of   requests   | A j |   ( for   resource   block )   and   Ω k (for server)
25:
                   If   | A j | 1   and   | Ω k | q k  then
26:
                         Matching   successful :   x i , j ( t ) = 1 ,   y i , k ( t ) = 1 ,   mark   d i as matched
27:
                  else
28:
                        For   each   conflicting   terminal   d i  do
29:
                               If   d i   as   low   priority   ( small   κ i ) then
30:
                                     η i , j ( t ) η i , j ( t ) + Δ p i , j κ i
31:
                                     η i , k ( t ) η i , k ( t ) + Δ p i , k κ i
32:
                              End If
33:
                         End For
34:
                         Recalculate   θ i , j , k ( t )   and   update   L i ( t )
35:
                  End If
36:
            End For
36:
      End While
38:
End for
Return   total   delay   T total ,   total   queue   backlog   Q total ,   x i , j ( t ) ,   y i , k ( t )
3.
Cloud–edge collaborative data processing matching iterative optimization considering resource competition
By increasing the matching cost of f j or c k , the terminal’s preference for the combination option s j , k is reduced, thus reducing the number of terminals applied to match the same combination option and resolving resource competition conflicts. In the process of matching costs, the data processing priorities of different service terminals should be considered, and the matching cost for low-priority services should be increased. The matching cost update formula of the resource block and server is expressed as follows:
η i , j ( t ) = η i , j ( t ) + Δ p i , j κ i η i , k ( t ) = η i , k ( t ) + Δ p i , k κ i
where Δ p i , j and Δ p i , k are the matching cost enhancement steps of d i and resource block f j and server c k , and κ i is the priority weight of d i . The larger κ i is, the higher the priority of d i .

5. Example Analysis

To verify the effectiveness of the proposed cloud–edge collaborative network architecture, the distributed cloud–edge collaborative control model, and the optimization algorithm based on processing delay and backlog perception, this section conducts experiments on the NS-3 simulation platform. The experimental objectives include verifying the adaptability of the network architecture, evaluating the effectiveness of the model and algorithm, and comparing its performance with the result of traditional methods. The experimental environment is configured with the NS-3 simulation platform (version 3.36), with hardware specifications of an Intel Core i5-4460 CPU, 3.8 GiB of memory, and a 64-bit Ubuntu 18.04.6 LTS operating system (running in a VMware virtualization environment with a 63.1 GB disk). The network topology is designed based on actual industrial operation and maintenance scenarios.
This paper builds a comprehensive network model, which includes a cloud system as the core hub. This is connected with three network switches, which are connected to build a high-availability internal network infrastructure. The model also includes three edge nodes, which are distributed in different geographical locations to provide services and data processing closer to users. In addition, two wireless nodes are integrated into the network to support the connection of mobile devices and sensors. The network architecture of edge–cloud collaboration is shown in Figure 4.
Based on the intelligent operation and maintenance scenario of a company’s actual production equipment, this paper considers various data acquisition terminals such as label data, operation data, maintenance data, and decommissioning data; the simulation is conducted using NS-3. Two simulation scenarios are as follows:
(1)
In scenario 1, data is sent from the sensor at end 1 → switch 2 → edge device 1→ switch 2 → switch 1 → cloud-side complex operation → output results to switch 1 0→ switch 2 → edge device 1 → switch 2 → return to the actuator at end 1.
(2)
In, scenario 2 data is sent from the sensor at end 1→ switch 2→ edge device 1→ switch 2 → switch 3 → switch 1 → cloud-side complex operation → output the result to switch 1 → switch 3 → switch 2 → edge device 1 → switch 2 → return to the actuator at end 1.
Use NetAnim to visualize the architecture model built in ns-3, as shown in Figure 5.
The initial IP addresses of each point are shown in Table 2.

5.1. Simulation Parameter

The initial values of the parameters and the initial coordinate settings for each node are shown in Table 3 and Table 4.

5.2. Contrast Algorithm Setting

In this paper, the performance of the proposed algorithm is verified, and three different cloud–edge collaborative data processing algorithms are used for comparison. The first algorithm is based on processing-delay-first resource block allocation (PDPRA), which is characterized by the transfer of all data to edge servers for processing. The second algorithm is the computing-resource-based algorithm (CSA), in which the terminal prefers the computing resources of the server but does not consider the quality of the channel or the backlog of the terminal data queue. The third algorithm, simple neural network-based resource allocation (Simple NN), employs real-time channel status and node load to generate resource allocation strategies. Although Simple NN offers basic online reasoning, it lacks dynamic adaptability and integration. Moreover, none of these three comparison algorithms can solve the problem of resource competition.
As illustrated in Figure 6, the proposed algorithm demonstrates superior performance over PDPRA, CSA, and Simple NN, with its task execution time curve consistently occupying the lowest region. The algorithm achieves an average execution time reduction of 28–35% compared to the benchmark methods. This can be mainly attributed to its multi-objective dynamic optimization framework, which integrates real-time channel quality monitoring, node load balancing, and queue backlog awareness. Through a conflict-aware weighted resource-matching strategy, the algorithm dynamically optimizes the trade-off between transmission and computation during iterative updates.
The performance comparison in Figure 7 shows that the proposed algorithm reduces data processing latency by 11.05%, 12.18%, and 18.4%; decreases data queue backlog by 13.35%, 15.2%, and 22.7%; and outperforms PDPRA, CSA, and Simple NN, respectively. These improvements can be attributed to the new algorithm’s consideration of the differences in processing delay caused by resource blocks and the differences in terminal data queue backlog caused by server selection, which are weighted according to different choices. In addition, the new algorithm can better solve the matching conflict, and updating the matching cost ensures a smaller delay and data queue backlog for high-priority terminals. In contrast, the PDPRA algorithm only optimizes processing latency while ignoring server resource competition and matching conflicts, resulting in suboptimal performance. The CSA algorithm neglects communication resource contention, whereas the Simple NN algorithm, constrained by its static neural network architecture, exhibits increased latency and worsened queue backlog in dynamic scenarios.
As demonstrated in Figure 8 and Figure 9, the proposed algorithm significantly outperforms PDPRA, CSA, and Simple NN, achieving 6.26%, 13.18%, and 19.2% reductions in data processing latency, and 13.57%, 16.16%, and 22.5% improvements in queue backlog mitigation, respectively. This improvement is due to the new algorithm’s integration of the communication and computing resources of the cloud and edge layers, which are organized into resource blocks and combined with cloud–edge servers through the Cartesian product. This approach allows terminals to better select servers and resource blocks and transfer data to edge or cloud servers for computation, significantly reducing data processing delays and data queue backlogs.
Figure 10 shows the weighted sum comparison of data processing delay and data queue backlog when the number of terminals increases in a large-scale device access scenario. As the number of terminals grows, the number of resource blocks that can be occupied by terminals decreases accordingly, resulting in an upward trend in the weighted sums of the three comparison algorithms, namely, PDPRA, CSA, and Simple NN. When the number of terminals exceeds 200, the number of terminals choosing the same server increases significantly, the processing capacity of communication and computation resource blocks decreases significantly, and the competition between terminals and servers for high-processing-capacity resource blocks intensifies, resulting in a steep increase in the weighted sum of the three comparison algorithms. In contrast, the proposed algorithm effectively mitigates the resource competition conflict by guiding terminals to preferentially match resource blocks with better processing capacity with server combinations through the matching cost iterative optimization mechanism in the case of declining resource processing capacity, and its weighting and growth are significantly smaller than those of the comparison algorithms.
Figure 11 shows the weighted sum comparison of data processing delay and data queue backlog when the number of edge servers increases in the large-scale device access scenario. When the number of edge servers increases, the total number of high-performance servers in the system rises, and the overall processing capacity of the edge layer increases significantly. Since the PDPRA, CSA, and Simple NN algorithms also achieve some processing capacity increase when server resources are sufficient, the weighted sums of all three algorithms show a decreasing trend. The proposed algorithm further dynamically selects the optimal servers through a delay- and backlog-aware mechanism to reduce the queue backlog while lowering the data processing delay when there is a surplus of server resources. As a result, its weighted sum performance decreases in a more pronounced fashion and is consistently superior to the values of the comparison algorithms.
At the same time, if there is a data queue backlog at the terminal of the current iteration, according to Equation (1), the backlog data will be transferred to the edge or cloud server for processing in the next iteration. Through the rational selection of resource blocks and servers by the new algorithm, the data queue backlog is kept at a low level, and the system performance is further optimized.

5.3. Simulation Result

The comparison of latency performance between Path 1 and Path 2 is shown in Figure 12.
The delay in Path 2 is significant due to the following reasons:
(1)
Additional Path Switching: In Path 2, the data needs to pass through three distinct network switches. The presence of multiple-path switching results in a longer transmission path, thereby increasing overall latency.
(2)
Increased Network Devices: Path 2 contains a relatively large number of network devices. Each time data is transferred from one switch to another, packets need to be processed and forwarded, which introduces delays in communication and coordination between different devices.
(3)
Potential Congestion: The involvement of additional network devices and path switching in Path 2 heightens the potential risk of network congestion. This congestion can lead to packet queuing during transmission, which further exacerbates latency.
Taking these factors into account, the latency of Path 2 is relatively high due to the introduction of additional network switches and path switching, as well as more complex data transmission routes. High latency can adversely affect applications and real-time settings that require low latency. Therefore, designing a network architecture requires a careful balance between performance requirements and network topology to select the most appropriate transmission path.
The selection of transmission media directly influences the delay characteristics of the system, and the reasonable selection can balance performance and cost factors while satisfying application requirements, making it a crucial decision in network design and optimization.
As verified by simulation, the proposed delay-aware distributed cloud-edge scheduling framework exhibits significant performance improvements in dynamic network environments. As shown in Figure 8 and Figure 9, the algorithm outperforms the traditional approaches (PDPRA, CSA, Simple NN) by reducing the data processing latency by 6.26–19.2%, and improving the queue backlog alleviation rate by 13.57–22.5%. This superiority stems from its unique integration of the following components:
(1)
Dynamic Network Modeling: Real-time monitoring of channel quality and node load, enabling adaptive task allocation and reducing transmission delays caused by outdated network state information.
(2)
Distributed 2D Matching: By transforming 3D resource matching (endpoint-resource block–server) into a 2D problem, the algorithm efficiently resolves shared resource conflicts and reduces overall task execution time.
(3)
Priority-Aware Resource Competition Solution: An iterative matching cost adjustment mechanism ensures that high-priority tasks are given priority for resource allocation, thus minimizing latency for latency-sensitive applications.
These results highlight the effectiveness of the framework in balancing computational requirements and network latency, addressing the central challenge facing traditional approaches that ignore dynamic transmission delays.
Although the proposed method shows good performance in simulation, its practical application still faces the following challenges. First, when an edge node fails suddenly (e.g., power interruption of industrial sensors or malfunctions in computation units), the existing architecture lacks real-time fault sensing and dynamic rescheduling mechanisms, and tasks relying on the node face the risk of timeout due to delayed migration to redundant nodes, potentially increasing task failure rates. Second, the movement of the equipment (such as AGVs, inspection drones) triggers channel fading, and the algorithm’s delay prediction model fails to track channel parameter changes in real time, potentially increasing task assignment error rates.
To further validate the algorithm’s universality in dynamic abnormal scenarios, the following experiments will be carried out in the future research. First, we will inject high-frequency random node failure events and sudden traffic congestion models into the NS-3 simulation environment, simulate extreme scenarios such as sensor node cluster failures in smart factories or sudden reductions in the bandwidth of industrial buses, and test the algorithm’s failure-aware delay and task rescheduling success rate. Second, we will develop a “mobile trajectory prediction” model to address the challenges of equipment mobility, and develop a “mobile trajectory prediction” model to address the challenges of equipment mobility. Further, to address the challenges of device mobility, we develop a joint mechanism of “mobile trajectory prediction and task pre-migration” to train LSTM prediction models with historical mobile data. The above experiments will systematically verify the robustness of the algorithm in dynamic disruption and mobile scenarios, and provide theoretical support for industrial-scale deployment.

6. Conclusions

In this paper, we present a novel cloud–edge collaborative scheduling method that explicitly accounts for network transmission delay, a critical factor often overlooked in task execution. Traditional methods neglect the influence of network transmission delay on task execution efficiency. To bridge this gap, we propose a distributed cloud–edge collaborative scheduling method that explicitly considers network transmission delay to optimize task allocation and execution order. Experiments show that this method can significantly boost task execution efficiency and enhance the overall performance of the system. This work offers a novel optimization solution for task scheduling in distributed cloud–edge systems, significantly contributing to the promotion of green computing and advancing sustainable development objectives. Future research will aim to develop scheduling strategies that incorporate dynamic network disruptions (e.g., node failures, traffic congestion) and equipment mobility. By integrating probabilistic models or adaptive learning mechanisms, such as reinforcement learning, these strategies aim to enhance the algorithm’s resilience in uncertain environments, ensuring stable performance under volatile conditions. Such advancements are expected to further propel the field toward more efficient, robust, and sustainable distributed computing systems, aligning with the growing demands for reliability and adaptability in emerging networked applications.

Author Contributions

Conceptualization and methodology, W.Z. and C.W.; writing, W.X.; supervision, Y.L.; data curation and validation, W.X. and G.S.; investigation and visualization, W.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number (62173082, U22B20115), and the National Key Research and Development Program of China, grant number (2022YFB4100802).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author(s).

Acknowledgments

The authors would like to appreciate the editors and the anonymous reviewers for their insightful suggestions to improve the quality of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The objective of the optimization problem (Equation (17)) is to minimize the weighted sum of average processing delay and data queue backlog. Due to the inclusion of integer variables in the constraint conditions, it is difficult to solve directly. Therefore, introducing Lagrange multipliers λ j ( t ) , μ i ( t ) , v k ( t ) 0 , a Lagrange function is constructed:
L = t = 1 T i = 1 l V Q i t + j = 1 J k = 1 K x i , j ( t ) y i , k ( t ) T i , j , k , s t     + t = 1 T j = 1 J λ j t ( i = 1 I x i , j ( t ) 1 )     + t = 1 T k = 1 K μ k t ( i = 1 I y i , k ( t ) q k )
The dual problem involves max λ , μ , v and min L . By solving the dual problem, the original problem is transformed into a state of unconstrained optimization.
Decompose the global problem into terminal-level sub-problems. The local optimization objective for each terminal d i is as follows:
min x i , j ( t ) , y i , k ( t ) t V Q i t + j , k x i , j ( t ) y i , k ( t ) T i , j , k , s t                               + t , j λ i t x i , j ( t ) + t , k v k t y i , k ( t )
The terminal selects the resource block server combination ( f j , c k ) for maximizing the preference value, and the preference value is defined in Equation (18).
Next, we will prove the convergence of the algorithm.
Theorem A1 (Convergence). 
The algorithm converges to a stable match after a finite number of iterations.
According to Equations (18) and (19), the preference values satisfy the following:
θ i , j , k ( t + 1 ) = θ i , j , k ( t ) Δ p i , j + Δ p i , k κ i < θ i , j , k ( t )
The preference value strictly decreases over time.
In each iteration, if terminal d i switches to a better match ( f j , c k ) , the total cost will be reduced by at least the following value:
Δ C ε = min ( j , k ) ( θ i , j , k ( t ) θ i , j , k ( t ) ) > 0
The number of resource blocks and server combinations is limited (set as J × K ), so there is a lower bound C min for the total cost. The algorithm must terminate after a finite number of iterations T 0 , satisfying the following criterion:
C ( T 0 ) C ( 0 ) T 0 ε
Definition A1 (Nash equilibrium condition). 
When the preference values of all terminals are stable, they satisfy the following:
θ i , j , k ( t ) θ i , j , k ( t ) , ( j , k ) ( j , k ) , i D
The current match of each terminal is the local optimal selection. When the preference values of all terminals no longer change, it satisfies the following:
θ i , j , k ( t ) = θ i , j , k ( t ) , ( j , k ) ( j , k ) , i D
According to Equation (19), at this point, Δ p i , j = Δ p i , k = 0 , which means there is no resource competition conflict. Combining Equations (A3)–(A5), the algorithm must converge to a stable state while satisfying monotonicity, boundedness, and limited resources.
Theorem A2 (Queue Stability). 
If the system satisfies  E [ A i ( t ) ] E [ U i ( t ) ] , then the queue backlog  Q i ( t )  is bounded on average and satisfies the following:
lim T 1 T t = 1 T E [ Q i ( t ) ] B V
where  B  is a constant and  V  is a weight parameter that balances delay and queue backlog.
Define the Lyapunov function as the sum of squares of the queue backlog:
L ( t ) = 1 2 i = 1 I Q i 2 ( t )
Define the single time slot Lyapunov drift as follows:
Δ ( t ) = E [ L ( t + 1 ) L ( t ) | Q ( t ) ]
By substituting the queue update Equation (6) and expanding it, we can obtain the following:
Δ ( t ) i = 1 I { 1 2 E ( A i ( t ) U i ( t ) ) 2 + Q i ( t ) E [ A i ( t ) U i ( t ) ] }
We assuming that the data arrival and transmission volume meet the following conditions:
E [ A i 2 ( t ) + U i 2 ( t ) ] B
Combined with the V Q i ( t ) term in the optimization objective, the total drift and penalty term meet the following requirements:
Δ ( t ) + V E [ Q i ( t ) ] B + V E [ Q i ( t ) ] i = 1 I Q i ( t ) E [ U i ( t ) A i ( t ) ]
From E [ A i ( t ) ] E [ U i ( t ) ] , we can obtain the following:
i = 1 I Q i ( t ) E [ U i ( t ) A i ( t ) ] 0
V E [ Q i ( t ) ] B lim T 1 T t = 1 T E [ Q i ( t ) ] B V
Through Lyapunov drift analysis, the average boundedness of the queue backlog is proved, and the upper bound is inversely proportional to the trade-off parameter V.
Theorem A3 (delay optimality). 
Let the total cost of the global optimal solution be  C  and the total cost of the algorithm solution be  C alg , then there is a constant  B  such that the following prevails:
C alg C + B V
where  V  is the weight parameter to weigh the delay and queue backlog.
Defining the total cost as Equation (17), combined with Theorem A2, the total contribution of queue backlog terms can be obtained as follows:
1 T t = 1 T V Q i ( t ) V B V = B
Let the delay term of the global optimal solution be as follows:
D = 1 T t = 1 T j , k x i , j ( t ) y i , k ( t ) T i , j , k ( t )
The Equation (A17) satisfies:
D D alg + ε
where ε is the delay difference between the algorithm and the optimal solution.
According to Lyapunov drift theory, the total delay term satisfies the following:
1 T t = 1 T j , k x i , j ( t ) y i , k ( t ) T i , j , k ( t ) D + B V
Combining the results of queue backlog items, the total cost difference is shown in Equation (A17). The average value of the queue backlog is limited to O ( 1 / V ) , and increasing V can further reduce the backlog. The difference between the delay term of the algorithm and the global optimal solution is O ( 1 / V ) , indicating that the algorithm has asymptotic optimality in delay optimization. By adjusting V , the queue backlog and delay optimization can be balanced.

References

  1. Zheng, L.; Chen, J.; Liu, T.; Liu, B.; Yuan, J.; Zhang, G. A Cloud-Edge Collaboration Framework for Power Internet of Things Based on 5G networks. In Proceedings of the 2021 IEEE 9th International Conference on Information, Communication and Networks (ICICN), Xi’an, China, 25–28 November 2021; pp. 273–277. [Google Scholar]
  2. Mohiuddin, K.; Fatima, H.; Khan, M.A.; Khaleel, M.A.; Nasr, O.A.; Shahwar, S. Mobile learning evolution and emerging computing paradigms: An edge-based cloud architecture for reduced latencies and quick response time. Array 2022, 16, 100259. [Google Scholar] [CrossRef]
  3. Liu, G.; Dai, F.; Xu, X.; Fu, X.; Dou, W.; Kumar, N.; Bilal, M. An adaptive DNN inference acceleration framework with end–edge–cloud collaborative computing. Future Gener. Comput. Syst. 2023, 140, 422–435. [Google Scholar] [CrossRef]
  4. Zhou, J.; Kondo, M. An Edge-Cloud Collaboration Framework for Graph Processing in Smart Society. IEEE Trans. Emerg. Top. Comput. 2023, 11, 985–1001. [Google Scholar] [CrossRef]
  5. Zeng, F.; Zhang, K.; Wu, L.; Wu, J. Efficient Caching in Vehicular Edge Computing Based on Edge-Cloud Collaboration. IEEE Trans. Veh. Technol. 2023, 72, 2468–2481. [Google Scholar] [CrossRef]
  6. Zhou, H.; Wang, K.; Fei, Z.; Luo, H. The Technical Architecture and Mechanism of Power Grid Mobile Security Inspection Cloud-Edge Collaboration Based on the 5G Network. In Proceedings of the 2023 3rd International Conference on New Energy and Power Engineering (ICNEPE), Huzhou, China, 24–26 November 2023; pp. 1036–1040. [Google Scholar]
  7. Ren, C.; Gong, C.; Liu, L. Task-Oriented Multimodal Communication Based on Cloud–Edge–UAV Collaboration. IEEE Internet Things J. 2024, 11, 125–136. [Google Scholar] [CrossRef]
  8. Deng, X.; Li, J.; Liu, E.; Zhang, H. Task allocation algorithm and optimization model on edge collaboration. J. Syst. Archit. 2020, 110, 101778. [Google Scholar] [CrossRef]
  9. Liu, G.; Song, X.; Xin, C.; Liang, T.; Li, Y.; Liu, K. Edge–Cloud Collaborative Optimization Scheduling of an Industrial Park Integrated Energy System. Sustainability 2024, 16, 1908. [Google Scholar] [CrossRef]
  10. Reyana, A.; Kautish, S.; Alnowibet, K.A.; Zawbaa, H.M.; Wagdy Mohamed, A. Opportunities of IoT in Fog Computing for High Fault Tolerance and Sustainable Energy Optimization. Sustainability 2023, 15, 8702. [Google Scholar] [CrossRef]
  11. Zhang, Z.; Qu, T.; Zhao, K.; Zhang, K.; Zhang, Y.; Liu, L.; Wang, J.; Huang, G.Q. Optimization Model and Strategy for Dynamic Material Distribution Scheduling Based on Digital Twin: A Step towards Sustainable Manufacturing. Sustainability 2023, 15, 16539. [Google Scholar] [CrossRef]
  12. Shao, S.; Su, L.; Zhang, Q.; Wu, S.; Guo, S.; Qi, F. Multi task dynamic edge–end computing collaboration for urban Internet of Vehicles. Comput. Netw. 2023, 227, 109690. [Google Scholar] [CrossRef]
  13. Ma, X.; Zhao, J.; Jia, L.; Chen, X.; Li, Z. Optimal edge-cloud collaboration based strategies for minimizing valid latency of railway environment monitoring system. High-Speed Railw. 2023, 1, 185–194. [Google Scholar] [CrossRef]
  14. Zhu, Q.; Yu, X.; Zhao, Y.; Nag, A.; Zhang, J. Resource Allocation in Quantum-Key-Distribution-Secured Datacenter Networks With Cloud–Edge Collaboration. IEEE Internet Things J. 2023, 10, 10916–10932. [Google Scholar] [CrossRef]
  15. Ou, Q.; Wang, Y.; Song, W.; Zhang, N.; Zhang, J.; Liu, H. Research on network performance optimization technology based on cloud-edge collaborative architecture. In Proceedings of the 2021 IEEE 2nd International Conference on Big Data, Artificial Intelligence and Internet of Things Engineering (ICBAIE), Nanchang, China, 26–28 March 2021; pp. 274–278. [Google Scholar]
  16. Karimi, E.; Chen, Y.; Akbari, B. Task offloading in vehicular edge computing networks via deep reinforcement learning. Comput. Commun. 2022, 189, 193–204. [Google Scholar] [CrossRef]
  17. Li, Q.; Guo, M.; Peng, Z.; Cui, D.; He, J. Edge–Cloud Collaborative Computation Offloading for Mixed Traffic. IEEE Syst. J. 2023, 17, 5023–5034. [Google Scholar] [CrossRef]
  18. Tang, T.; Li, C.; Liu, F. Collaborative cloud-edge-end task offloading with task dependency based on deep reinforcement learning. Comput. Commun. 2023, 209, 78–90. [Google Scholar] [CrossRef]
  19. Wang, Q.; Chen, S.; Wu, M. Incentive-Aware Blockchain-Assisted Intelligent Edge Caching and Computation Offloading for IoT. Engineering 2023, 31, 127–138. [Google Scholar] [CrossRef]
  20. Chen, Z.; Ren, Y.; Sun, Q.; Li, N.; Han, Y.; Huang, Y.; Li, H.; I, C.-L. Computing Task Orchestration in Communication and Computing Integrated Mobile Network. In Proceedings of the 2023 IEEE Globecom Workshops (GC Wkshps), Kuala Lumpur, Malaysia, 4–8 December 2023; pp. 2073–2078. [Google Scholar]
  21. Hussain, M.; Wei, L.F.; Lakhan, A.; Wali, S.; Ali, S.; Hussain, A. Energy and Performance-Efficient Task Scheduling in Heterogeneous Virtualized Cloud Computing. Sustain. Comput. Inform. Syst. 2021, 30, 100517. [Google Scholar] [CrossRef]
  22. Hussain, M.; Wei, L.-F.; Abbas, F.; Rehman, A.; Ali, M.; Lakhan, A. A Multi-Objective Quantum-Inspired Genetic Algorithm for Workflow Healthcare Application Scheduling With Hard and Soft Deadline Constraints in Hybrid Clouds. Appl. Soft Comput. 2022, 128, 109440. [Google Scholar] [CrossRef]
  23. Bao, B.; Yang, H.; Yao, Q.; Guan, L.; Zhang, J.; Cheriet, M. Resource Allocation With Edge-Cloud Collaborative Traffic Prediction in Integrated Radio and Optical Networks. IEEE Access 2023, 11, 7067–7077. [Google Scholar] [CrossRef]
  24. Liu, S.; Qiao, B.; Han, D.; Wu, G. Task offloading method based on CNN-LSTM-attention for cloud–edge–end collaboration system. Internet Things 2024, 26, 101204. [Google Scholar] [CrossRef]
  25. Kim, T.; Sathyanarayana, S.D.; Chen, S.; Im, Y.; Zhang, X.; Ha, S.; Joe-Wong, C. MoDEMS: Optimizing Edge Computing Migrations for User Mobility. IEEE J. Sel. Areas Commun. 2023, 41, 675–689. [Google Scholar] [CrossRef]
  26. Pang, S.; Hou, L.; Gui, H.; He, X.; Wang, T.; Zhao, Y. Multi-mobile vehicles task offloading for vehicle-edge-cloud collaboration: A dependency-aware and deep reinforcement learning approach. Comput. Commun. 2024, 213, 359–371. [Google Scholar] [CrossRef]
  27. Hussain, M.; Wei, L.-F.; Rehman, A.; Abbas, F.; Hussain, A.; Ali, M. Deadline-Constrained Energy-Aware Workflow Scheduling in Geographically Distributed Cloud Data Centers. Future Gener. Comput. Syst. 2022, 132, 211–222. [Google Scholar] [CrossRef]
  28. Hussain, M.; Wei, L.-F.; Rehman, A.; Hussain, A.; Ali, M.; Javed, M.H. An Electricity Price and Energy-Efficient Workflow Scheduling in Geographically Distributed Cloud Data Centers. J. King Saud Univ.-Comput. Inf. Sci. 2024, 36, 102170. [Google Scholar] [CrossRef]
  29. Jiang, S.; Gao, H.; Wang, X.; Liu, J.; Zuo, K. Deep reinforcement learning based multi-level dynamic reconfiguration for urban distribution network: A cloud-edge collaboration architecture. Glob. Energy Interconnect. 2023, 6, 1–14. [Google Scholar] [CrossRef]
  30. Li, X.; Wang, Z. A Cloud-Edge Computing Method for Integrated Electricity-Gas System Dispatch. Processes 2023, 11, 2299. [Google Scholar] [CrossRef]
Figure 1. Cloud-side network architecture diagram.
Figure 1. Cloud-side network architecture diagram.
Sustainability 17 04887 g001
Figure 2. Cloud–edge network architecture diagram.
Figure 2. Cloud–edge network architecture diagram.
Sustainability 17 04887 g002
Figure 3. Delay queue-aware dynamic matching algorithm process.
Figure 3. Delay queue-aware dynamic matching algorithm process.
Sustainability 17 04887 g003
Figure 4. Edge–cloud collaborative system.
Figure 4. Edge–cloud collaborative system.
Sustainability 17 04887 g004
Figure 5. Architecture model visualization.
Figure 5. Architecture model visualization.
Sustainability 17 04887 g005
Figure 6. Algorithm performance comparison.
Figure 6. Algorithm performance comparison.
Sustainability 17 04887 g006
Figure 7. Data processing latency is weighted and compared with data queue backlog.
Figure 7. Data processing latency is weighted and compared with data queue backlog.
Sustainability 17 04887 g007
Figure 8. Comparison of data processing delay.
Figure 8. Comparison of data processing delay.
Sustainability 17 04887 g008
Figure 9. Data queue backlog comparison.
Figure 9. Data queue backlog comparison.
Sustainability 17 04887 g009
Figure 10. Weighting functions and comparisons as the number of terminals increases.
Figure 10. Weighting functions and comparisons as the number of terminals increases.
Sustainability 17 04887 g010
Figure 11. Weighting functions and comparisons as edge servers increases.
Figure 11. Weighting functions and comparisons as edge servers increases.
Sustainability 17 04887 g011
Figure 12. The comparison of latency performance between Path 1 and Path 2.
Figure 12. The comparison of latency performance between Path 1 and Path 2.
Sustainability 17 04887 g012
Table 1. Relevant symbol.
Table 1. Relevant symbol.
NotationDefinition
γ i , j , k t Signal-to-noise ratio (SNR) for terminal i to server k via resource block j in iteration t
P i ( t ) Data transmission power of d i
g i , j , k ( t ) d i Channel gain used f j to transmit data to c k
δ 2 Transmission channel noise power
x i , j ( t ) 0-1 variable, whether time slot t offloads tasks from terminal f j to server d i
y i , k ( t ) 0-1 variable, whether time slot t offloads tasks from terminal d i to server c k
R i , j , k ( t ) The transmission rate for d i transmitting data c k using f j
B i , j , k Channel bandwidth for transmitting data
T i , j , k , t r ( t ) The channel bandwidth for d i transmitting data c k via f j
T i , k , c ( t ) The latency c k to process the collected data from d i
T i , k , w ( t ) The latency waiting for c k it to complete its computational load.
l j ( t ) The computational resources provided by the resource block f j
α i The computational resources required to process a unit of d i data.
C j Server computing resources
D i , k Volume of data for the mission
u i , j A binary variable where u i , j = 1 denotes task offloading to the server and u i , j = 0 denotes local execution
z i , k Storage resources required for storage tasks
θ ν 2 E End-to-edge server transmission rate
θ i , j , k ( t ) Terminal d i preference values for resource block server combination options s j . k
Δ p i , j Step of matching cost enhancement between d i and resource block f j
Δ p i , k d i with server c k in matching cost boosting step and
κ i Priority weighting of d i
Table 2. Initial IP addresses.
Table 2. Initial IP addresses.
Device NameIPv4 AddressIPv6 AddressMAC Address
Cloud server10.1.1.100/242001:db8::10000:00:00:00:00:01
Switch 110.1.1.1/242001:db8::100:00:00:00:00:02
Switch 210.1.1.2/242001:db8::200:00:00:00:00:03
Edge device 110.1.1.10/242001:db8::1000:00:00:00:00:04
Edge device 110.1.1.11/242001:db8::1100:00:00:00:00:05
End 110.1.2.20/242001:db8::2000:00:00:00:00:06
End 210.1.2.21/242001:db8::2100:00:00:00:00:07
End 310.1.2.22/242001:db8::2200:00:00:00:00:08
Table 3. Initial value of the parameter.
Table 3. Initial value of the parameter.
ParameterValues
Simulation run time /s10
Application communication period /s(1, 10)
Application packet size512B
Application packet sending rate100 kb/s
The amount of data sent by the application118,260 B
Maximum transmission rateCable (100 Mbps), Fiber (1 Gbps), wireless (500 Mbps)
Channel delay2 µs
End movement velocity1 m/s
Table 4. Initial coordinates of each node (m).
Table 4. Initial coordinates of each node (m).
0 (cloud)(50.0, 0.0, 0.0)
1 (switch 1)(50.0, 25.0, 0.0)
2 (switch 2)(25.0, 50.0, 0.0)
3 (switch 3)(75.0, 50.0, 0.0)
4 (side 1)(0, 100.0, 0.0)
5 (side 2)(50.0, 100.0, 0.0)
6 (side 3)(100, 100.0, 0.0)
7 (end 1)(0.0, 75.0, 0.0)
8 (end 2)(100.0, 75.0, 0.0)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zheng, W.; Wang, C.; Xu, W.; Sun, G.; Luo, Y. A New Delay-Aware Distributed Cloud–Edge Scheduling Framework and Algorithm in Dynamic Network Environments. Sustainability 2025, 17, 4887. https://doi.org/10.3390/su17114887

AMA Style

Zheng W, Wang C, Xu W, Sun G, Luo Y. A New Delay-Aware Distributed Cloud–Edge Scheduling Framework and Algorithm in Dynamic Network Environments. Sustainability. 2025; 17(11):4887. https://doi.org/10.3390/su17114887

Chicago/Turabian Style

Zheng, Wei, Chenyang Wang, Wentao Xu, Guoxiang Sun, and Yanhong Luo. 2025. "A New Delay-Aware Distributed Cloud–Edge Scheduling Framework and Algorithm in Dynamic Network Environments" Sustainability 17, no. 11: 4887. https://doi.org/10.3390/su17114887

APA Style

Zheng, W., Wang, C., Xu, W., Sun, G., & Luo, Y. (2025). A New Delay-Aware Distributed Cloud–Edge Scheduling Framework and Algorithm in Dynamic Network Environments. Sustainability, 17(11), 4887. https://doi.org/10.3390/su17114887

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop