Next Article in Journal
A Review of Operation Optimization Objectives and Evaluation Methods for Park-Level Integrated Energy System with Mobile Robots
Previous Article in Journal
Development of Magnetic Hysteresis Loop Measurement System for Characterization of 3D-Printed Magnetic Cores
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quantitative Analysis and Verification of Edge Computing Offloading Strategy Based on Probabilistic Model Checking

1
School of Software, Northwestern Polytechnical University, Xi’an 710129, China
2
Yangtze River Delta Research Institute of NPU, Taicang 215400, China
3
School of Public Health, Shanghai Jiao Tong University School of Medicine, Shanghai 200025, China
4
Department of Computer Science, Wenzhou-Kean University, Wenzhou 325060, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2025, 14(11), 2236; https://doi.org/10.3390/electronics14112236
Submission received: 1 April 2025 / Revised: 17 May 2025 / Accepted: 28 May 2025 / Published: 30 May 2025
(This article belongs to the Special Issue Recent Advancements of High Performance Computing (HPC))

Abstract

:
Edge computing has become a key framework for meeting the demands of low latency and high reliability across various applications. To ensure efficient task execution in edge computing, many offloading strategies have been proposed; however, most lack quantitative analysis and formal verification to guarantee their correctness. This paper addresses this gap by presenting an integrated offloading strategy that combines delay-based, energy-efficient, and energy-delay trade-off approaches, enhancing both the clarity and verifiability of offloading methods. Furthermore, we apply probabilistic model checking using PRISM to rigorously analyze and validate the correctness of the proposed hybrid strategy. Our approach provides practical solutions and valuable insights, promoting the development of reliable and efficient offloading strategies for edge computing systems.

1. Introduction

Edge computing is an emerging computing paradigm aimed at fulfilling the demanding ultra-low latency needs of contemporary users. At the heart of this paradigm lies the offloading strategy, which plays a crucial role in edge computing systems. These offloading strategies are primarily classified into three categories: delay-based, energy-efficient, and energy-delay trade-off. Their purpose is to tackle the challenges posed by resource constraints in mobile devices and excessive energy consumption.
In the field of edge computing, the development of offloading strategies is pivotal in achieving low-latency and high-performance computing. As edge devices become increasingly heterogeneous and the network conditions vary, there is a growing demand for efficient decision-making processes to determine which computational tasks should be offloaded and where they should be executed. However, the complexity of this decision-making process, coupled with the inherent uncertainties in edge environments, underscores the need for a quantitative approach. In addition, the majority of current research efforts primarily focus on proposing the three aforementioned offloading strategies for different user scenarios. These strategies involve resource allocation algorithms and exploration algorithms to evaluate their impact on latency improvement or energy consumption reduction using various datasets. Although these existing schemes are clear and useful, they may not provide a complete picture. Quantitative analysis can complement this by offering empirical evidence and deeper insights into the system’s behavior. Furthermore, there is a notable lack of analysis regarding the correctness of the offloading strategy designs themselves. Fortunately, the application of quantitative analysis and verification through formal methods offers a means to assess the correctness of these strategies. Henceforth, in order to address these challenges, this research draws motivation from the recognition that probabilistic model checking provides a systematic means to analyze and validate edge computing offloading strategies [1]. In light of this, this paper introduces a verification method specifically designed for offloading strategies. The method abstracts the fundamental properties that an offloading strategy should adhere to, and employs probabilistic model checking for quantitative analysis and verification. By leveraging probabilistic models, we evaluate the correctness of edge computing offloading strategies, addressing uncertainties and risks in the design process. Our findings provide actionable guidance for designing and verifying offloading strategies in edge computing.
The key component of edge computing systems is the development of efficient and accurate offloading strategies. Researchers have proposed various offloading strategies tailored to different scenarios, such as single-user or multi-user, focusing on either delay reduction (delay-based offloading), energy consumption optimization (energy-efficient offloading), or a balance between the two (energy-delay trade-off). These strategies are typically evaluated through data comparison and performance analysis to assess their effectiveness and improvements in specific or multiple scenarios. However, a crucial aspect often overlooked in most research is the analysis and verification of the correctness of the designed strategies using mathematical logic. Most of the existing work revolves around proposing new strategies without conducting rigorous analysis. Therefore, it is essential to verify whether the designed offloading strategy is correct and aligns with the desired objectives of the designer. This verification serves as a prerequisite for performance analysis and further enhancements of offloading strategies.
An effective approach to designing a verification model for offloading strategies is to encompass the considerations of delay, energy consumption, and the balance between the two. This can be achieved by configuring the transmission cost parameters appropriately, allowing for quantitative analysis. In this paper, the investigation of correctness primarily focuses on the probabilistic nature of potential failures or issues within the strategy itself. Through comprehensive evaluation, quantitative analysis, and the application of the formal verification method, the success rate of the strategy can be determined from a mathematical statistical perspective. By utilizing probabilistic modeling techniques, the correctness of newly proposed strategies by researchers can be assessed effectively.
The main contributions of this paper are as follows:
  • Proposing an integrated offloading strategy covering delay-based, energy-efficient, and energy-delay trade-off approaches to improve strategy clarity and verifiability.
  • Modeling the integrated strategy probabilistically and verify its correctness via the PRISM model checker, demonstrating effectiveness.
  • Offering technical guidance and insights for correctness verification and evaluation, advancing theoretical foundations for trustworthy edge computing offloading validation.
The structure of the rest of the paper is as follows. Section 2 provides an introduction to probabilistic model checking, along with an overview of the automated verification tools utilized in this paper. Additionally, it outlines the fundamental concepts and components of the offloading strategy. Section 3 focuses on modeling the offloading strategy, presenting a detailed framework that captures their essential characteristics and behaviors. Section 4 abstracts the key properties of offloading strategies and conducts quantitative analysis and verification using the methods of probabilistic model checking, aiming to assess the correctness of the strategies based on the established model framework. Section 5 then gives the related work, and Section 6 concludes the paper and gives the future work.

2. Background

In this section, we present the basic principles of the offloading strategy and provide an explanation of the probabilistic model checking approach employed in this paper.

2.1. Offloading Strategies

In this paper, edge computing is a distributed paradigm that processes data closer to its source to reduce latency and network congestion. The “edge” includes different levels such as local devices, nearby edge servers, and remote cloud servers. Accordingly, we categorize computation into three levels based on task location, including local (on-device processing), edge server (offloading to nearby nodes with moderate resources and low latency), and cloud server (offloading to remote data centers with high computing power but higher latency). These levels form the basis of our system model for analyzing delay, energy consumption, and offloading decisions. Figure 1 shows the workflow of offloading strategies in edge computing systems [2,3,4], where three different kinds of offloading strategies are outlined as summarized in Table 1.
(1) Delay-based Offloading Strategy: This strategy focuses on minimizing the overall execution time delay. If the local execution delay T l is lower than the offloading execution delay T o , which can be either the time delay in edge servers T m or the time delay in the cloud server T c , the computation is performed locally. Conversely, if the offloading execution delay T o is lower, the computation is offloaded to either the edge servers or the cloud for faster execution.
(2) Energy-efficient Offloading Strategy: In this strategy, the primary consideration is minimizing energy consumption. If the energy consumption of local execution E l is lower than that of offloading execution E o , where E o can be either the energy consumption in edge servers E m or the cloud server E c , the computation is executed locally. Otherwise, if the energy consumption of offloading execution E o is lower, the computation is offloaded to either the edge servers or the cloud for more energy-efficient execution.
(3) Energy-delay Trade-off Offloading Strategy: This strategy dynamically adapts based on user requirements or specific scenarios, taking into account the trade-off between delay and energy consumption. The strategy evaluates the overall cost by considering both delay and energy consumption, represented as C = α T + ( 1 α ) E [5]. Note that when combining metrics with different units, such as energy consumption E (measured in, for example, Joules or Watts) and latency T (measured in milliseconds or seconds), it is important to ensure that the composite metric remains meaningful and interpretable. Hence, in the above formula, C is the composite metric that represents the combination of energy consumption E and latency T; α and 1 α are weighting factors, denoting the relative importance of energy consumption and latency in the composite metric. Here, energy consumption refers to the energy used, measured in a standardized unit (Joules), to ensure unit consistency, while latency represents the time taken for data transmission or task execution, also measured in a standardized unit (seconds) to maintain unit consistency.
In brief, to address the issue of different units for energy consumption and latency, we adopt the standardized units for both metrics in the composite formula. This standardization ensures that the units are consistent, allowing for a meaningful combination of these two distinct metrics. The weighting factors, α and 1 α , are used to emphasize the relative importance of each metric in the composite metric, providing a clear and interpretable representation of the trade-off between energy consumption and latency. By standardizing the units and clearly defining the weighting factors, the composite metric C can effectively capture the trade-off between energy consumption and latency while maintaining unit consistency. As a result, this approach enables a more intuitive and interpretable assessment of system correctness. Furthermore, users can calculate the total cost to determine the most suitable offloading strategy for a given situation. Specifically, the weight coefficient α allows users to prioritize either minimizing delay or minimizing energy consumption, thereby striking a balance between the two objectives.

2.2. Probabilistic Model Checking

Probabilistic model checking is a formal verification technique that uses probabilistic models and statistical methods to rigorously analyze system behavior, performance, and reliability under uncertainty. By mathematically modeling state transitions and probabilistic events (e.g., data transmission delays or resource availability), it enables quantitative assessment of properties like correctness, security, and efficiency, helping identify design flaws early. This approach is particularly valuable in edge computing networks—such as IoT, smart grids, and vehicular systems—where real-time, safety-critical operations demand reliable evaluation of stochastic events (e.g., data loss or task delays). Combining quantitative analysis with probabilistic modeling ensures robust system design and risk mitigation in dynamic, uncertain environments.
In this paper, we leverage probabilistic model checking with PRISM [6,7] to quantitatively verify the correctness of offloading decisions in edge computing. PRISM supports multiple probabilistic models, including DTMC (discrete-time Markov chains with probabilistic transitions), MDP (extending DTMC with decision making), and CTMC (continuous-time variant with dense time modeling). Here, we give one of the probability formulas used in this paper as an example:
P = ? [ F T ϕ ]
This computes the probability that a given Linear Temporal Logic (LTL) property ϕ holds eventually (denoted by F) within T time steps. Here, LTL is a formal logic for specifying and verifying temporal properties of system behaviors over future states.

3. Modeling the Integrated Offloading Strategy

In this section, we present the stateflow of the integrated offloading strategy, illustrate its automaton, and formalize the model in PRISM.

3.1. The Stateflow of the Integrated Offloading Strategy

To enable quantitative analysis, we model the stateflow of three integrated offloading strategies (Figure 2), whose stages are as follows.
  • Data integrity verification. Upon receiving a task submission, the strategy first verifies the success of the submission. If confirmed, it proceeds to the strategy selection phase. Otherwise, the system initiates a resubmission process. If the task fails after the maximum attempts, the offloading process is terminated.
  • Strategy selection. The choice of offloading strategy depends on the coefficient α in the transmission cost. The two branches illustrated in Figure 2 correspond to distinct strategies based on the value of α . Specifically, the left branch represents the delay-based and energy-efficient strategies when α is either 0 or 1, whereas the right branch corresponds to the energy-delay trade-off strategy applied when α takes any value between 0 and 1.
Detailed explanations for the left branch in Figure 2, which corresponds to the delay-based and energy-efficient offloading strategies outlined in Table 1, are as follows. When decisions are made based on either energy consumption or latency—that is, for the energy-efficient or delay-based offloading strategies—the key step is to determine the minimum transmission cost. If the minimum cost, m i n c o s t , matches the local execution cost C l o c a l shown in Table 2, the task is executed locally and then completed. If  m i n c o s t corresponds to a remote option, the task is offloaded to either an edge server or a cloud server. At this point, it is necessary to check whether sufficient resources are available to process the task. If resources are adequate, data processing begins and the system continuously verifies whether the task has finished. If the task remains incomplete, processing continues until completion. However, if there are not enough resources, the task enters a waiting stage where resource availability is monitored. Once resources become sufficient, data processing resumes. If the waiting time exceeds the offloading latency benefit, the system switches promptly to local execution; otherwise, it continues to wait for resources.
Conversely, the right branch in Figure 2 corresponds to the energy-delay trade-off offloading strategy described in Table 1. This strategy balances energy consumption and delay based on the user’s choice. If the user opts for local execution, the strategy provides a smooth line chart illustrating the probability of success or failure depending on the selected coefficient α . If offloading is chosen, a decision-making process similar to that used in the delay-based or energy-efficient strategies is applied, alongside a corresponding smooth probability chart related to the choice of α . These charts are comparable to those presented in Figures 4–6 in Section 4.
To complement the stateflow diagram in Figure 2, Table 2 summarizes the key parameters adopted in our offloading strategy model. The core formulations are presented below [5]:
The local computation time is given by:
t local = m s local
The edge server computation time comprises both transmission and processing components:
t mec = b B + m s mec
The offloading condition for latency optimization is:
m · 1 s local 1 s mec > b B
Energy consumption models are defined as:
E local = P local · t local
E mec = P transmit · b B + P connect · m s mec
The composite cost functions for each computing mode are:
C local = α · t local + ( 1 α ) · E local
C mec = α · t mec + ( 1 α ) · E mec
C cloud = α · t cloud + ( 1 α ) · E cloud
The optimal strategy minimizes the cost:
C min = min ( C local , C mec , C cloud )
It is worth noting that the use of deterministic computation time or some other assumptions above are aimed at simplifying the modeling and analysis process to gain a clearer understanding of the fundamental workings of the edge computing strategy in this study. For instance, although computation time in real systems may often be stochastic or probabilistic, modeling them deterministically allows for initial assessments and performance analyses. This modeling approach contributes to an intuitive understanding.

3.2. The Automaton of the Offloading Strategies

In this subsection, we present the corresponding automaton in Figure 3 based on the stateflow discussed earlier in Figure 2. Figure 3 illustrates the automaton, where the cumulative sum of probability values between the subsequent nodes and the current node in each state is equal to 1. For instance, the state s 11 in the automaton represents a probability of p 11 entering the state s 21 of the energy-efficient or delay-based offloading strategy after the successful checking of data, which corresponds to α = 0 or 1, and a probability of p 13 entering the state s 31 of energy-delay trade-off, whose coefficient α belongs to ( 0 ,   1 ) . After passing through the probability branches mentioned above, if it is in the subsequent state of state s 21 , the automaton continues to provide different probability branch scenarios. Specifically, there is a probability of 1 p 21 transitioning to state s 23 , where local computing is performed directly, and a probability of p 21 transitioning to state s 22 , where data processing is offloaded to the edge or cloud server. To clarify, all conditional decisions depicted in Figure 2 can be converted into the probability branching model described earlier, while any repetitive structures can be translated into self-loops associated with states, such as s 12 , s 24 , and  s 25 .
The dashed blue boxes in Figure 3 represent the choices made based on resource adequacy when deciding whether to offload to the edge server or the cloud server, considering factors such as delay or energy consumption. Table 3 provides the names of the states and their corresponding explanations, as depicted in Figure 3. In Figure 3, to be specific, states s 22 and s 22 signify decisions related to offloading data to the edge server and the cloud server, respectively, balancing the trade-off between delay and energy consumption. The explanations for specific details within these two states closely resemble those for states s 22 and s 23 . The selection process for offloading decisions, which hinges on factors like delay and energy consumption, remains uncertain depending on different probabilities.
Furthermore, in order to provide a clearer representation of the transition probability relationships and information between various states in the edge computing offloading probability model depicted in Figure 3, we give matrix q, where the specific probability values are determined by us. In the transition probability matrix q of the probability model, the rows and columns correspond to the states of the edge computing offloading strategy probability model and have been labeled within matrix q. For instance, what is meant by the entry in the first row and second column is that the probability value of transitioning from state s 11 to state s 12 equals 1 p 11 p 13 . For brevity, we have omitted the relationships between states s 0 , s 31 , s 32 , s 33 , s 34 , and other states in matrix q, which are shown in Figure 3.
q =       s 11 s 12       s 21 s 22     s 23     s 24 s 25 s 26 s s u c   s f a i l s 11 s 12 s 21 s 22 s 23 s 24 s 25 s 26 s s u c s f a i l ( 0 1 p 11 p 13 p 11 0 0 0 0 0 0 0 1 p 12 p 14 p 14 0 0 0 0 0 0 0 p 12 0 0 0 p 21 1 p 21 0 0 0 0 0 0 0 0 0 0 p 22 1 p 22 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 p 25 0 1 p 25 0 0 0 0 0 0 1 p 23 p 24 p 23 p 24 0 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 )

3.3. The Model in PRISM

In this section, we present the formal specification of our dynamic offloading model using the PRISM probabilistic model checker. The model is implemented as a Continuous-Time Markov Chain (CTMC) that captures the key components and decision processes involved in edge computing offloading. The PRISM model consists of several key components as below.

3.3.1. Constants and Parameters

The model begins with the declaration of various constants that represent system parameters and environmental conditions, which include parameters such as s t e p (state transition count), m (resource required for a computation task), s _ l o c a l , s _ m e c , and  s _ c l o u d (execution speeds for local, edge server (mec), and cloud), B _ m e c and B _ c l o u d (edge and cloud network bandwidth), b (task size), and various power-related parameters. Listing 1 gives the details.

3.3.2. Time and Energy Calculations

The model includes formulas for calculating execution time and energy consumption for different offloading locations related to Formulas (1)–(9) in Section 3.1. These formulas define the time ( t _ l o c a l , t _ m e c , t _ c l o u d ) and energy consumption ( E _ l o c a l , E _ m e c , E _ c l o u d ) required for task execution in different offloading scenarios. The formulas for calculating the cost ( C _ l o c a l , C _ m e c , C _ c l o u d ), minimum cost ( m i n _ c o s t ), and maximum cost ( m a x _ c o s t ) are also defined. For brevity, here, we give several definitions in PRISM as an instance in Listing 2.
Listing 1: Main model constants and parameters.
Electronics 14 02236 i001
Listing 2: Part of time and energy calculations.
Electronics 14 02236 i002

3.3.3. State Variables and Transitions

The core of the model is defined in the DynamicOffload module, which contains the state variables and transition rules. It includes state variables like choice, s (representing different states), t a s k _ h a n d l i n g _ m e c , and  t a s k _ h a n d l i n g _ c l o u d (number of tasks being handled in edge and cloud server), t _ m and t _ c (time waiting for edge and cloud server allocation), s u b m i t _ t i m e s (number of task submissions), and various Boolean variables to track the success of different events. In addition, the model defines transitions between different states based on probabilistic choices and conditions. For example, it includes transitions for task submission, resubmission, local handling, offloading to the edge or cloud server, waiting for resource allocation, and returning results. Listing 3 shows core module and state transitions.
The proposed model formalizes a probabilistic state transition mechanism that governs the complete offloading lifecycle. The system evolves through distinct phases including task submission, potential resubmission, local/edge/cloud execution, resource allocation waiting periods, and result return procedures. Central to this framework are three core components: (1) strategy selection through optimization parameters α { 0 , 1 } for pure energy/delay minimization or α ( 0 , 1 ) for balanced trade-off, implemented via transition probabilities P = { p 11 , , p 32 } ; (2) cost-aware offloading decisions that select between local ( choose _ local ), edge ( offload _ mec ), or cloud ( offload _ cloud ) execution based on minimum computed costs { C local , C mec , C cloud } ; and (3) temporal control mechanisms incorporating maximum wait thresholds ( t mec max , t cloud max ) with fallback to local execution when exceeding t wait . State transitions are explicitly defined through success codes S + = { 9 local , 19 mec , 29 cloud } , failure state 11, and situational branching governed by situation ( x ) = I x { 0 , 1 } + 2 · I 0 < x < 1 where x = α . This integrated approach ensures adaptive decision making under dynamic environmental constraints while maintaining probabilistic guarantees throughout the offloading process.
Listing 3: Core module and state transitions.
Electronics 14 02236 i003

3.3.4. Transition Probabilities

The model uses several probability values to govern the transitions between states, as shown in Table 4. These probabilities reflect the likelihood of different outcomes in the offloading process, such as successful task submission, resource allocation, and processing completion.
The complete model captures the dynamic nature of offloading decisions in edge computing environments, considering factors such as resource availability, network conditions, and energy-time trade-offs. The CTMC formulation allows for probabilistic analysis of different offloading strategies and their impact on system performance.

4. Quantitative Analysis and Verification

In this section, we first present the properties that need to be analyzed and verified, followed by a quantitative analysis and verification of the experimental parameters and results.

4.1. Property Specification

We formalize the following probabilistic properties to verify the correctness of our offloading decision strategy.
Overall success probability: This represents the probability that the entire strategy can eventually successfully handle the task within a given number of steps and within the specified range of submitted task times.
P = ? [ F s t e p ( h a n d l e _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ]
Overall failure probability: This depicts the probability that the entire strategy fails to process the task successfully within the specified number of steps and the number of submissions is within the specified range, even though all the conditions of the strategy are normal.
P = ? [ F s t e p ( s = 11 & ( h a n d l e _ s u c = f a l s e ) & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ]
Local processing success probability: This indicates the probability that the strategy chooses to offload to local devices and successfully processes the task within the specified number of steps.
P = ? [ F s t e p ( l o c a l _ h a n d l e d _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ]
Edge server success probability: This describes the probability that the strategy chooses to offload to the edge server and successfully processes the task within the specified number of steps.
P = ? [ F s t e p ( m e c _ h a n d l e d _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ]
Cloud server success probability: This shows the probability that the strategy chooses to offload to the cloud server and successfully processes the task within the specified number of steps.
P = ? [ F s t e p ( c l o u d _ h a n d l e d _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ]
Local processing ratio: This expresses the proportion of the probability of successful task processing by local devices to the probability of overall successful processing in the strategy.
P = ? [ F s t e p ( l o c a l _ h a n d l e d _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ] P = ? [ F s t e p ( h a n d l e _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ]
Edge server ratio: This means the proportion of the probability of successful task processing by the edge server to the probability of overall successful processing in the strategy.
P = ? [ F s t e p ( m e c _ h a n d l e d _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ] P = ? [ F s t e p ( h a n d l e _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ]
Cloud server ratio: This outlines the proportion of the probability of successful task processing by the cloud server to the probability of overall successful processing in the strategy.
P = ? [ F s t e p ( c l o u d _ h a n d l e d _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ] P = ? [ F s t e p ( h a n d l e _ s u c = t r u e & s u b m i t _ t i m e s m a x _ s u b m i t _ t i m e s ) ]

4.2. Verification Results and Analysis

In this subsection, we analyze the properties discussed in the previous subsection to demonstrate the usability of the offloading strategies. Additionally, we discuss the curves depicted in the figures that represent the verification results.
Figure 4 illustrates the overall success rate of strategies based on the delay-efficient or energy-efficient offloading strategy. The graph depicts the success rate of strategies completed at different time steps for different task sizes, where b takes the values of 3, 9, 15, and 21. Figure 5 presents the success rates of three different offloading objects, including local offloading, edge offloading, and cloud offloading.
Specifically, the analysis of the success rates of these strategies reveals that the success rate of different task sizes varies depending on the corresponding offloading strategy. It is evident that when other network factors such as bandwidth and computing power remain constant, local computing has the highest success rate but is more suitable for small tasks. In contrast, edge or cloud offloading is more suitable for medium to large tasks, but the overall success rate is lower due to factors such as data transmission delay.
Figure 6 shows the success rate data of strategies when users prefer to balance delay and energy consumption (energy-delay trade-off offloading strategy). It shows the success probabilities of overall computing, local computing, edge computing, and cloud computing. Compared to the overall graph based on delay and energy consumption strategies, the success rate of overall strategies has slightly decreased, mainly due to the potential impact of external environment parameters when balancing the strategies. To investigate the proportion of the three possible offloading computing methods in the overall success rate of strategies, we analyze the data and find that even with an increasing time step, the probability of successful local offloading is still greater than that of edge and cloud offloading shown in Figure 7. The overall curve can assist users in selecting a delay and energy consumption ratio based on the device’s battery level and the pending data task.

5. Related Work

Recent advances in edge computing offloading strategies have primarily evolved along two interrelated directions: algorithm-driven optimization and scenario-specific adaptation. In the domain of algorithmic optimization, various AI-based approaches have been proposed to enhance offloading performance in terms of latency, energy efficiency, and resource utilization. For instance, Wang et al. [8] introduced TPMOA, a deep reinforcement learning method for wireless VR systems, which integrates viewpoint prediction with multi-objective optimization to jointly reduce latency and energy consumption. Yang et al. [9] proposed KDECN, which combines an LSTM-based traffic demand predictor (TRDP) and a DRL-based resource allocator (L-CORA) to optimize dynamic edge network scheduling. Gong et al. [10] leveraged MOEA/D for multi-objective offloading across cloud–edge-end systems, while He et al. [11] formulated a KKT-based analytical solution derived from queueing models to support optimal offloading decisions under constrained conditions.
On the other hand, scenario-specific adaptation approaches have focused on addressing unique environmental and application constraints. Zhang et al. [12] proposed MILMST for infrastructure-less networks, which employs multi-agent imitation learning and Steiner tree-based migration to enhance offloading in the absence of fixed infrastructure. Li et al. [13] tackled high-mobility scenarios in MEC by integrating Lyapunov optimization with fully connected LSTM (FC-LSTM) scheduling to balance latency and energy consumption. Task dependency challenges have also been addressed: Li et al. [14] proposed a graph-based workflow partitioning method, achieving 7.8–9.5% energy savings, while Tang et al. [15] introduced a two-stage offloading strategy combining priority-based task grouping and game-theoretic scheduling for delay-sensitive IoT workloads.
In addition to performance-oriented offloading, recent research in edge and cloud computing optimization has extended toward system-level innovation, particularly in learning-driven resource management, formal verification, and quantum acceleration. For instance, Xu et al. [16] introduced CERACE, which applies prioritized deep Q-networks (P-DQN) for cost-efficient resource allocation, achieving a 30% cost reduction. Yang et al. [17] employed a hybrid heuristic and multi-agent DRL strategy for MEC caching to minimize latency. Similarly, Bo et al. [18] proposed MADSASPG, a MADRL-NOMA scheme for vehicular networks, reducing latency by 35% and energy consumption by 25%, while Su et al. [19] demonstrated that distributed edge control for HVAC systems can match centralized performance.
To improve the reliability of complex edge–cloud systems, formal verification techniques have been explored. Wunderlich et al. [20] extended classical temporal logics (LTL/CTL) by incorporating weight accumulations, enabling quantitative analysis in probabilistic model checking tools such as PRISM. Schneider et al. [21] adapted bounded model checking for imprecise probabilistic timed graph transition systems (IPTGTSs), enhancing verification under uncertainty in cyber–physical systems. Finally, at the frontier of computational acceleration, Jeon et al. [22] proposed QPMC, a quantum-enhanced probabilistic model checking framework that achieves quadratic speedup through quantum amplitude estimation, laying the groundwork for scalable verification of large-scale stochastic systems.
While existing research has made significant progress in optimizing edge offloading strategies, most efforts focus on performance improvement, with limited attention to the formal verification of correctness. To address this gap, this paper proposes a unified strategy that integrates three representative types of offloading strategies and employs probabilistic model checking to conduct quantitative analysis and correctness verification, aiming to provide theoretical guidance for the design and homologous evolution of offloading strategies.

6. Conclusions and Future Work

This paper presented a hybrid offloading strategy that combined delay-based, energy-efficient, and energy-delay trade-off approaches. These strategies were modeled using automata and analyzed with the PRISM model checker to verify various properties. Quantitative analysis showed that, within the energy-delay trade-off strategy, users could select the most appropriate approach based on the probability of successful offloading and their specific needs. Similarly, the success rates for energy-efficient and delay-based strategies were evaluated. This method has provided valuable guidance to ensure the correctness of the design and implementation of offloading strategies.
In future work, we will enhance the hybrid offloading strategy by incorporating network conditions, resource variability, and application demands, addressing current assumptions to better model real-world edge computing scenarios.

Author Contributions

Conceptualization and Validation, J.Y. and Y.F.; Formal Analysis, J.Y., Y.F. and Q.W.; Visualization, J.Y. and Y.Z.; Supervision, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Wenzhou-Kean University internal research (Grant No. ISRG2024008), and Basic Research Programs of Taicang (Grant No. TC2024JC47).

Data Availability Statement

No datasets were generated or analyzed during the current study.

Conflicts of Interest

We declare that we have no conflicts of interest related to this paper.

References

  1. Liu, H.; Xu, H.; Gao, H.; Bian, M.; Miao, H. An Novel Approach to Evaluate the Reliability of Cloud Rendering System Using Probabilistic Model Checker PRISM: A Quantitative Computing Perspective. In Proceedings of the 14th EAI International Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services, Melbourne, Australia, 7–10 November 2017; Gu, T., Kotagiri, R., Liu, H., Eds.; ACM: Singapore, 2017; pp. 78–85. [Google Scholar]
  2. He, X.; Li, T.; Jin, R.; Dai, H. Delay-Optimal Coded Offloading for Distributed Edge Computing in Fading Environments. IEEE Trans. Wirel. Commun. 2022, 21, 10796–10808. [Google Scholar] [CrossRef]
  3. Liu, J.; Yang, P.; Chen, C. Intelligent energy-efficient scheduling with ant colony techniques for heterogeneous edge computing. J. Parallel Distrib. Comput. 2023, 172, 84–96. [Google Scholar] [CrossRef]
  4. Mashhadi, F.; Monroy, S.A.S.; Bozorgchenani, A.; Tarchi, D. Optimal auction for delay and energy constrained task offloading in mobile edge computing. Comput. Netw. 2020, 183, 107527. [Google Scholar] [CrossRef]
  5. Zhang, J.; Yang, F.; Wu, Z.; Zhang, Z.; Wang, Y. Multi-Access Edge Computing and Its Key Technologies; The People’s Posts and Telecommunications Press (Posts & Telecom Press): Beijing, China, 2019. [Google Scholar]
  6. Kwiatkowska, M.Z.; Norman, G.; Parker, D. PRISM 4.0: Verification of Probabilistic Real-Time Systems. In Proceedings of the 23rd International Conference Computer Aided Verification; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2011; Volume 6806, pp. 585–591. [Google Scholar]
  7. Ye, K.; Cavalcanti, A.; Foster, S.; Miyazawa, A.; Woodcock, J. Probabilistic modelling and verification using RoboChart and PRISM. Softw. Syst. Model. 2022, 21, 667–716. [Google Scholar] [CrossRef]
  8. Wang, J.; Xia, H.; Xu, L.; Zhang, R.; Jia, K. DRL-based latency-energy offloading optimization strategy in wireless VR networks with edge computing. Comput. Netw. 2025, 258, 111034. [Google Scholar] [CrossRef]
  9. Yang, K.; Wang, X.; He, Q.; Zhao, L.; Liu, Y.; Tarchi, D. Knowledge-Defined Edge Computing Networks Assisted Long-Term Optimization of Computation Offloading and Resource Allocation Strategy. IEEE Trans. Wirel. Commun. 2024, 23, 5316–5329. [Google Scholar] [CrossRef]
  10. Gong, Y.; Bian, K.; Hao, F.; Sun, Y.; Wu, Y. Dependent tasks offloading in mobile edge computing: A multi-objective evolutionary optimization strategy. Future Gener. Comput. Syst. 2023, 148, 314–325. [Google Scholar] [CrossRef]
  11. He, Z.; Xu, Y.; Liu, D.; Zhou, W.; Li, K. Energy-efficient computation offloading strategy with task priority in cloud assisted multi-access edge computing. Future Gener. Comput. Syst. 2023, 148, 298–313. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Zhang, S.; Zhang, Q.; Fan, J. An intention-driven task offloading strategy based on imitation learning in pervasive edge computing. Comput. Netw. 2025, 257, 110998. [Google Scholar] [CrossRef]
  13. Li, Y.; Cheng, S.; Zhang, H.; Liu, J. Dynamic adaptive workload offloading strategy in mobile edge computing networks. Comput. Netw. 2023, 233, 109878. [Google Scholar] [CrossRef]
  14. Li, X.; Chen, T.; Yuan, D.; Xu, J.; Liu, X. A Novel Graph-Based Computation Offloading Strategy for Workflow Applications in Mobile Edge Computing. IEEE Trans. Serv. Comput. 2023, 16, 845–857. [Google Scholar] [CrossRef]
  15. Tang, J.; Qin, T.; Xiang, Y.; Zhou, Z.; Gu, J. Optimization Search Strategy for Task Offloading From Collaborative Edge Computing. IEEE Trans. Serv. Comput. 2023, 16, 2044–2058. [Google Scholar] [CrossRef]
  16. Xu, Z.; Zhong, Z.; Shi, B. Deep Reinforcement Learning Based Resource Allocation Strategy in Cloud-Edge Computing System. In Proceedings of the International Joint Conference on Neural Networks, IJCNN 2022, Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar]
  17. Yang, Y.; Lou, K.; Wang, E.; Liu, W.; Shang, J.; Song, X.; Li, D.; Wu, J. Multi-Agent Reinforcement Learning Based File Caching Strategy in Mobile Edge Computing. IEEE/ACM Trans. Netw. 2023, 31, 3159–3174. [Google Scholar] [CrossRef]
  18. Bo, J.; Zhao, X. Vehicle Edge Computing Task Offloading Strategy Based on Multi-Agent Deep Reinforcement Learning. J. Grid Comput. 2025, 23, 13. [Google Scholar] [CrossRef]
  19. Su, B.; Li, X.; Wang, S.; Cao, J. Distributed Optimal Control for HVAC Systems Adopting Edge Computing-Strategy, Implementation, and Experimental Validation. IEEE Internet Things J. 2022, 9, 11858–11867. [Google Scholar] [CrossRef]
  20. Wunderlich, S. Probabilistic Model Checking for Temporal Logics in Weighted Structures. Ph.D. Thesis, University of Dresden, Dresden, Germany, 2024. [Google Scholar]
  21. Schneider, S.; Maximova, M.; Giese, H. Bounded model checking for interval probabilistic timed graph transformation systems against properties of probabilistic metric temporal graph logic. J. Log. Algebr. Methods Program. 2024, 137, 100938. [Google Scholar] [CrossRef]
  22. Jeon, S.; Cho, K.; Kang, C.G.; Lee, J.; Oh, H.; Kang, J. Quantum Probabilistic Model Checking for Time-Bounded Properties. Proc. ACM Program. Lang. 2024, 8, 557–587. [Google Scholar] [CrossRef]
Figure 1. The workflow of offloading strategies in edge computing systems.
Figure 1. The workflow of offloading strategies in edge computing systems.
Electronics 14 02236 g001
Figure 2. The stateflow of the integrated offloading strategy.
Figure 2. The stateflow of the integrated offloading strategy.
Electronics 14 02236 g002
Figure 3. The automaton of the offloading strategy relevant to Figure 2.
Figure 3. The automaton of the offloading strategy relevant to Figure 2.
Electronics 14 02236 g003
Figure 4. The probability of successful execution of the strategy α = 0 or 1.
Figure 4. The probability of successful execution of the strategy α = 0 or 1.
Electronics 14 02236 g004
Figure 5. The probability of unsuccessful execution of the strategy α = 0 or 1.
Figure 5. The probability of unsuccessful execution of the strategy α = 0 or 1.
Electronics 14 02236 g005
Figure 6. The probability of successful handling of the offloading strategy 0 < α < 1 .
Figure 6. The probability of successful handling of the offloading strategy 0 < α < 1 .
Electronics 14 02236 g006
Figure 7. The rate result of successful handling of the offloading strategy 0 < α < 1 .
Figure 7. The rate result of successful handling of the offloading strategy 0 < α < 1 .
Electronics 14 02236 g007
Table 1. The judgements of the delay and energy consumption in the offloading strategy.
Table 1. The judgements of the delay and energy consumption in the offloading strategy.
Strategy TypeNotationOffloading Decision
Delay-based min ( T l , T m , T c ) = T i i = l : local execution
i = m : edge server offloading
i = c : cloud server offloading
Energy-efficient min ( E l , E m , E c ) = E i
Energy-delay trade-off C = α · T + ( 1 α ) · E User-defined
Table 2. The metrics and descriptions in the offloading strategy.
Table 2. The metrics and descriptions in the offloading strategy.
MetricsNotationsExplanations
Delay Tmthe quantity of computing resources
s l o c a l the local execution speed
t l o c a l the local execution time
Bthe current network bandwidth
bthe current computation task size
s m e c the computation speed at edge server
t m e c the time required to offload
and execute on edge server
s c l o u d the computation speed at cloud server
t c l o u d the time required to offload
and execute on cloud
Consumption E P l o c a l the power consumed by the mobile device
during operation
P t r a n s m i t the transmission power during uploading
P c o n n e c t the power consumed by the mobile device to
maintain connection with the network
E l o c a l the energy consumption cost of local computing
E m e c the energy comsumption cost of edge computing
E c l o u d the energy comsumption cost of cloud computing
Composite Metric C C l o c a l the composite metric of the cost of local computing
C m e c the composite metric of the cost of edge computing
C c l o u d the composite metric of the cost of cloud computing
Table 3. The state name and its corresponding explanations in Figure 3.
Table 3. The state name and its corresponding explanations in Figure 3.
State NameExplanations
s 0 the start of the offloading strategy
s 11 checking data
s 12 resubmitting data
s f a i l strategy failing to execute
s 21 delay-based or energy-efficient
s 22 offload to edge or cloud server
s 22 entering the same probability workflow as the edge server
s 22 entering the same probability workflow as the cloud server
s 23 offloading to local
s 24 processing data
s 25 waiting to allocate resources
s 26 returning the results
s 31 energy-delay trade-off
s 32 choosing to offload to local
s 33 choosing to offload to edge server
s 34 choosing to offload to cloud server
s s u c successful strategy execution
Table 4. The name and value of the probability adopted in our model.
Table 4. The name and value of the probability adopted in our model.
Probability NameProbability Value
p 11 0.98
p 12 0.95
p 13 0.97
p 14 0.57
p 21 0.94
p 22 0.93
p 23 0.93
p 24 0.55
p 25 0.56
p 31 1/3
p 32 1/3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yin, J.; Fei, Y.; Wu, Q.; Zhao, Y. Quantitative Analysis and Verification of Edge Computing Offloading Strategy Based on Probabilistic Model Checking. Electronics 2025, 14, 2236. https://doi.org/10.3390/electronics14112236

AMA Style

Yin J, Fei Y, Wu Q, Zhao Y. Quantitative Analysis and Verification of Edge Computing Offloading Strategy Based on Probabilistic Model Checking. Electronics. 2025; 14(11):2236. https://doi.org/10.3390/electronics14112236

Chicago/Turabian Style

Yin, Jiaqi, Yuan Fei, Qiangyu Wu, and Yue Zhao. 2025. "Quantitative Analysis and Verification of Edge Computing Offloading Strategy Based on Probabilistic Model Checking" Electronics 14, no. 11: 2236. https://doi.org/10.3390/electronics14112236

APA Style

Yin, J., Fei, Y., Wu, Q., & Zhao, Y. (2025). Quantitative Analysis and Verification of Edge Computing Offloading Strategy Based on Probabilistic Model Checking. Electronics, 14(11), 2236. https://doi.org/10.3390/electronics14112236

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop