Abstract
This paper proposes a cooperative framework for resource allocation in multi-access edge computing (MEC) under a partial task offloading setting, addressing the joint challenges of learning performance and system efficiency in heterogeneous edge environments. In the proposed architecture, selected users act as edge servers (SEs) that collaboratively assist others alongside a central server (CS). A joint optimization problem is formulated to integrate model training with resource allocation while accounting for data freshness and spatial correlation among user tasks. The correlation-aware formulation penalizes outdated and redundant data, leading to improved robustness against non-i.i.d. distributions. To solve the NP-hard problem efficiently, a projected gradient descent (PGD) method is developed. The simulation results demonstrate that the proposed cooperative approach achieves a balanced delay of s, close to edge-only computing ( s) and lower than the CS-only mode, while improving clustering accuracy to (up to higher than the baseline). Moreover, it reduces the central server load by nearly half, ensuring scalability and latency compliance within 3GPP limits. These findings confirm that cooperation between SEs and the CS substantially enhances reliability and performance in distributed Edge-AI system.
1. Introduction
1.1. State-of-the-Art and Motivations
Recent advances in artificial intelligence (AI) have sparked innovation in various technological domains. Most conventional AI solutions rely on centralized computing architectures that depend on large-scale data collection. While these centralized models offer powerful learning, perception, and decision-making capabilities, they have their own limitations in dynamic and uncertain environments. Transferring data to a central server and accumulating it for computing is time-consuming and costly. Additionally, providing adequate computing resources on a central server is challenging, especially under high demand or time-sensitive conditions. To address these challenges, researchers have proposed more distributed and adaptive architectures. One prominent approach is multi-access edge computing (MEC), which brings computing and storage resources closer to end devices at the edge of the network. This configuration improves task offloading efficiency and enables real-time system responsiveness. Edge AI enables the execution of AI models directly at the network edge, allowing real-time decision-making with reduced latency and lower reliance on centralized servers. This approach is particularly valuable in time-sensitive and resource-constrained environments [,]. Due to the significant advantages of Edge AI, such as enhanced operational efficiency, improved data privacy, and ultra-low latency, global interest and investment in this field have surged in recent years. According to market analysis, the global Edge AI market is projected to exceed USD 66 billion by 2030 [,,].
Despite its benefits, Edge AI and MEC are faced with several significant challenges. One of the most critical issues is the propagation of uncertainty in distributed environments. In this regard, stochastic geometry models can provide a powerful analytical framework for characterizing spatial randomness and evaluating network performance [,,,,]. Furthermore, due to asynchronous updates and decentralized model execution, local errors can accumulate and spread across the network, ultimately degrading the overall model accuracy and reliability. Another major challenge lies in the non-independent and identically distributed (non-i.i.d) nature of user data. While various solutions have been proposed in the context of federated learning (FL) to mitigate this issue, concerns remain regarding the generalization and reliability of global models trained on highly heterogeneous local datasets []. Such data distribution disparities can significantly disrupt model convergence and performance. While most existing works on non-i.i.d data in FL and machine learning (ML) focus on non-uniformity of local data distributions (e.g., label skew or quantity skew), fewer studies address the non-independence aspect, where user data may be statistically correlated or dependent across clients [,,].
These challenges highlight the need for robust and correlation-aware learning frameworks that maintain accuracy, adaptability, and convergence under decentralized and statistically dependent conditions. In such environments, traditional centralized or federated approaches often fall short, particularly when dealing with heterogeneous and temporally dynamic data [,,].
To address these limitations, we propose a distributed learning framework tailored for partial task offloading in MEC environments. Unlike conventional paradigms, our approach enables user devices to selectively offload task-related data to edge or central servers (CS) via data-level cooperation. This setup leverages local computation, respects delay constraints, and accommodates statistical heterogeneity across users. A key novelty of our model lies in its integration of spatial and directional task correlations. This is often overlooked in existing works, i.e, the data is treated as being i.i.d which is not an accurate assumption. We introduce a correlation-aware loss function that explicitly incorporates data freshness and penalizes contributions from distant or weakly correlated users. This design enhances the relevance of updates, reduces the impact of outdated or biased data, and improves overall learning robustness in Edge AI systems. Using stochastic geometry analysis, we can analytically guarantee that users can be served by at least one local server at any time. In summary, the main question addressed in this paper is as follows: How can user tasks in a network be served efficiently under a partial offloading MEC scenario, with priority for edge servers, while taking into account the characteristics of an AI-driven edge model, including communication latency and the challenges posed by non-i.i.d data?
1.2. Contributions
Different from previous works [,,,,,], this paper introduces a novel cooperative computation framework for partial offloading in MEC environments. The main contributions are as follows:
- We propose a cooperative partial offloading model in MEC environments, where tasks can be processed either locally by neighboring users or centrally by a server, taking spatial and directional correlations into account where a closed-form upper bound for spatial correlation is derived to constrain offloading decisions based on user proximity and sensing overlap.
- To minimize learning loss under delay and resource constraints, a novel optimization problem is formulated that incorporates freshness-aware weighting, correlation modeling, and allocation decisions. Moreover, to improve robustness in non-i.i.d. settings, we integrate earth mover’s distance (EMD) into the loss function to capture distributional dissimilarity among users’ data.
- To ensure scalability, we develop a coordination-free solution method suitable for practical deployment in distributed MEC systems.
- Leveraging stochastic geometry, we provide tractable analytical characterizations of coverage probability and delay distribution, which not only enable probabilistic guarantees on task offloading but also ensure the generalizability of the proposed framework to large-scale and heterogeneous MEC networks.
- We show that using the proposed framework, we can significantly reduce the computation load on the central server compared to baseline schemes and for a given delay threshold, we can give service to considerably higher number of users.
2. Related Works
In this section, we first review related works on MEC systems and their resource management strategies, followed by a discussion on studies focusing on Edge AI. Regardless of the method used to solve the optimization problem in MEC-based systems, task offloading is typically classified as either full or partial. In full offloading, the entire task is computed at the edge server, while in partial offloading, the task is split between local and edge computing server to balance latency and resource usage [,,]. In [] the authors investigate parallel task offloading in fog-enabled IoT networks, where computational tasks are divided into multiple sub-tasks and executed concurrently across heterogeneous fog nodes. They formulate the resource allocation problem with the objective of minimizing overall task latency while ensuring simultaneous completion of all sub-tasks, which is essential for efficient utilization of distributed resources. To address the inherent instability caused by interdependent task assignments, the authors design a matching-based allocation framework that maintains stable associations between task-originating devices and helper nodes. Through extensive simulations, the proposed method is shown to reduce average latency by up to 52% under high workload conditions, outperforming several state-of-the-art offloading strategies and demonstrating its suitability for large-scale, delay-sensitive IoT systems. In [] the authors propose a computation offloading scheme for IoT applications involving dependent tasks. This scheme consists of two main components: a multi-queue priority algorithm that schedules dependent sub-tasks, and a deep reinforcement learning method based on Actor–Critic for making dynamic offloading decisions. The framework is designed to minimize task completion time and energy consumption in edge environments by handling task dependencies and fluctuating network conditions efficiently. By leveraging the Lyapunov method, the stochastic optimization problem is stated in [] with the aim of providing a joint task offloading and resource allocation framework for edge-assisted machine learning inference. This framework focuses on minimizing end-to-end latency while preserving inference accuracy and queue stability. The model considers local and edge inference options, whereby each console can adapt the quality of uploaded data and dynamically decide on offloading and computing strategies. The problem is decomposed into three sub-problems: offloading and channel allocation; data quality adjustment; and computational resource assignment. These sub-problems are solved using convex optimization and low-complexity heuristic methods. The simulation results demonstrate substantial improvements in latency reduction and stability under dynamic edge environments.
In [], the authors proposed a method called the Decentralized Distributed Sequential Neural Network (DDSNN), tailored for low-power devices in wireless sensor networks, where conventional deep models face strict memory and energy constraints. By sequentially partitioning a LeNet model across multiple nodes, DDSNN enables fully decentralized inference without the need for compression or centralized coordination. In a predictive maintenance case study involving industrial pump vibration data, the framework preserved full precision, achieved 99% accuracy, and reduced inference latency by nearly 50% compared to the baseline. Although the accuracy gain over the non-distributed model was marginal, the authors emphasize that in highly resource-constrained settings, even slight improvements are significant, making DDSNN a practical and scalable solution.
The approach proposed in [] addresses FL in heterogeneous edge environments by allowing each edge device to perform a different number of local updates, adapting to their computational capacity. This idea lays a foundation for handling heterogeneity in edge systems and can be extended further to include other real-world factors. For instance, while the original model focuses on computational diversity, it does not account for communication delays or queuing effects that are common in distributed systems. Additionally, the influence of data heterogeneity is not explicitly modeled.
Authors in [] introduce a dynamic client selection strategy for FL, aiming to improve training efficiency under resource constraints in edge environments. The approach incorporates client-side characteristics such as computation power and data volume into a selection metric to adaptively involve suitable participants in each communication round. This method improves convergence speed and reduces communication cost. However, the model assumes relatively consistent client availability and does not explicitly address delay variability or the impact of severe data heterogeneity, which are common in real-world edge networks. Moreover, the paper focuses primarily on client selection policies rather than modeling deeper aspects such as uncertainty propagation or the effects of delayed or imbalanced updates on global performance. These aspects open opportunities for extending this methodology toward more delay-aware and distribution-sensitive learning frameworks.
To improve personalization in FL, the authors in [] propose an adversarial training framework combined with data-free knowledge distillation. Their method leverages earth mover’s distance (EMD) to align local and global data distributions, effectively addressing the challenge of non-i.i.d data across clients while preserving privacy. Although this framework improves global model adaptation, it does not consider delay sensitivity, dynamic task partitioning, or computing constraints common in edge environments. Nevertheless, the idea presents a promising direction that can be extended to such realistic, delay-aware, and resource-constrained edge AI settings.
With the aim of addressing label distribution skew in FL, the authors in [] propose a novel learning approach that leverages knowledge distillation and a label-invariant teacher-student framework. Their method focuses on mitigating performance degradation due to non-i.i.d label distributions without directly sharing model parameters. This work sheds light on the importance of decoupling label heterogeneity from the model optimization process. While the approach effectively addresses label skew, it can be extended to incorporate delay sensitivity and partial offloading in edge computing environments—dimensions must be considered.
Furthermore, authors in [] have explored pruning-aware collaborative inference of large AI models at the edge, where model partitioning and resource optimization are jointly considered to balance accuracy, latency, and energy consumption. While this provides important insights into enabling efficient edge inference, it does not address aspects such as training under heterogeneous data distributions and stochastic task offloading, which remain central to advancing edge intelligence.
To facilitate a more effective comparison between this work and prior studies, we present Table 1.
Table 1.
Comparison of related works with our proposed framework.
Symbol Notation: and denote the probability density function (PDF) and cumulative distribution function (CDF) of x. denotes the probability of x. is the absolute value, and is the ceiling function.
3. System Model and Parameters
This article focuses on a task offloading scenario in which data-generating equipment, referred to as requesting entities (REs), can delegate their computational tasks either to nearby local edge devices, referred to as serving entities (SEs), or to a centralized server (CS). The goal is to enable efficient and delay-sensitive learning by leveraging Edge AI frameworks, where training occurs across distributed nodes with limited resources and potentially heterogeneous data distributions. In this regard, as can be seen in Figure 1, we consider a central server for task computing.
Figure 1.
A hybrid edge architecture in which some entities request task execution (REs) and others provide computing resources (SEs). A CS is also available for additional offloading. Typical use cases include smart campuses or factories, where IoT devices and robots can be considered as REs, and where attendees smartphones can be considered as SEs performing partial task offloading.
The set of equipment whose tasks needed to be addressed is represented by . In addition, the set of equipment that can provide computing resources for local task computing is denoted by . We assume that the devices are randomly distributed over an area of size (in square meters), following a Poisson Point Process (PPP) [,,]. Let and denote the spatial intensities of SEs and REs, respectively, measured in devices per square meter. According to the properties of the PPP, the probability of observing exactly k devices of a given type—either SEs or REs—within a region of area is given by where denotes the spatial density (intensity) of SEs or REs, respectively. Furthermore, we assume that the spatial distributions of SEs and REs are independent (In our model, REs represent IoT devices or sensors that are typically deployed in large numbers and remain stationary within a given environment (e.g., a campus or factory). In contrast, SEs correspond to mobile devices, such as smartphones or laptops, which are carried by human users. Since the placement of REs is governed by specific deployment requirements, while the mobility of SEs is driven by human movement patterns, it is reasonable to assume that the spatial distributions of REs and SEs are mutually independent. Moreover, the locations of individual REs and SEs are also assumed to be independent of one another. This independence assumption is widely adopted in stochastic geometry models, as it enhances both realism and analytical tractability) of each other, consistent with the properties of independent homogeneous PPPs [,,].
The size of the task for RE r is denoted by (in byte), while the computing capacity of SE s is represented by (in CPU cycles/s). Similarly, the computing capacity of the CS is denoted by . In this paper, we assume that the task data consists of image-based content. (For instance, one can envision a scenario involving sensors and mobile robots deployed across environments such as a university campus, smart factory, or smart city. These devices are responsible for capturing images of their surroundings to support perception and decision-making tasks.) Accordingly, we adopt a widely used benchmark dataset to model the image data in our experiments. In line with this, we define a dynamic field of view (FoV) for each RE r, represented by an angular direction at time slot t. This direction evolves over time according t: where is the initial viewing angle and is a constant angular increment per time slot. In addition, each RE is assumed to have a symmetric viewing range with a total width of , meaning that at time t, RE r can observe all the objects located within the angular sector If the angular displacement between two consecutive time slots satisfies , then the task based on the data collected at time t is assumed to be independent of the task based on the data collected at time . Otherwise, a dependency exists, which will be addressed in the subsequent sections.
Furthermore, the maximum visual sensing range for all REs is assumed to be equal and is represented by L. In addition, the time that RE r collects the data for its task, is denoted by . Furthermore, the maximum range of each edge server is denoted by . In this paper, we assume that each RE can be covered by at least servers but can only be assigned to one, the details of which is provided in Section 5. A data sample from the dataset of RE r is denoted by , where represents the dataset associated with RE r. The size of each REs dataset is modeled as a uniform random variable:
where and denote the minimum and maximum number of samples, respectively. Here, , where N is the total number of samples in the global dataset, and quantifies the degree of quantity skew across the REs. Here we assume that for the transmitted , the received version at SEs is represented by , and at the CS, by , where the effect of wireless channel transmission such as additive noise and fading have in been incorporated within them.
We use notation to represent the assignment variable, where if RE r is assigned to SE s, and otherwise. Let represent the fraction of the data task from RE r that is offloaded to the CS. Consequently, denotes the fraction of the task offloaded to the assigned SE s. Accordingly, the number of data samples offloaded to the CS by RE r is , and we denote this subset as . Similarly, the number of samples offloaded to SE s is , and we denote this subset as . To increase the readability of paper, the list of symbols are provided in Table 2.
Table 2.
List of system model parameters.
4. The Proposed Problem
4.1. Constraints
Each RE must be assigned to one SE, as a result,
Furthermore, the following constraints ensure that the total computational load assigned to each SE and the CS does not exceed their respective computing capacities:
where denotes the number of CPU cycles required per byte []. In addition, we consider computing delay, transmission delay, and queuing delay. The computing delay is calculated by:
The transmission delay also is obtained by the following:
where V and are the transmission data rate between REs and SEs and the CS, in byte per second, respectively. By considering the queuing model for both the CS and the SEs, the total queuing delay experienced by RE r under partial offloading is modeled by Pollaczek–Khinchin formula. We denote by the task generation rate of each RE (in tasks per second). The net arrival rates to CS and to SE s are then
where denotes the average offloading fraction across REs (for large systems one may use ). For the CS, the service time of a task originating from RE r (seconds) is
Consequently, the first and second moments of the service time at the CS are
By the Pollaczek–Khinchin formula for an queue, the mean waiting time in queue (excluding service) at the CS is
Analogously, for SE s with service time we have
Finally, the expected queuing delay experienced by RE r (in seconds) under partial offloading and soft assignments is given by the mixture
Finally, by the following constraint, we guarantee that total delay must be less than threshold of maximum delay :
4.2. Data Freshness and Distribution-Aware Loss Modeling
Due to the geographical distribution of REs, each RE observes a distinct data distribution, leading to statistical heterogeneity (non-i.i.d) across the system. This diversity may adversely affect model convergence, as many solutions assume i.i.d data and are developed based on this assumption. We refer to this scenario as non-i.i.d.-blind and we try to develop a non-i.i.d.-aware model. To do so, we formulate a delay-aware and distribution-sensitive loss model that incorporates statistical dissimilarity, temporal dynamics, and communication delay penalties. The global loss function at the CS is given by the following:
where is the local loss. In parallel, the loss function at each SE s is computed as follows:
where the delay penalty term is defined as and denotes the time when the sample from RE r was generated. This function prioritizes fresher data by assigning smaller penalties to more recent samples. The dissimilarity coefficient quantifies the statistical distance between RE r’s local distribution and the global distribution using EMD:
To model temporal dynamics, assuming that at time slot , only classes have been observed (i.e., ), while the remaining classes are unseen, the estimated class probabilities at time t are as follows:
where controls the decay of outdated distributions and ensures normalization (as shown in Appendix A). Furthermore, if all classes have been seen at time slot , , we assume that . Although this work focuses on quantity skew due to non-uniform sensing rates and task sizes, feature skew is partly reflected through the spatial randomness of RE datasets and wireless noise distortions. Label skew is not directly relevant here since the framework operates in an unsupervised setting. Future extensions could explicitly incorporate feature-level or domain-shift variations to capture broader heterogeneity conditions in Edge AI.
In addition, each RE is assumed to have a symmetric FoV with total angular width , centered around its viewing angle . That is, at time t, RE r can observe all objects located within an angular sector of width centered at .
We consider REs uniformly distributed within a circular region of radius L:
We incorporate both spatial and directional correlations between the tasks of REs r and using a physically meaningful and dimensionally consistent formulation. The correlation depends on their positions , and viewing angles , , with angular difference . As illustrated in Figure 2, the task correlation is modeled as follows:
where is the spatial correlation length and is the angular correlation parameter, ensuring dimensional consistency. To ensure the correlation remains below a threshold , we require
Figure 2.
An illustrative example of how the FoV of REs overlap and how their spatial correlation influences their observations. The yellow star represents a typical point that is visible to both users.
This leads to the following geometric condition:
For system design, we consider the expected spatial configuration. Under uniform distribution, the expected squared distance is . This yields the following simplified angular correlation:
where represents the spatial correlation baseline. The fundamental design constraint becomes the following:
provided the argument lies in . This closed-form expression provides a practical design guideline, clearly showing the trade-offs between spatial coverage (L), correlation parameters (, ), and the correlation threshold () (more details are provided in Appendix B).
4.3. Problem Formulation
With the aim of minimizing the total loss of the system, i.e., the loss of the CS and SEs over variables , , and , subject to the constraints discussed above, we state the following optimization problem:
where constraint (26b) guarantees that each RE is assigned to exactly one SE, while (26c) ensures that the total delay does not exceed the maximum allowable threshold . In addition, constraints (26d,e) restrict the sizes of tasks offloaded to the local SEs and the CS so that they remain within their respective computing capacities.
5. Feasibility Analysis
In this section, we present a mathematical analysis to evaluate the feasibility of the system and ensure its consistency. Based on the properties of a homogeneous PPP, the PDF of the distance D between an RE and its nearest serving entity SE is given by
where denotes the density of SEs. The expected distance can be derived as
To guarantee that REs are within a maximum distance threshold from at least one SE with high probability, we impose
where is the outage tolerance (i.e., the probability that an RE is not covered within ). Rearranging the above condition yields the following requirement on the SE density:
This condition provides a lower bound on the spatial density of SEs to achieve a coverage probability of at least within the distance threshold . To satisfy the requirement that each RE is covered by at least servers, we analyze the coverage under a homogeneous PPP. Let denote the number of SEs within distance of a typical RE. Then, . To ensure , we must have the following:
and finally we have the following:
which implicitly provides a lower bound on as a function of and the target reliability . To illustrate, consider required SEs per RE. When the reliability target is , the corresponding coverage intensity is approximately . Hence, the minimum server density should satisfy the following:
For , this yields the following:
which is equivalent to one SE per area on average. This corresponds to approximately 20 SEs per square kilometer, ensuring each RE has at least three SEs in range with reliability.
If the reliability requirement is increased to , the coverage intensity increases to about , resulting in the following:
For , this reduces to , which still guarantees that each RE is covered by three SEs with probability at least .
These results numerically confirm that moderate SE densities, on the order of servers/m2, are sufficient to maintain reliable multi-server coverage within typical urban cell sizes. Hereupon the set of servers to which RE r can be assigned based on the physical distance is denoted by . Assuming identical task sizes and SE capacities , the total required computing resources in the network is . If all REs adopt a uniform offloading strategy , then the computing loads are split as for the CS and for the SEs. Assuming the total CS capacity is and the total SE capacity is , the maximum number of REs the network can support under this strategy is
While this analysis provides an upper bound, it does not incorporate the binary task assignment variables , which govern the actual RE-to-SE allocation decisions. As a result, the derived expression represents an idealized scenario. The practical feasibility of this result depends on whether the local constraints at each SE can be satisfied under the discrete assignment structure. Analytically, for we have Figure 3 illustrates this trade-off across values of .
Figure 3.
Effect of on the maximum number of REs supported by the network.
6. Solution Method
6.1. Dataset Description
To evaluate the performance of the proposed model, we use the widely adopted MNIST dataset in image classification and ML-based systems [,,]. The MNIST dataset consists of 70,000 gray-scale images of handwritten digits (0–9), each of size pixels, with 60,000 samples for training and 10,000 for testing [,]. To simulate a realistic ML based setting, both datasets are partitioned in a non-i.i.d fashion across multiple edge nodes. Each node is assigned a unique subset of the data to reflect user-specific distributions, capturing the impact of data heterogeneity and decentralized learning on model performance.
6.2. Proposed Solution
We address the joint optimization problem in (26), which involves the offloading ratios , task assignment variables , and learning model parameters . The combinatorial nature of the binary assignment variables makes direct optimization computationally prohibitive. To overcome this challenge, we employ a continuous relaxation framework, in which discrete decision variables are parameterized through smooth mappings and optimized via projected gradient descent (PGD).
6.2.1. Joint and Disjoint Optimization Protocols
To evaluate the effectiveness of the proposed optimization, two schemes are developed. In the joint optimization approach, as the proposed method, all the parameters are updated simultaneously using PGD, which allows end-to-end coordination between learning dynamics and resource allocation decisions. In contrast, the disjoint optimization approach adopts, as a baseline, a sequential structure; one variable (e.g., ) is fixed while the remaining parameters are optimized iteratively; and the process repeats by substituting the latest updates in subsequent optimization rounds until convergence. Both optimization protocols operate under identical dataset partitions, computing capacities, and latency constraints, ensuring a fair and consistent comparison of their convergence behavior and performance.
6.2.2. Assignment Relaxation
Let denote the distance between RE r and SE s. The normalized distance is defined as
The normalized load of SE s at time t is
where represents the fraction of available capacity at SE s, and . Based on these features, we define an affinity score between RE r and SE s:
which encourages assignments toward nearby and less-loaded servers. Matrix is then expressed as a linear parametric function:
The relaxed assignment is obtained via a softmax mapping:
which ensures lies on the probability simplex.
6.2.3. Offloading Ratio Relaxation
For each RE r, we define an aggregate edge suitability score:
which increases when nearby SEs have higher available capacity. The central-server logit, collected into the vector is parameterized as
leading to the relaxed offloading ratio
where is the sigmoid function. Thus, higher edge suitability reduces , prioritizing task processing at the edge.
6.2.4. Joint Optimization
The relaxed decision variables and the learning model parameters are optimized on the augmented objective:
Here, denotes the unsupervised learning loss, composed of a reconstruction term and a distribution-alignment term to address non-i.i.d. data across servers:
where denotes the reconstructed task from latent representation z, and is a statistical divergence measure. The penalty terms ensure feasibility with respect to capacity and delay constraints:
6.2.5. PGD Mathematical Details
The optimization problem in (26) is of the constrained form,
where is the total system loss and is the feasible set induced by (26b–e). To solve this, PGD alternates between a gradient step
and a projection step
where is the learning rate and denotes Euclidean projection onto :
The gradient step reduces the objective in the unconstrained space, while the projection enforces feasibility with respect to capacity and delay. Under mild assumptions (e.g., Lipschitz continuity of ), PGD converges to a first-order stationary point. PGD updates at each iteration k, and the relaxed logits and the model parameters are updated via
The sigmoid and softmax mappings ensure that and remain valid throughout the updates. Finally, discrete assignments are obtained as
The overall procedure is summarized in Algorithm 1.
| Algorithm 1 Joint PGD-based offloading and assignment optimization |
|
6.2.6. Computational Complexity and Scalability Discussion
The computational complexity of the proposed PGD-based joint optimization mainly arises from gradient evaluations and projection operations. Each iteration involves operations, where R and S denote the numbers of REs and SEs, respectively, and is the number of learnable model parameters. Since the projection step is performed in closed form for and , the overall per-iteration complexity scales linearly with network size. This makes the framework scalable to large MEC deployments. In terms of energy efficiency, cooperative task processing reduces redundant transmissions and central computations, leading to an estimated decrease in total energy consumption compared to the fully centralized baseline, as verified in our simulations. Hence, the PGD formulation achieves a balanced trade-off between convergence speed, scalability, and energy efficiency for real-time Edge AI.
6.2.7. Implementation
The entire framework is implemented in PyTorch. Both the unsupervised reconstruction/alignment objective and the penalty terms are differentiable tensors. Using autograd, gradients are propagated through all components, enabling end-to-end training of via stochastic PGD updates. The learning component is implemented as a lightweight convolutional neural network (CNN) to ensure compatibility with edge devices. The model consists of two convolutional layers with ReLU activations, followed by two fully connected layers. All the trainable parameters are optimized jointly using the Adam optimizer with learning rates , , and . The overall optimization minimizes the total loss , which combines the unsupervised reconstruction and distribution-alignment terms with capacity and delay penalties as defined in (26). This design enables a stable, end-to-end training process while maintaining low computational overhead suitable for resource-constrained edge environments.
6.2.8. Convergence and Penalty Analysis
The convergence of the proposed PGD scheme can be characterized using standard results from constrained optimization theory. Let the total loss function be continuously differentiable with an -Lipschitz gradient, i.e.,
Then, for a fixed learning rate , the PGD iteration
is guaranteed to converge to a first-order stationary point satisfying . In our setting, the feasible set arises from the capacity and delay constraints, while the sigmoid and softmax relaxations ensure that remain differentiable and bounded during optimization.
Furthermore, the penalty terms in (26) provide a smooth relaxation of the original hard constraints:
where and act as trade-off coefficients. By jointly incorporating both penalty components, the optimizer effectively balances constraint satisfaction and model performance within a unified objective surface, leading to improved numerical stability and faster convergence compared to treating each constraint separately.
7. Performance Evaluation
In this part, we evaluate the performance of the proposed MEC-based Edge AI framework under different system configurations, focusing on the interplay between task offloading, computation capacity, and data heterogeneity. The source code and data is available in [].
7.1. Task Offloading and Delay
In Figure 4, we obtain the loss value for CS and SEs. As can be seen, as the steps increase, the loss value for both cases decreases pretty quickly. Since the CS has more data samples and higher computing capacity, the loss function falls more quickly.
Figure 4.
The loss values for and servers.
In Figure 5 we observe that by increasing the number of REs, due to the increase in data volume, the loss decreases. This will contribute in better accuracy as can be seen in Figure 6. note that since our framework follows an unsupervised learning paradigm, the term “accuracy” in all plots refers to clustering accuracy (CA), computed as the optimal label-alignment accuracy between the predicted clusters and ground-truth classes [].
Figure 5.
Loss function for diffrent number of REs.
Figure 6.
Accuracy of model vs. the number of REs.
In Figure 6, we illustrate the effect of the number of REs on overall accuracy and as can be seen, more REs will result in an improved accuracy. On the other hand, more REs mean larger delay. As such, a threshold on the delay will impose an upper bound on the number of REs and simultaneously on the accuracy. For example, 30 REs can provide 99% accuracy. On the other hand, this number of REs causes a delay as large as 0.05 s, as can be seen in Figure 7, which is within the allowed value by 3GPP [,]. In addition, we set [CPU cycle/s] for the CS, [CPU cycle/s] , and [Mega bytes], .
Figure 7.
Effect of number of REs on the total delay in second for the joint scenario (proposed) and disjoint (baseline).
7.2. Comparison with the Baseline Scenario
Now we would like to compare our proposed scheme with a baseline scenario in which the task offloading is performed in a disjoint manner, i.e., each RE is restricted to either the CS or a single SE without coordination. Such a rigid allocation increases the reliance on the CS and leads to inefficient utilization of edge resources. In contrast, our proposed joint framework allows flexible task distribution across CS and SEs, which not only balances the load but also maximizes edge-side computing.
As can be seen in Figure 7, the proposed method causes less delay than the baseline for different values of REs. For example, if the target delay threshold is set to 0.045 s, the proposed scheme can accommodate up to 30 REs, while the baseline could only afford half of this.
To further demonstrate the practical advantage of the proposed cooperative optimization, we compare three deployment modes as only SEs computing, only CS offloading, and the proposed joint scheme with cooperative computing. The results in Figure 8 and Figure 9 show that the cooperative configuration achieves an excellent trade-off and while maintaining an average delay near to the edge-only configuration (0.042 s vs. 0.033 s), it significantly improves clustering average accuracy to 99.2%, surpassing both baselines. These results confirm that the adaptive coordination between SEs and the CS enhances system efficiency, reduces task congestion, and provides stable model performance under heterogeneous network conditions.
Figure 8.
Average total delay for the three modes: only SEs, only CS, and the proposed cooperative scheme. The cooperative mode achieves a balanced delay of 0.042 s, lower than the only CS case (0.060 s) and close to only SEs (0.033 s), in the case of 30 REs.
Figure 9.
Accuracy comparison for only SEs, only CS, and cooperative modes. The proposed method achieves 99.2% accuracy, outperforming only SEs (85%) and only CS (99.7%), in the case of 30 REs.
In Figure 10, we have compared the proposed method and baseline method in terms of the fraction of tasks that are offloaded on CS. As can be seen, for different values of REs, the load on CS is cut into half when using the proposed framework, which is important from a practical point of view.
Figure 10.
Effect of number of REs on the mean of for the joint scenario (proposed) and disjoint (baseline).
7.3. Comparing Non-i.i.d.-Blind and Non-i.i.d.-Aware Scenarios
In this subsection, we first evaluate the performance of the proposed loss model in mitigating the effect of quantity skew, i.e., non-i.i.d.-aware, as formulated in (18) and (19), against the non-i.i.d.-blind scenario that does not account for it. Quantity skew arises when clients (REs) have highly imbalanced numbers of local samples. Unlike the balanced case, where each client contributes equally, the aggregation here is biased toward clients with larger datasets.
In Figure 11, we evaluate the effect of on accuracy for both the non-i.i.d.-blind and non-i.i.d.-aware cases. As can be seen, for all the values of , the proposed scheme improves the accuracy over the baseline scenario.
Figure 11.
Effect of quantity skew coefficient on the total accuracy of model for non-i.i.d.-blind (baseline) and -aware models (proposed), for 30 REs.
Moreover, for very low values of , corresponding to extremely small sample sizes at the clients, accuracy is less due to insufficient data. As increases, the number of available samples grows, which enhances performance and leads to higher accuracy up to an optimal point. Beyond this point, however, the variance in the distribution of data across clients becomes significant, introducing instability and larger errors, which ultimately causes accuracy to decrease. Overall, this demonstrates that the proposed scheme effectively mitigates the negative effects of quantity skew and highlights a non-monotonic relationship between and accuracy.
Finally, we investigate the impact of the correlation threshold on overall system accuracy, considering both the coverage overlap and spatial distribution of REs. Our analysis demonstrates that increasing the maximum allowable correlation between mutually visible REs considerably affects the system’s clustering accuracy, as shown in Figure 12. Specifically, when the correlation threshold is raised, the system incorporates more highly correlated data from overlapping FoVs, which amplifies the non-i.i.d nature of the collected datasets. This increased correlation leads to model overfitting and reduced generalization capability, ultimately degrading clustering performance.
Figure 12.
Effect of correlation on the system accuracy for non-i.i.d.-blind and -aware models.
As far as non-i.i.d.-blind and -aware scenarios are concerned, we can see that the accuracy of the non-i.i.d.-aware case has considerably improved compared to the non-i.i.d.-blind scenario that does not enforce the FoV constraint derived in (22). The difference, in some cases, is above 10% which is very significant in clustering and demonstrates the critical importance of properly regulating angular separation between REs through mathematical formulation of the correlation threshold.
Comparing the last 2 figures, we can see that in contrast to Figure 12, the improvement in accuracy in Figure 11 is not significant. However, it is important to note that for many applications, even a slight improvement in accuracy is critical. For example, in the context of smart manufacturing, this has the potential to reduce miss-clustering in defect detection. Similarly, in the field of healthcare, marginal gains have been shown to enhance diagnostics reliability [,].
8. Conclusions
This work presented a cooperative framework for partial task offloading in Edge-AI systems, leveraging stochastic geometry to capture spatial randomness and to guide correlation-aware resource allocation. A key methodological contribution was the derivation of a closed-form upper bound on spatial correlation, which enabled constraints ensuring that only relevant and timely contributions are included in the global model. We further formulated and solved a joint optimization problem over task assignments, offloading ratios, and learning parameters, and proposed a practical PGD method. Through feasibility analysis and simulations, we demonstrated that the framework effectively addresses the challenges of latency guarantees, accuracy, and scalability in heterogeneous MEC environments. In particular, the results confirm that explicitly handling non-i.i.d. data distributions and spatial correlations leads to superior performance compared to baseline models where learning and resource allocation are decoupled.
Author Contributions
Conceptualization, A.N.; Methodology, A.N.; Software, A.N.; Validation, H.S.; Formal analysis, H.S.; Investigation, A.N.; Resources, H.S.; Writing—original draft, A.N.; Writing—review & editing, H.S.; Visualization, H.S. and A.N.; Supervision, H.S.; Project administration, H.S.; Funding acquisition, H.S. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported by Qatar Research Development and Innovation Council (QRDI) under grant ARG01-0511-230129.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The source code and data of this paper can be found in [].
Conflicts of Interest
The authors declare no conflicts of interest.
Appendix A
We assume that the class distribution for each RE r at time t satisfies the normalization condition:
Assume that at time slot , only classes have been observed (i.e., ), while the remaining classes are unseen. To model the evolution of the class distribution over time while maintaining the normalization constraint, we define the estimated distribution at time t as follows:
where is a decay parameter controlling the memory of previous distributions, and is obtained by
Appendix B
To justify the validity and dimensional consistency of (22), this appendix provides the derivation of the spatial–directional correlation function used in the paper. We consider a homogeneous field of REs distributed according to a Poisson process within a circular region of radius L. Each RE observes its environment within a FoV characterized by an angle and a total width . Let two REs, located at and , have an angular separation and Euclidean distance .
Following isotropic Gaussian field models in spatial statistics [,] and stochastic geometry [], the pairwise correlation between these REs can be expressed as
where denotes the spatial correlation length and the angular correlation parameter, both ensuring dimensional consistency.
For uniformly distributed REs, the average correlation over spatial and angular randomness can be written as
While the above integral does not admit a closed-form solution, a tractable and physically meaningful approximation can be obtained by replacing with its expected value under the uniform distribution, . This substitution gives
which depends only on the normalized spatial scale and the angular separation . This simplification yields a closed-form, dimensionless expression that is well suited for system-level design.
To ensure that the correlation between any two REs remains below a desired threshold , we impose . Using (A5) and solving for yields
Therefore, to suppress excessive task similarity across users, the FoV width should satisfy
This condition provides a physically interpretable trade-off between sensing range, allowable correlation, and FoV overlap. The approximation has been numerically validated and shown to remain within a small deviation of the full triple integral (A4), confirming its suitability for system-level design and analysis.
References
- Ficili, I.; Giacobbe, M.; Tricomi, G.; Puliafito, A. From sensors to data intelligence: Leveraging IoT, cloud, and edge computing with AI. Sensors 2025, 25, 1763. [Google Scholar] [CrossRef] [PubMed]
- Bourechak, A.; Zedadra, O.; Kouahla, M.N.; Guerrieri, A.; Seridi, H.; Fortino, G. At the confluence of artificial intelligence and edge computing in IoT-based applications: A review and new perspectives. Sensors 2023, 23, 1639. [Google Scholar] [CrossRef] [PubMed]
- IEEE Future Networks. Edge Platforms and Services Evolving into 2030. 2021. Available online: https://futurenetworks.ieee.org/podcasts/edge-platforms-and-services-evolving-into-2030 (accessed on 23 July 2025).
- Grand View Research. Edge AI Market Size, Share & Growth|Industry Report, 2030. 2024. Available online: https://www.grandviewresearch.com/industry-analysis/edge-ai-market-report (accessed on 23 July 2025).
- MarketsandMarkets. Edge AI Hardware Industry Worth $58.90 Billion by 2030. 2025. Available online: https://www.marketsandmarkets.com/PressReleases/edge-ai-hardware.asp (accessed on 23 July 2025).
- Yang, J.; Chen, Y.; Lin, Z.; Tian, D.; Chen, P. Distributed Computation Offloading in Autonomous Driving Vehicular Networks: A Stochastic Geometry Approach. IEEE Trans. Intell. Veh. 2024, 9, 2701–2713. [Google Scholar] [CrossRef]
- Gu, Y.; Yao, Y.; Li, C.; Xia, B.; Xu, D.; Zhang, C. Modeling and Analysis of Stochastic Mobile-Edge Computing Wireless Networks. IEEE Internet Things J. 2021, 8, 14051–14065. [Google Scholar] [CrossRef]
- Tran, D.A.; Do, T.T.; Zhang, T. A stochastic geo-partitioning problem for mobile edge computing. IEEE Trans. Emerg. Top. Comput. 2020, 9, 2189–2200. [Google Scholar] [CrossRef]
- Hmamouche, Y.; Benjillali, M.; Saoudi, S.; Yanikomeroglu, H.; Renzo, M.D. New Trends in Stochastic Geometry for Wireless Networks: A Tutorial and Survey. Proc. IEEE 2021, 109, 1200–1252. [Google Scholar] [CrossRef]
- Zhang, Y.; Chen, G.; Du, H.; Yuan, X.; Kadoch, M.; Cheriet, M. Real-time remote health monitoring system driven by 5G MEC-IoT. Electronics 2020, 9, 1753. [Google Scholar] [CrossRef]
- Li, T.; Sahu, A.K.; Talwalkar, A.; Smith, V. Federated learning: Challenges, methods, and future directions. IEEE Signal Process. Mag. 2020, 37, 50–60. [Google Scholar] [CrossRef]
- Solans, D.; Heikkila, M.; Vitaletti, A.; Kourtellis, N.; Anagnostopoulos, A.; Chatzigiannakis, I. Non-i.i.d data in federated learning: A survey with taxonomy, metrics, methods, frameworks and future directions. arXiv 2024, arXiv:2411.12377. [Google Scholar]
- Lu, Z.; Pan, H.; Dai, Y.; Si, X.; Zhang, Y. Federated learning with non-iid data: A survey. IEEE Internet Things J. 2024, 11, 19188–19209. [Google Scholar] [CrossRef]
- Jimenez-Gutierrez, D.M.; Hassanzadeh, M.; Anagnostopoulos, A.; Chatzigiannakis, I.; Vitaletti, A. A thorough assessment of the non-iid data impact in federated learning. arXiv 2025, arXiv:2503.17070. [Google Scholar] [CrossRef]
- Su, W.; Li, L.; Liu, F.; He, M.; Liang, X. AI on the edge: A comprehensive review. Artif. Intell. Rev. 2022, 55, 6125–6183. [Google Scholar] [CrossRef]
- Shi, Y.; Yang, K.; Jiang, T.; Zhang, J.; Letaief, K.B. Communication-efficient edge AI: Algorithms and systems. IEEE Commun. Surv. Tutor. 2020, 22, 2167–2191. [Google Scholar] [CrossRef]
- Letaief, K.B.; Shi, Y.; Lu, J.; Lu, J. Edge artificial intelligence for 6G: Vision, enabling technologies, and applications. IEEE J. Sel. Areas Commun. 2021, 40, 5–36. [Google Scholar] [CrossRef]
- Lin, F.P.C.; Hosseinalipour, S.; Michelusi, N.; Brinton, C.G. Delay-aware hierarchical federated learning. IEEE Trans. Cogn. Commun. Netw. 2023, 10, 674–688. [Google Scholar] [CrossRef]
- Wang, S.; Tuor, T.; Salonidis, T.; Leung, K.K.; Makaya, C.; He, T.; Chan, K. Adaptive federated learning in resource constrained edge computing systems. IEEE J. Sel. Areas Commun. 2019, 37, 1205–1221. [Google Scholar] [CrossRef]
- Xiao, H.; Xu, C.; Ma, Y.; Yang, S.; Zhong, L.; Muntean, G.M. Edge intelligence: A computational task offloading scheme for dependent IoT application. IEEE Trans. Wirel. Commun. 2022, 21, 7222–7237. [Google Scholar] [CrossRef]
- Qiao, D.; Guo, S.; Zhao, J.; Le, J.; Zhou, P.; Li, M.; Chen, X. ASMAFL: Adaptive staleness-aware momentum asynchronous federated learning in edge computing. IEEE Trans. Mob. Comput. 2024, 24, 3390–3406. [Google Scholar] [CrossRef]
- Fan, W.; Chen, Z.; Hao, Z.; Wu, F.; Liu, Y. Joint Task Offloading and Resource Allocation for Quality-Aware Edge-Assisted Machine Learning Task Inference. IEEE Trans. Veh. Technol. 2023, 72, 6739–6752. [Google Scholar] [CrossRef]
- Fan, W.; Li, S.; Liu, J.; Su, Y.; Wu, F.; Liu, Y. Joint Task Offloading and Resource Allocation for Accuracy-Aware Machine-Learning-Based IIoT Applications. IEEE Internet Things J. 2023, 10, 3305–3321. [Google Scholar] [CrossRef]
- Khalili, A.; Zarandi, S.; Rasti, M. Joint Resource Allocation and Offloading Decision in Mobile Edge Computing. IEEE Commun. Lett. 2019, 23, 684–687. [Google Scholar] [CrossRef]
- Kuang, Z.; Li, L.; Gao, J.; Zhao, L.; Liu, A. Partial Offloading Scheduling and Power Allocation for Mobile Edge Computing Systems. IEEE Internet Things J. 2019, 6, 6774–6785. [Google Scholar] [CrossRef]
- Zhang, S.; Gu, H.; Chi, K.; Huang, L.; Yu, K.; Mumtaz, S. DRL-Based Partial Offloading for Maximizing Sum Computation Rate of Wireless Powered Mobile Edge Computing Network. IEEE Trans. Wirel. Commun. 2022, 21, 10934–10948. [Google Scholar] [CrossRef]
- Malik, U.M.; Javed, M.A.; Frnda, J.; Rozhon, J.; Khan, W.U. Efficient matching-based parallel task offloading in iot networks. Sensors 2022, 22, 6906. [Google Scholar] [CrossRef]
- Bolat, Y.; Murray, I.; Ren, Y.; Ferdosian, N. Decentralized Distributed Sequential Neural Networks Inference on Low-Power Microcontrollers in Wireless Sensor Networks: A Predictive Maintenance Case Study. Sensors 2025, 25, 4595. [Google Scholar] [CrossRef]
- Zhao, Y.; Li, M.; Lai, L.; Suda, N.; Civin, D.; Chandra, V. Federated learning with non-i.i.d data. arXiv 2018, arXiv:1806.00582. [Google Scholar]
- Lai, P.; He, Q.; Xia, X.; Chen, F.; Abdelrazek, M.; Grundy, J.; Hosking, J.; Yang, Y. Dynamic User Allocation in Stochastic Mobile Edge Computing Systems. IEEE Trans. Serv. Comput. 2022, 15, 2699–2712. [Google Scholar] [CrossRef]
- Lyu, Z.; Xiao, M.; Xu, J.; Skoglund, M.; Di Renzo, M. The larger the merrier? Efficient large AI model inference in wireless edge networks. arXiv 2025, arXiv:2505.09214. [Google Scholar] [CrossRef]
- Wu, Y.; Zheng, J. Modeling and Analysis of the Uplink Local Delay in MEC-Based VANETs. IEEE Trans. Veh. Technol. 2020, 69, 3538–3549. [Google Scholar] [CrossRef]
- Wu, Y.; Zheng, J. Modeling and Analysis of the Local Delay in an MEC-Based VANET for a Suburban Area. IEEE Internet Things J. 2022, 9, 7065–7079. [Google Scholar] [CrossRef]
- Cheng, Q.; Cai, G.; He, J.; Kaddoum, G. Design and Performance Analysis of MEC-Aided LoRa Networks with Power Control. IEEE Trans. Veh. Technol. 2025, 74, 1597–1609. [Google Scholar] [CrossRef]
- Dhillon, H.S.; Ganti, R.K.; Baccelli, F.; Andrews, J.G. Modeling and Analysis of K-Tier Downlink Heterogeneous Cellular Networks. IEEE J. Sel. Areas Commun. 2012, 30, 550–560. [Google Scholar] [CrossRef]
- Andrews, J.G.; Baccelli, F.; Ganti, R.K. A Tractable Approach to Coverage and Rate in Cellular Networks. IEEE Trans. Commun. 2011, 59, 3122–3134. [Google Scholar] [CrossRef]
- Savi, M.; Tornatore, M.; Verticale, G. Impact of processing-resource sharing on the placement of chained virtual network functions. IEEE Trans. Cloud Comput. 2019, 9, 1479–1492. [Google Scholar] [CrossRef]
- Mu, Y.; Garg, N.; Ratnarajah, T. Federated learning in massive MIMO 6G networks: Convergence analysis and communication-efficient design. IEEE Trans. Netw. Sci. Eng. 2022, 9, 4220–4234. [Google Scholar] [CrossRef]
- Wang, H.; Kaplan, Z.; Niu, D.; Li, B. Optimizing federated learning on non-i.i.d data with reinforcement learning. In Proceedings of the IEEE INFOCOM 2020-IEEE Conference on Computer Communications, Toronto, ON, Canada, 6–9 July 2020; pp. 1698–1707. [Google Scholar]
- He, C.; Annavaram, M.; Avestimehr, S. Group knowledge transfer: Federated learning of large cnns at the edge. Adv. Neural Inf. Process. Syst. 2020, 33, 14068–14080. [Google Scholar]
- Deng, L. The MNIST database of handwritten digit images for machine learning research [best of the web]. IEEE Signal Process. Mag. 2012, 29, 141–142. [Google Scholar] [CrossRef]
- MNIST Dataset on Papers with Code. Available online: https://www.kaggle.com/datasets/hojjatk/mnist-dataset (accessed on 23 July 2025).
- Saeedi, H.; Nouruzi, A. Stochastic-Geometric-based Modeling for Partial Offloading Task Computing in Edge AI Systems. GitHub Repository. Available online: https://github.com/alinouruzi/Stochastic-Geometric-based-Modeling-for-Partial-Offloading-Task-Computing-in-Edge-AI-Systems (accessed on 21 September 2025).
- Xie, J.; Girshick, R.; Farhadi, A. Unsupervised deep embedding for clustering analysis. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 16–24 June 2016; pp. 478–487. [Google Scholar]
- 3GPP. 5G; Service Requirements for the 5G System (3GPP TS 22.261 Version 16.14.0 Release 16). Technical Specification ETSI TS 122 261 V16.14.0, ETSI. 2021. Available online: https://www.etsi.org/deliver/etsi_ts/122200_122299/122261/16.14.00_60/ts_122261v161400p.pdf (accessed on 29 July 2025).
- 5G Hub. 5G Ultra Reliable Low Latency Communication (URLLC). 2023. Available online: https://5ghub.us/5g-ultra-reliable-low-latency-communication-urllc/ (accessed on 29 July 2025).
- Wang, Q.; Haga, Y. Research on Structure Optimization and Accuracy Improvement of Key Components of Medical Device Robot. In Proceedings of the 2024 International Conference on Telecommunications and Power Electronics (TELEPE), Frankfurt, Germany, 29–31 May 2024; pp. 809–813. [Google Scholar] [CrossRef]
- Zrubka, Z.; Holgyesi, A.; Neshat, M.; Nezhad, H.M.; Mirjalili, S.; Kovács, L.; Péntek, M.; Gulácsi, L. Towards a single goodness metric of clinically relevant, accurate, fair and unbiased machine learning predictions of health-related quality of life. In Proceedings of the 2023 IEEE 27th International Conference on Intelligent Engineering Systems (INES), Nairobi, Kenya, 26–28 July 2023; pp. 000285–000290. [Google Scholar] [CrossRef]
- Cressie, N. Statistics for Spatial Data; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
- Dai, R.; Akyildiz, I.F. A spatial correlation model for visual information in wireless multimedia sensor networks. IEEE Trans. Multimed. 2009, 11, 1148–1159. [Google Scholar] [CrossRef]
- Haenggi, M. Stochastic Geometry for Wireless Networks; Cambridge University Press: Cambridge, UK, 2013. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).