1. Introduction
In today’s highly interconnected and volatile economic landscape, the ability to make accurate, real-time decisions is more essential than ever. Financial markets react within milliseconds to new information, and delays in data processing or transmission can amplify systemic vulnerabilities. Traditional centralized computing infrastructures, while powerful, often fail to meet the ultra-low latency requirements of modern economic systems [
1,
2], especially during periods of financial turbulence [
3,
4].
Edge computing has emerged as a transformative solution to this challenge by enabling data processing at or near the source [
5,
6], significantly reducing the response time [
7,
8,
9] and improving operational resilience [
10,
11]. When integrated with Internet of Things (IoT) technologies, edge computing facilitates localized analytics [
12,
13,
14], real-time optimization [
15,
16,
17], and faster reaction cycles in economic systems [
18,
19]. This distributed model is particularly relevant for institutions and platforms that manage financial flows, risk signals, and decision triggers at scale.
One of the critical applications of edge computing in the financial domain lies in the management of financial contagion risk, the rapid transmission of financial shocks from one institution or market segment to others [
4,
20,
21,
22]. During such events, the ability to localize decisions, filter signals at the edge, and dynamically adjust resource allocation can mitigate systemic collapse and improve stability.
The main purpose of this research is to explore how edge computing infrastructures influence the accuracy, responsiveness, and quality of economic decision-making under uncertainty, particularly in the context of financial contagion. Unlike previous studies that regularly explore technological and financial domains separately, this work introduces a hybrid analytical framework that integrates economic behavior signals with real-time edge performance metrics, employing machine learning techniques for predictive and interpretative modeling.
This study contributes to the literature by (i) introducing composite indicators that link computation dynamics with economic decision outputs; (ii) simulating real-time decision environments using IoT-augmented edge systems and synthetic economic scenarios; and (iii) providing empirical evidence on how edge-enhanced infrastructure can be leveraged in order to mitigate systemic risk propagation in complex economic systems.
To achieve this goal, the study addresses the following research questions (RQs):
RQ1: How can composite indicators be constructed to effectively reflect the interplay between edge system performance and real-time economic decision variables?
RQ2: What are the significant relationships between edge computing performance metrics and the outcomes of economic decision-making processes?
RQ3: Which machine learning algorithms provide the most accurate and robust classification of economic decision outcomes in edge-enhanced environments, and how can these models be optimized for effective risk management?
The structure of the paper is as follows.
Section 2 reviews the relevant literature on financial contagion, edge computing, and real-time economic systems.
Section 3 details the methodology, including data description, composite indicator construction, and machine learning models.
Section 4 presents the empirical results and visual analysis.
Section 5 offers a comprehensive discussion on the economic and technological implications of the findings. Finally,
Section 6 concludes this study by answering the research questions, outlining the key contributions, acknowledging limitations, and suggesting avenues for future research.
Unlike previous studies that focus either on technical aspects of edge computing (e.g., latency reduction, blockchain integration) or on network-based models of financial contagion (e.g., exposure matrices, risk spillover effects), this study proposes a hybrid framework that integrates edge performance metrics into economic decision modeling under uncertainty. By constructing composite indicators and applying machine learning classification on real-time decision outcomes, we offer a novel systems-level approach to understanding and mitigating financial contagion risk through technological responsiveness.
3. Materials and Methods
This study employed a data-driven analytical framework designed to explore the intersection between edge computing performance and economic decision-making. The primary objective was to evaluate how real-time computational metrics influence decision accuracy and systemic economic outcomes under uncertain and dynamic conditions. A combination of statistical techniques and machine learning algorithms was used to identify patterns, assess predictive capabilities, and reduce systemic complexity.
The dataset was obtained from Kaggle [
38] and was adapted to simulate real-time economic environments enhanced through edge computing and IoT infrastructure.
The dataset comprises 500 observations and includes both raw and engineered variables that reflect the real-time interaction between digital infrastructure and economic behavior. Although the dataset contains 500 synthetic observations, its design emulates real-time decision-making dynamics in edge-enhanced economic environments, making it suitable for exploratory modeling and interpretability-focused research. Core indicators were grouped into two main categories:
Economic indicators, which capture the behavioral and transactional dimensions of the system:
Transaction volume: simulates the load of economic activity per decision window;
Market behavior index: approximates aggregated sentiment or fluctuation signals;
Financial metric: reflects the monetary intensity or return associated with each decision.
Technological indicators, which measure edge system performance in real time:
- d.
Edge processing latency: the delay incurred at the computation node;
- e.
System throughput: the amount of successfully processed data;
- f.
Resource utilization and workload distribution efficiency: indicates system strain and balance;
- g.
Decision accuracy: the performance of the edge system in making correct decisions;
- h.
Decision outcome: the binary result of the decision process (0 = negative, 1 = positive).
The variable decision accuracy is included directly in the simulated dataset provided by Kaggle [
38] and reflects the correctness level of economic decisions based on system conditions (e.g., latency, throughput, workload balance). It is generated independently from the target variable decision outcome and is not derived from any predictive model trained in this study.
To capture deeper interactions between system dynamics and economic outcomes, four composite indicators were constructed:
These derived metrics are grounded in operational logic and are intended to improve the analytical capacity of the models. The ratio from relation (1) reflects the computational effectiveness of the system in handling economic activity. A higher value indicates that the infrastructure can support a greater number of successful outputs relative to the volume of incoming transactions, serving as a proxy for digital scalability and processing responsiveness.
The construction of the transaction efficiency indicator is grounded in operational research on throughput-based performance evaluation in edge-enhanced financial systems. As shown by Li et al. [
25], latency and the response time are critical dimensions in the security and operational viability of real-time financial networks, particularly in contexts such as supply chain finance. Our indicator builds on this logic by quantifying how efficiently the system handles the transaction load relative to the processing throughput, thus offering a scalable proxy for digital responsiveness.
In relation (2), we describe the latency per transaction metric. This indicator assesses the average processing delay incurred per transaction. It provides insight into the temporal cost of system decisions, with elevated values potentially signaling bottlenecks or inefficiencies in real-time operations that could degrade the decision quality under pressure.
The latency per transaction indicator is derived based on established principles in edge-computing architectures that are aimed at reducing network transmission delays and improving responsiveness in financial systems. As shown in [
39], positioning the edge nodes closer to users and applying predictive caching can reduce overall latency by up to 38%, with edge processing delays maintained under 120 ms. Our indicator captures this temporal performance dimension by normalizing edge latency with respect to the transaction volume, thus reflecting real-time responsiveness under varying economic loads.
The indicator from relation (3) captures the balance between how intensively system resources are being used and how efficiently workloads are being distributed. A mismatch, such as high utilization but low distribution efficiency, may suggest resource saturation or poor orchestration within the edge environment.
The utilization to efficiency ratio introduced in this study reflects the balance between how intensively system resources are used and how effectively workloads are distributed across the edge infrastructure. While previous studies, such as [
40], have employed mobile edge computing for financial risk evaluation, our indicator goes a step further by quantifying operational saturation versus orchestration efficiency, providing a novel proxy for real-time system stability in economic contexts.
The composite score from relation (4) embeds both the correctness and the nature of the decision (positive or negative). By combining these elements, the score quantifies decision effectiveness in a way that is sensitive to the outcome polarity, allowing for a nuanced interpretation of decision performance in response to system or market conditions.
These derived variables provide additional granularity in modeling the decision environment and help reveal the underlying mechanisms that may influence economic behavior and systemic risk, especially under high-frequency or stress-prone conditions.
To evaluate the relationship between edge computing metrics and economic decision outcomes, the analysis followed a structured multi-step methodology, integrating both descriptive statistics and predictive modeling techniques. The analytical workflow is illustrated in
Figure 1, which outlines six core phases designed to explore, preprocess, and model the dataset in a robust and interpretable manner.
The first stage focused on descriptive analysis and data exploration, including the computation of summary statistics and the visualization of distributions and boxplots for key variables. This was followed by a correlation analysis using both Pearson and Kendall methods to capture linear and monotonic relationships between economic and technological indicators. The third phase addressed potential multicollinearity among predictors, assessed via the Variance Inflation Factor (VIF) and Tolerance scores.
To address potential multicollinearity among predictors and to simplify the model without compromising informational value, a dimensionality reduction step was incorporated using Principal Component Analysis (PCA). This dimensionality reduction aimed to consolidate performance indicators into a single composite feature to streamline model training and interpretation.
In the model development stage, five supervised learning algorithms were trained to classify binary decision outcomes: Logistic Regression, Random Forest, Support Vector Machines (SVMs), Neural Networks, and XGBoost. These models were subsequently evaluated in the final validation phase based on classification performance metrics including accuracy, precision, recall, F1-Score, and ROC-AUC (Receiver Operating Characteristic–Area Under the Curve).
To classify economic decision outcomes in edge-enhanced environments, we employed five supervised machine learning algorithms. Each of these models offers different mechanisms for learning patterns in the data and handling non-linear relationships, noise, and feature interactions. Below, we briefly describe the technical foundation of each algorithm.
Logistic Regression (LR) is a linear probabilistic model used for binary classification problems [
41]. It estimates the probability
by applying the logistic function to a linear combination of output features [
42], according to Equation (5):
In Equation (5), is the model intercept, and are the regression coefficients associated with each feature .
LR provides interpretability and a strong baseline for classification, especially when the relationship between features and the target variable is approximately linear [
43].
The Support Vector Machine with RBF Kernel (SVM-RBF) constructs a hyperplane that separates data points of different classes by maximizing the margin between them [
42]. The radial basis function (RBF) kernel allows the model to capture non-linear relationships by projecting data into a higher-dimensional space, according to Equation (6):
Random Forest (RF) is an ensemble learning method that builds a collection of decision trees using bootstrapped samples and random feature selection. The final prediction is obtained through majority voting (for classification). RF is robust to overfitting and handles feature interactions well [
44]:
where, in Equation (7),
represents the
decision tree’s prediction.
Extreme Gradient Boosting (XGBoost) is a highly efficient and scalable gradient boosting framework that sequentially adds decision trees to minimize a regularized objective function [
45] (8):
where, in Equation (8),
penalizes the model’s complexity. XGBoost supports missing values, handles imbalance well, and includes built-in regularization, making it ideal for real-time classification under uncertainty. Also,
is the loss function, and
and
are the hyperparameters that control regularization.
represents the minimum loss reduction required to make a further partition on a tree node. A higher
value leads to fewer splits, thus producing simpler trees and reducing the overfitting risk. In this study,
was set to 0.2, which encourages the model to create splits only when they yield meaningful improvements in predictive accuracy, leading to more generalizable trees in noisy, synthetic data environments.
corresponds to the L2 regularization term on leaf weights. This penalizes extreme weight values and stabilizes the model when dealing with multicollinearity or outlier-sensitive data. A value of
was selected to maintain moderate regularization strength, ensuring that the model remains flexible while avoiding excessive sensitivity to specific features or noise. Together, these parameters contribute to a better trade-off between bias and variance, ultimately enhancing the robustness of the XGBoost classifier in edge-enhanced decision scenarios.
A Neural Network (NN) is a feed-forward multilayer perceptron (MLP) that is trained via backpropagation. It consists of an input layer, one or more hidden layers, and an output layer with a sigmoid activation function for binary classification [
46]. The forward pass computes, according to Equation (9):
where, in Equation (9),
and
are weight matrices,
and
are biases,
represent the sigmoid activation function for binary classification, and
is the Rectified Linear Unit activation function [
47].
All machine learning algorithms were implemented using the Python 3.10 programming language. The primary libraries utilized were Scikit-learn (for Logistic Regression, Support Vector Machine, Random Forest), XGBoost (for gradient boosting), and TensorFlow/Keras (for the construction and training of the Neural Network model). The LR model was implemented using the <<LogisticRegression>> class from the <<sklearn.linear_model>> module, with default settings for binary classification. The SVM-RBF kernel was implemented using the SVC class from <<sklearn.svm>>, with the kernel parameter set to “rbf” and the gamma parameter was tuned via internal scaling. The RF classifier was implemented using the <<RandomForestClassifier>> class from <<sklearn.ensemble>>, with 100 trees and a fixed random seed to ensure reproducibility. The XGBoost model was implemented using the <<XGBClassifier>> class from the <<xgboost>> package. Parameters such as the learning rate, the maximum tree depth, and the number of estimators were fine-tuned to optimize predictive performance. The NN model was built as a feed-forward MLP using the Sequential API from the <<TensorFlow/Keras>> library. The network architecture consisted of one hidden layer with ReLU activation and one output layer with sigmoid activation for binary classification.
The confusion matrix is a commonly used diagnostic tool in classification tasks, capturing the distribution of predicted versus actual class labels. It is applicable to both binary and multiclass classification problems. Unlike aggregate metrics such as overall accuracy, the confusion matrix allows for a more granular analysis by highlighting how well each individual class is predicted. This is particularly relevant when the model has difficulty with specific classes, which might not be reflected in overall accuracy results. For instance, a model may consistently misclassify a particular category, a limitation that standard accuracy metrics cannot reveal [
48,
49]. According to
Table 1, the confusion matrix consists of four core components: (i) TP—correctly predicted positive instances; (ii) FP—negative instances incorrectly predicted as positive; (iii) FN—positive instances incorrectly predicted as negative; and (iv) TN—correctly predicted negative instances.
This structure enables a detailed breakdown of classification performance, serving as the foundation for several derived metrics such as accuracy (10), recall (11), precision (12), and F1 Score (13), which are used to assess model robustness across different types of error. The evaluation metrics used to assess classification models are computed based on the components of the confusion matrix [
48,
50].
Before fitting the classification models, it is essential to assess multicollinearity among the independent variables, as high intercorrelations can distort coefficient estimates, reduce model interpretability, and inflate variance. Two widely used diagnostic measures for multicollinearity are the VIF and Tolerance. The VIF quantifies how much the variance of a regression coefficient is inflated due to multicollinearity. It is calculated for each predictor
as follows:
where, in Equation (10),
is the coefficient of determination obtained when
is regressed on all other independent variables. A VIF value of 1 indicates no correlation with other variables, whereas values above 5 (and especially above 10) are often interpreted as indicative of problematic multicollinearity [
51,
52].
The Tolerance is the reciprocal of VIF (Equation (15)):
Low Tolerance values (typically
) suggest that the predictor is highly collinear with others and may be redundant in the model [
52,
53].
These diagnostics were applied to all input variables prior to model estimation. Based on the computed VIF and Tolerance values, variables exhibiting high multicollinearity were either excluded or transformed using PCA to ensure model robustness and avoid instability during training. PCA is an unsupervised statistical method that transforms the original correlated variables into a new set of uncorrelated variables called principal components, which are linear combinations of the original variables. The first few principal components capture most of the variance in the data, allowing for a more compact and stable representation of the feature space. Mathematically, the transformation is defined as follows:
where, in Equation (16),
is the
kth principal component, and
are the eigenvectors of the covariance matrix of the original variables
.
To assess the performance and generalizability of the classification models, the dataset was randomly split into training (80%) and testing (20%) subsets [
48,
54]. All models were trained on the training set and evaluated on the unseen test set to ensure a fair assessment of the generalization capability. The evaluation was conducted using both threshold-dependent and threshold-independent performance metrics derived from the confusion matrix, including accuracy, precision, recall, F1-Score, and ROC-AUC.
These metrics offer complementary perspectives: while accuracy provides a global measure of correctness, precision and recall are particularly useful in understanding the balance between false positives and false negatives. The F1-Score captures the trade-off between the two, and the ROC-AUC reflects the model’s discriminative power across all classification thresholds.
4. Results
In practical terms, the problem addressed in this study relates to how financial and economic institutions can respond rapidly and effectively to fast-moving, uncertain events, such as market turbulence or financial contagion, where milliseconds can determine substantial gains or losses. Traditional decision-making architectures, often relying on centralized data processing, may introduce latency that compromises responsiveness and risk mitigation. In contrast, edge computing enables data to be processed near their source, thus reducing latency and enhancing local autonomy in decision-making. In this context, the real-time decision quality becomes a proxy for system resilience. This study models the problem by simulating how variations in the edge system performance, such as processing delays, throughput, and resource efficiency, can impact the accuracy and reliability of economic decisions, with implications for systemic stability.
4.1. Simulation Setup and Experimental Configuration
All simulations and model training were conducted in Python 3.10 using the Google Colab environment. Additionally, the full source code, including preprocessing, feature engineering, model training, and evaluation steps, is openly available on GitHub at [
https://github.com/IonutNica/Edge_Computing, accessed on 1 May 2025] to facilitate replication and transparency. The implementation relied on widely used machine learning libraries: Scikit-learn (for Logistic Regression, Support Vector Machine, and Random Forest), XGBoost (for gradient boosting), and TensorFlow/Keras (for Neural Network modeling). The dataset includes synthetic observations simulating real-time economic behavior and edge infrastructure dynamics. The data were split randomly into training (80%) and testing (20%) subsets to ensure fair evaluation of generalization performance.
Logistic Regression was implemented using the <<LogisticRegression>> class from <<sklearn.linear_model>>, with default settings for binary classification.
The Support Vector Machine with RBF kernel used the SVC(kernel=”rbf”) class from <<sklearn.svm>>, with <<gamma=“scale”>> for automatic kernel coefficient adjustment.
Random Forest was implemented via <<RandomForestClassifier>> from <<sklearn.ensemble>>, using <<n_estimators = 100>> and a fixed random seed to ensure reproducibility.
XGBoost was developed using the XGBClassifier from the xgboost library, with tuned hyperparameters: <<max_depth = 6>>, <<learning_rate = 0.1>>, <<lambda = 1>>, <<gamma = 0.2>>, and <<n_estimators = 150>>.
The Neural Network was constructed using the Sequential API from <<tensorflow.keras.models>> consisting of one hidden layer with 32 neurons (ReLU activation) and one output layer with sigmoid activation. It was compiled with binary cross-entropy loss and trained over 50 epochs with a batch size of 16.
Model performance was evaluated on the test set using standard classification metrics: accuracy, precision, recall, F1-Score, and ROC-AUC.
4.2. Descriptive Statistics
Following the data preparation and feature engineering steps described in the previous section, a sequence of statistical and machine learning analyses was conducted to uncover patterns and predictive signals in real-time economic decision-making supported by edge computing. The results are structured in several stages: exploratory data analysis and descriptive statistics; correlation and multicollinearity assessments; dimensionality reduction via PCA; and finally, predictive modeling using various supervised learning algorithms.
Table 2 presents the descriptive statistics of the dataset used to evaluate the interplay between the edge computing performance and economic decision-making under risk conditions. The transaction volume displays a high variability (mean = 5591.86; std. dev. = 2523.18), suggesting substantial fluctuations in economic activity across instances, which may indicate varying system loads or market demands in different real-time contexts. The market behavior index shows a mean value near zero (
0.006) with a relatively balanced distribution (std. dev. = 0.595), indicating that market dynamics oscillate around an equilibrium, with no significant trend toward bullish or bearish conditions. The financial metric exhibits a positively skewed distribution (mean = 111.21; median = 71.27), reflecting the existence of high outliers in financial intensity across scenarios. This may suggest that some decisions are associated with extreme financial performance outcomes. The edge processing latency varies widely (mean = 164.40 ms; std. dev. = 149.76), highlighting inconsistent processing delays at the edge level, a key concern for time-sensitive economic decisions. Resource utilization and workload distribution efficiency demonstrate a discrepancy in scale. While utilization is normalized (mean = 0.55), efficiency varies more widely (mean = 101.00), indicating that higher workload distribution efficiency is not necessarily associated with proportional resource use. Decision accuracy centers around a high average (73.36%) but shows significant variance, which could reflect instability in the decision-support system under changing computational or market stress. The system throughput values are expressed in normalized units, scaled to simulate microsecond-level data transmission rates typically encountered in edge-computing environments. Its interpretation should be relative (e.g., in ratios), rather than absolute. Transaction efficiency, a ratio between throughput and volume, shows small values, which are consistent with the system throughput scale, but indicate how well the infrastructure handles the data load. The latency per transaction (mean = 0.040) indicates an average of 40 ms per transaction, highlighting edge computing’s potential in real-time scenarios. The utilization to efficiency ratio (mean = 0.023) shows low correspondence between resource use and workload efficiency, suggesting optimization issues in the edge architecture. The decision quality Score, computed as the product of decision accuracy and decision outcome, has a right-skewed distribution (mean = 34.83; median = 7.33), implying that while most decisions are moderate or poor in quality, a few reach very high performance.
4.3. Correlation, Exploratory Analysis and Multicollinearity Diagnostics
The Pearson correlation heatmap from
Figure 2 reveals several insightful relationships between technological and economic indicators in the dataset.
Notably, transaction efficiency is negatively correlated with the transaction volume (0.66) and the latency per transaction (0.47), suggesting that increased transaction loads and processing delays hinder efficiency. A strong positive correlation (0.88) exists between the decision outcome and the decision quality score, validating the constructed metric. Additionally, system throughput correlates moderately with both the transaction efficiency (0.53) and the latency per transaction (0.42), emphasizing its role in real-time performance. Most variables exhibit weak pairwise associations, implying low multicollinearity and diverse influences on decision outcomes, suitable for robust multivariate modeling.
According to
Figure 3, the Kendall correlation heatmap provides a non-parametric assessment of the monotonic relationships among the variables. It confirms several key findings from the Pearson analysis, but with subtle differences in strength. The decision quality score remains strongly correlated with the decision outcome (0.81), reinforcing the reliability of this composite indicator. A moderate positive correlation exists between the edge processing latency and the latency per transaction (0.71), and between resource utilization and the utilization to efficiency ratio (0.56), indicating consistent directional trends in these dimensions. Transaction efficiency exhibits a negative correlation with transaction volume (
0.50), suggesting that efficiency diminishes as the volume increases. Overall, Kendall’s tau supports the robustness of relationships detected via the Pearson analysis while being less sensitive to outliers and distributional assumptions, making it a valuable complement in evaluating dependencies among real-time economic indicators.
Figure 4 presents both the histogram and the boxplot of the standardized transaction volume. The histogram indicates a relatively uniform distribution with no strong skewness or pronounced peaks, suggesting a balanced spread of values across the sample. The boxplot reinforces this observation, showing no significant outliers and a symmetric distribution around the median. This stable pattern supports the reliability of transaction volume as a consistent metric in the analysis of economic decisions facilitated by edge computing.
The visualization of the market behavior index reveals a nearly symmetric and uniform spread across its range, according to
Figure 5. The histogram displays a slight bimodal tendency, suggesting some variation in behavioral clustering within the market. Meanwhile, the boxplot indicates a well-balanced distribution with no extreme outliers, and a median that aligns closely with the center of the interquartile range. This implies that market sentiment, as captured by this index, fluctuates consistently around a central value, making it a stable input for real-time economic modeling based on edge computing.
Figure 6 provides insights into the distribution of the financial metric variable. The histogram shows a left-skewed distribution, indicating that lower financial metric values are more frequent in the dataset. This asymmetry suggests that most economic scenarios captured by the system are associated with lower financial outputs or returns. The boxplot further confirms this skewness, with the median located closer to the lower quartile and a longer right tail. These characteristics may reflect the volatility or imbalance in financial outcomes within edge-enabled decision-making environments.
Figure 7 displays the distribution of the edge processing latency variable, highlighting a strongly right-skewed pattern in the histogram. This indicates that most latency values are low, while a smaller number of cases experience significantly higher delays. The boxplot reinforces this observation, showing a concentration of data on the lower end with a long right whisker, suggesting the presence of high-latency outliers. Such variability in processing time may reflect uneven computational loads across edge nodes, which could affect real-time decision-making efficiency in edge-enabled economic systems.
The distribution of the resource utilization variable, shown in
Figure 8, shows an approximately symmetric pattern centered around the mean, indicating a consistent and balanced use of computational resources across the edge computing infrastructure. The histogram indicates that resource utilization values are well dispersed without strong skewness, while the boxplot confirms the absence of significant outliers. This balanced spread implies the efficient allocation of resources within the network, which is essential for maintaining system stability and performance in real-time economic environments.
The observed left-skewness in the distribution of workload distribution efficiency from
Figure 9 indicates that, in several instances, computational tasks are unequally balanced across the edge infrastructure. Such inefficiencies may lead to localized system slowdowns or bottlenecks, which in turn hinder the system’s capacity to process economic data promptly. In high-volatility contexts, this latency can delay or change critical financial decisions, amplifying systemic vulnerabilities and increasing the risk of financial contagion propagation across interconnected institutions or markets.
Figure 10 shows that decision accuracy values are relatively uniformly distributed, with no extreme skewness or pronounced outliers. This suggests that the system maintains a balanced performance in terms of decision-making, with a substantial proportion of observations concentrated around the median. Such consistency is fundamental in real-time economic environments, where the reliability of automated or assisted decisions under edge computing architectures directly impacts risk management and systemic stability.
Figure 11 illustrates the distribution of the system throughput variable, which appears approximately uniform and symmetric across its range. The boxplot confirms the absence of significant outliers, suggesting stable data flow rates within the edge computing environment. This stability is decisive in real-time decision-making systems, as it ensures consistent performance and low latency, factors that are essential for timely responses during financial stress or contagion propagation scenarios.
As depicted in
Figure 12, the variable decision outcome follows a perfectly binary distribution, with values clustered exclusively at
1 and 1, corresponding to the two possible decision outcomes. The histogram highlights an almost equal frequency of both outcomes, while the boxplot confirms the absence of intermediate values or variability. This binarized structure is essential for classification tasks such as Logistic Regression, enabling clear evaluation of the predictive performance of decision-making systems under edge computing conditions.
Figure 13 illustrates the distribution of latency per transaction, showing a clear right-skewed pattern where most values are concentrated toward the lower latency end. This distribution supports the premise that reduced latency can enhance real-time responsiveness, which is critical in economic contexts where rapid decision-making is essential to prevent the amplification of systemic risks such as financial contagion. The observed variation, including occasional higher latencies, provides a useful framework for assessing the stability and resilience of edge-assisted economic infrastructures.
The distribution of the utilization-to-efficiency ratio from
Figure 14 is notably left-skewed, with a concentration of observations at lower ratio values. This pattern may reflect an operational scenario in which high efficiency is maintained despite conservative or modest resource usage, an important attribute in edge-based economic environments aiming to balance cost and responsiveness. From the perspective of economic decision-making, such distributions can suggest systems that are structurally optimized to mitigate overloads and ensure continuity during periods of financial stress or contagion transmission, underscoring the strategic value of resource calibration in real-time infrastructures.
Figure 15 illustrates a bimodal distribution for the decision quality score, with peaks near the extremes, suggesting a polarization in decision outcomes, either highly effective or largely ineffective. This pattern highlights the sensitivity of real-time economic systems to both system-level computational performance and contextual economic volatility. Such distributional traits are especially relevant in financial contagion scenarios, where abrupt shifts in accuracy and outcomes can propagate across interconnected agents or nodes, amplifying systemic risk. The variation in decision quality emphasizes the need for adaptive edge-driven architectures that can stabilize performance, even in uncertain environments.
Table 3 presents the Variance Inflation Factor (VIF) and the corresponding Tolerance values for each predictor variable used in the regression analysis. VIF values above 5 typically signal potential multicollinearity concerns which may distort coefficient estimates and reduce model interpretability. Notably, transaction efficiency (VIF = 6.18), decision accuracy (VIF = 5.96), and edge processing latency (VIF = 5.55) exceed this threshold, indicating a high degree of linear dependency with other variables in the model. These findings suggest that additional variable selection, dimensionality reduction, or regularization techniques may be necessary to ensure model stability, especially in scenarios that simulate real-time decision-making under interconnected economic and computational conditions.
4.4. Dimensionality Reduction via PCA
Most variables demonstrated acceptable multicollinearity thresholds, although indicators such as decision accuracy, edge processing latency, and transaction efficiency exhibited VIF values exceeding 5, signaling moderate to high multicollinearity. To mitigate this, we applied PCA on these three variables and extracted a single composite feature labeled performance PCA, which retained 96.46% of the total variance according to
Figure 16. This new feature encapsulates the essential variation across performance-critical indicators, ensuring dimensionality reduction while preserving relevant information.
To further interpret the nature of this component,
Figure 17 presents the feature loadings on the first principal component. The results clearly indicate that the edge processing latency dominates the composite dimension (loading ≈ 0.9999), while the contributions of decision accuracy (≈0.0087) and transaction efficiency (≈0.0000) are negligible. This distribution confirms that edge latency dominates the performance variation captured by the PCA score. Nonetheless, retaining the PCA-based composite variable helps maintain a unified representation of performance in the predictive models and ensures consistency with the multicollinearity diagnosis based on the VIF. This supports the validity of using the PCA-based score as a robust proxy for real-time system responsiveness in subsequent predictive models.
Figure 18 illustrates the distribution of the composite performance scores obtained through PCA, grouped by decision outcomes. The results reveal distinct distributional characteristics between the two groups. Although both “Negative” and “Positive” decisions share comparable interquartile ranges, their median values differ slightly, suggesting a potential performance threshold that separates decision outcomes. The spread of the PCA scores is notably wider for the “Positive” outcomes, indicating greater variability in the underlying performance features when the decisions were successful. This variability could reflect a more dynamic and responsive computational environment supporting effective economic decisions. The figure supports the hypothesis that performance-related features, such as latency, decision accuracy, and transaction efficiency, aggregated via PCA, are meaningfully associated with the quality of economic decision-making in edge computing contexts.
To ensure the robustness of all of the machine learning models applied in this study, multicollinearity among predictor variables was evaluated using the Variance Inflation Factor (VIF) and Tolerance statistics. As presented in
Table 4, all of the VIF values were below the commonly accepted threshold of 5, and the Tolerance values were well above 0.1. Although all of the variables presented in
Table 4 pass the multicollinearity diagnostics, specifically, with all VIF values below 5 and Tolerance scores above 0.1, the variable decision quality score was excluded from the subsequent machine learning models. Despite its acceptable multicollinearity metrics, this variable is mathematically constructed using the target variable (decision outcome), which could introduce data leakage and artificially inflate the model’s predictive performance. Therefore, to ensure model integrity and avoid circular reasoning, the decision quality score was retained only for exploratory and descriptive analysis.
4.5. Predictive Modeling and Performance Evaluation
Table 5 provides a comparative overview of various machine learning models evaluated for predicting binary decision outcomes in the context of real-time economic systems enhanced by edge computing.
The classification models were evaluated using five standard performance metrics: accuracy, precision, recall, F1-Score, and ROC-AUC. Accuracy measures the overall proportion of correct predictions across both classes and provides a general assessment of model correctness. precision reflects the proportion of predicted positive decisions that are truly positive, indicating how reliable the model is when signaling a favorable economic decision. Recall (also known as sensitivity) measures the model’s ability to correctly identify all actual positive decisions, which is essential for minimizing missed opportunities in high-impact contexts. The F1-Score is the harmonic mean of precision and recall, balancing both metrics when there is a trade-off between false positives and false negatives. Finally, the ROC-AUC provides a threshold-independent measure of the model’s discriminative ability, summarizing how well the model distinguishes between positive and negative decision outcomes across different probability cutoffs.
Among the evaluated classifiers, XGBoost achieved the strongest overall performance, with 97% accuracy, 1.00 precision, 0.94 recall, 0.97 F1-Score, and an ROC-AUC of 0.998. This consistent superiority across all performance metrics highlights XGBoost’s ability to model the complex, nonlinear relationships that characterize real-time, edge-enabled economic decision environments. RF closely followed, also reaching high performance levels (accuracy = 95%, ROC-AUC = 0.997), while the NN demonstrated solid but slightly lower scores (e.g., ROC-AUC = 0.965). In contrast, LR, traditionally used as a baseline model, showed the weakest performance, with only 64% accuracy and an ROC-AUC of 0.663, indicating a limited capacity to capture the underlying nonlinear dynamics. Even Support Vector Machines, which are typically robust to noise and effective in high-dimensional settings, failed to outperform XGBoost, scoring only 77% on all core metrics. These results confirm that the proposed framework, based on engineered edge-computing indicators and XGBoost, not only outperforms traditional models statistically but also offers greater reliability in real-time risk-sensitive applications, such as the early detection of financial instability or contagion propagation.
Table 6 presents the confusion matrix for the Random Forest model. Out of all of the actual positive decision outcomes, 46 were correctly classified (true positives), while five were incorrectly predicted as negative (false negatives), indicating a small but non-negligible rate of missed opportunities. All of the negative cases were correctly identified (true negatives = 49), and there were no false positives, demonstrating excellent specificity. This outcome suggests that the Random Forest model is highly reliable for identifying unfavorable decisions, while still maintaining a strong ability to detect positive ones.
Table 7 displays the confusion matrix for XGBoost, which shows further improved performance. It correctly classified 48 positive cases and 49 negative cases, with only three false negatives and no false positives. This perfect precision (no FPs) combined with very high recall (only three FNs) is particularly valuable in risk-sensitive environments, where false alarms must be minimized and undetected risks (FNs) can propagate into systemic failures. From a risk management perspective, this result highlights the robustness of XGBoost in ensuring both accuracy and early detection of potentially dangerous decision scenarios.
In contexts such as financial contagion prevention, the trade-off between false positives (resource misallocation) and false negatives (undetected shocks) is critical. XGBoost’s ability to eliminate false positives and minimize false negatives offers a decisive advantage for real-time, edge-enhanced financial decision systems. By contrast, although RF performs very well, the existence of five false negatives could translate into unflagged risks in practice.
This result provides valuable insights into how edge-enabled decision frameworks can support the design of adaptive and resilient economic infrastructures in peripheral regions. By integrating low-latency, high-precision computational mechanisms (as demonstrated through the performance of XGBoost), local institutions can significantly enhance their capacity to identify early signals of economic stress and respond with greater autonomy. This technological augmentation aligns with the principles of place-based policy, promoting decentralized, context-sensitive solutions that reinforce systemic stability at the regional level. Furthermore, by minimizing false positives and negatives in decision classification, the proposed framework contributes to more efficient resource allocation, reducing vulnerability to financial contagion and supporting long-term socioeconomic resilience in underdeveloped or structurally fragile territories.
5. Discussion
In this study, ML algorithms were applied to predict economic decision outcomes using a comprehensive set of features derived from real-time edge computing environments. The dependent variable, the decision outcome, is a binary indicator that reflects whether a decision was classified as positive or negative. This classification task aligns naturally with supervised learning methods, where the goal is to learn underlying patterns in the data that contribute to decision performance. The independent variables used in the models encompass both economic metrics (e.g., the transaction volume, the financial metric, and the market behavior index) and technological performance indicators (e.g., resource utilization, edge processing latency, system throughput, and performance PCA). These variables capture the hybrid nature of the system, integrating edge infrastructure dynamics with economic behavioral signals. The core purpose of employing ML models is threefold: (i) predictive decision support; (ii) systemic risk mitigation; and (iii) feature importance assessment. From a strategic perspective, the ML algorithms serve a broader analytical role in (i) evaluating decision efficiency under real-time constraints, especially in scenarios with potential systemic instability; (ii) identifying signals indicative of suboptimal or risky decisions, which may propagate across interconnected financial or operational networks; and (iii) supporting edge-based infrastructure monitoring, where fluctuations in latency or throughput might lead to degraded decision quality and increased exposure to contagion risk. Moreover, the comparative model performance shows that more complex algorithms, such as XGBoost and Random Forest, significantly outperform traditional Logistic Regression in predictive accuracy and ROC-AUC. This suggests that nonlinear relationships and interaction effects between variables play a substantial role in shaping decision outcomes, an insight highly relevant for policy designers and infrastructure engineers alike.
The increasing reliance on machine learning algorithms and edge computing in financial decision-making raises several ethical concerns, particularly in high-stakes environments where misclassifications may result in systemic harm. Automated models, while efficient, may perpetuate biases embedded in training data or overlook qualitative contextual factors that human judgment would otherwise consider. In real-time financial ecosystems, this could lead to the unfair exclusion of vulnerable actors, the misallocation of capital, or the amplification of existing inequalities.
Furthermore, the opacity of complex models such as XGBoost and Neural Networks may limit transparency and explainability, posing challenges for accountability and regulatory compliance. Ensuring that such systems are interpretable, auditable, and aligned with the principles of fairness and inclusivity is critical, especially when deployed in sectors that affect public welfare and economic stability.
As such, any deployment of edge-enhanced decision systems in financial contexts must be accompanied by robust ethical oversight frameworks, including algorithmic audits, transparency protocols, and mechanisms for human-in-the-loop validation. These safeguards are essential for building trust, avoiding unintended consequences, and ensuring that technological advancements contribute positively to equitable and resilient economic development.
These findings have significant economic implications, particularly in the context of financial contagion and systemic risk propagation. In highly interconnected economic ecosystems, decisions made under suboptimal technological or informational conditions can lead to cascading effects, not only within a single system but across multiple institutions or sectors. The ability to predict and interpret decision outcomes in real-time thus becomes a critical capability for early warning systems and crisis prevention frameworks.
The integration of performance metrics derived from edge computing (e.g., latency, utilization, and throughput) into predictive economic models offers a novel approach to understanding operational fragility. For instance, elevated latency per transaction or inefficient resource-to-efficiency ratios may signal early signs of stress in the digital infrastructure, potentially correlating with diminished decision accuracy or higher error rates in automated financial systems. These inefficiencies, if left undetected, can amplify exposure to financial contagion, especially in markets where high-frequency transactions and automated decisions dominate.
Moreover, from a macroeconomic perspective, the results suggest that system-level technical variables are not just peripheral operational concerns but can serve as leading indicators of economic disruption. Edge computing environments, due to their decentralized and adaptive nature, provide a rich stream of granular data that can be used to map emerging vulnerabilities in near real-time. Thus, incorporating such data into machine learning frameworks enhances our ability to forecast not only individual decision failures but also the broader structural risks they may entail.
6. Conclusions
This study examined the intersection of edge computing performance and economic decision-making in real-time environments, with particular attention to the mitigation of financial contagion risk. By constructing a dataset that integrates economic behavior signals with edge infrastructure metrics, and applying multiple machine learning models, we addressed three key research questions.
Regarding RQ1, we successfully designed four composite indicators, transaction efficiency, the latency per transaction, the utilization-to-efficiency ratio, and the decision quality score, that effectively capture the interplay between system performance and economic decision dynamics. These indicators allowed for granular modeling of computational and economic interactions.
For RQ2, correlation analysis and PCA revealed significant and interpretable relationships between edge performance (e.g., latency, resource use) and decision outcomes, highlighting that low latency and balanced resource utilization are associated with more favorable economic decisions.
With respect to RQ3, our results demonstrated that ensemble-based models such as XGBoost and Random Forest outperformed traditional classifiers, achieving high levels of predictive accuracy (up to 97%) and excellent ROC-AUC scores. These findings underscore the effectiveness of ML-based architecture in forecasting and managing decision risk under edge-enhanced conditions.
The economic implications of this research are multifaceted. At the micro level, localized decision-making enabled by edge computing can reduce reaction times and mitigate error propagation in automated economic processes. At the macro level, edge performance metrics may serve as early warning signals for systemic fragility, supporting the development of real-time monitoring tools and adaptive response systems.
In the context of financial contagion, our results suggest that technological inefficiencies, such as uneven workload distribution or delayed edge response, can indirectly influence the transmission of shocks across economic agents. Thus, integrating edge monitoring into financial infrastructure design could provide a new layer of resilience against cascading failures.
While this study did not simulate financial contagion events explicitly, the strong association between improved edge system performance (e.g., reduced latency, balanced resource utilization) and higher decision quality suggests a potential pathway for mitigating contagion risk. In complex economic networks, faster and more accurate local decisions can prevent the amplification of shocks across interconnected agents. Future research could build on this foundation by modeling systemic contagion dynamics under varying edge performance scenarios, offering more direct empirical validation of this critical link.
Based on the findings of this study, several actionable recommendations can be formulated to enhance economic resilience and improve decision-making accuracy in edge-enabled environments:
Integrate edge monitoring into financial regulation frameworks: Regulators should require financial institutions to monitor and report key edge performance metrics (e.g., processing latency, system throughput, and resource efficiency) as part of operational risk assessments. These indicators can serve as early warning signals for systemic stress or contagion risk;
Promote edge infrastructure standardization in financial ecosystems: Establishing common standards for edge computing performance and data interoperability in financial platforms will ensure consistency and reliability across institutions, especially during high-volatility periods;
Assign stress testing for digital infrastructure under simulated contagion scenarios: Financial institutions should be required to perform digital stress tests simulating edge node failures or high-latency conditions to understand potential vulnerabilities in decentralized environments;
Incentivize decentralized and adaptive risk management solutions: Governments and central banks should support the development of edge-based decision-support systems that can adapt in real time to economic shocks, improving reaction speed and local autonomy in decision-making.
Despite these contributions, several limitations must be acknowledged. First, the dataset used was synthetic and simulated, which may not fully reflect the complexity of real-world market behavior. Second, the analysis focused on binary outcomes and did not explore sequential decision processes or temporal feedback loops. It is important to acknowledge that the synthetic dataset, although carefully constructed to reflect key features of real-time economic dynamics, may not capture the full complexity, heterogeneity, and emergent behavior characteristic of real-world systems. As such, the generalizability of the results must be interpreted cautiously. Future studies should aim to validate these findings using empirical datasets derived from operational financial systems or IoT-enabled economic infrastructures.
Furthermore, while this study focused on binary decision outcomes (positive or negative), real-world economic decision-making often involves multi-class or continuous outcomes, representing a richer spectrum of financial states and risk levels. Generalizing the current framework to multiclass classification, ordinal decision outcomes, or sequential decision processes (e.g., using reinforcement learning) would improve applicability and realism in dynamic economic environments.
In addition, scalability remains a critical concern, as deploying resource-intensive machine learning models (such as XGBoost) across heterogeneous and constrained edge devices may encounter technical limitations. Computational costs associated with real-time inference could impact system responsiveness and energy efficiency. Moreover, although XGBoost achieved the highest predictive performance, its lower interpretability compared to simpler models (e.g., Logistic Regression) could pose challenges for regulatory compliance, auditability, and stakeholder trust. Future research should explore lightweight model alternatives (e.g., model pruning or distillation), scalable distributed edge architectures, and explainable AI techniques (e.g., SHAP values) to overcome these challenges. These limitations open valuable avenues for future research at the intersection of edge computing, economic decision-making, and financial risk management.
Also, future research could address these limitations by (i) validating findings on real-time data from financial markets or IoT-based economic systems; (ii) extending the framework to multiclass or time-series classification tasks; (iii) incorporating reinforcement learning or adaptive edge optimization algorithms for dynamic resource management; and (iv) exploring cross-sectoral contagion mechanisms beyond the financial domain, such as supply chain disruptions or energy markets.
Overall, this study provides a robust foundation for understanding how edge computing can shape economic decision architectures and offers actionable insights into the design of intelligent, real-time financial systems in an increasingly connected and uncertain world.