Next Article in Journal
A Variational-Mode-Decomposition-Cascaded Long Short-Term Memory with Attention Model for VIX Prediction
Previous Article in Journal
Enhancing Seismic Repairability of Precast RC Frames Through an Innovative Replaceable Plastic Hinge Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Deep Learning-Based Load Forecasting Model for Logical Range

1
School of Computer Science and Engineering, Xi’an University of Technology, Xi’an 710048, China
2
Shaanxi Key Laboratory for Network Computing and Security Technology, Xi’an 710048, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(10), 5628; https://doi.org/10.3390/app15105628
Submission received: 15 April 2025 / Revised: 10 May 2025 / Accepted: 16 May 2025 / Published: 18 May 2025

Abstract

:
The Logical Range is a mission-oriented, reconfigurable environment that integrates testing, training, and simulation by virtually connecting distributed systems. In such environments, task-processing devices often experience highly dynamic workloads due to varying task demands, leading to scheduling inefficiencies and increased latency. To address this, we propose GCSG, a hybrid load forecasting model tailored for Logical Range operations. GCSG transforms time-series device load data into image representations using Gramian Angular Field (GAF) encoding, extracts spatial features via a Convolutional Neural Network (CNN) enhanced with a Squeeze-and-Excitation network (SENet), and captures temporal dependencies using a Gated Recurrent Unit (GRU). Through the integration of spatial–temporal features, GCSG enables accurate load forecasting, supporting more efficient resource scheduling. Experiments show that GCSG achieves an R 2 of 0.86, MAE of 4.5, and MSE of 34, outperforming baseline models in terms of both accuracy and generalization.

1. Introduction

In recent years, enhancing the capability of test and training operations under integrated land, sea, air, space, and electromagnetic warfare environments has become a focal point in military competition. To support such complex missions, the concept of the Logical Range has emerged. A Logical Range is a mission-oriented, reconfigurable range system that integrates testing, training, and simulation resources through distributed networking and high-performance computing technologies [1,2]. Unlike traditional test ranges, Logical Ranges can dynamically reorganize functional elements across geographically dispersed platforms to meet the demands of new tasks that a single range cannot fulfill alone.
This flexible architecture enables cross-domain resource integration and real-time mission composition, but it also presents new technical challenges. During joint operations, such as cross-regional flight testing or collaborative weapon system evaluation, Logical Range platforms often suffer from uneven workload distribution, high concurrency, and complex system heterogeneity. These factors can cause delays in task scheduling and execution, leading to reduced responsiveness and efficiency. Consequently, accurate load forecasting becomes crucial for optimizing task allocation and maintaining system stability under fluctuating conditions.
Although traditional statistical approaches [3,4,5,6] have been applied for time-series load prediction, they typically rely on strict assumptions (e.g., Poisson distribution) and fail to handle the stochastic, nonlinear nature of real-world Logical Range workloads [7]. To overcome these limitations, more adaptable, learning-based forecasting frameworks are needed to support intelligent resource scheduling in such dynamic environments.
To capture temporal patterns in such data, autoregressive models and probabilistic statistical approaches [8,9,10,11] have been employed with some success. However, in dynamically evolving Logical Range environments, their predictive performance often deteriorates due to their limited adaptability to abrupt changes and temporal nonstationarity [12].
Recently, deep learning-based prediction models—notably those based on recurrent neural networks (RNNs)—have demonstrated strong capabilities in learning nonlinear temporal dependencies directly from historical data [13,14,15,16,17,18]. Nevertheless, many of these approaches fail to account for spatial correlations that may exist across subsystems or geographically distributed nodes. Recent studies have highlighted the importance of jointly modeling both spatial and temporal features, which can significantly improve the accuracy of forecasting in nonstationary and highly volatile load environments [19].
To address the aforementioned challenges, we propose GCSG: a novel hybrid deep learning model that integrates Gramian Angular Field (GAF) encoding, a Convolutional Neural Network with Squeeze and Excitation (CNN-SENet), and a Gated Recurrent Unit (GRU). This model is specifically designed to handle the complexity of load forecasting in Logical Range environments, offering several key innovations:
First, we employ the GAF approach to transform 1D nonstationary load sequences into 2D images, enabling CNNs to effectively capture spatial correlations. Second, we develop a CNN-SENet module that adaptively recalibrates channel-wise feature responses, enhancing the model’s ability to focus on informative features while suppressing noise, which is crucial for robust feature extraction in complex environments. Finally, we introduce an attention-based spatiotemporal fusion mechanism that dynamically combines spatial features (from CNN-SENet) with temporal dependencies (from GRU), significantly improving the prediction accuracy under highly dynamic workloads by capturing both transient patterns and long-term trends in the data.
The main contributions of this study are as follows:
(1)
We propose a domain-specific hybrid forecasting model tailored to the Logical Range context. Combining spatial and temporal representation learning through a GAF–CNN-SENet–GRU architecture, our approach effectively addresses the unique challenges of dynamic load behavior and distributed task execution observed in modern test and training infrastructures.
(2)
We design an innovative spatiotemporal attention fusion module that adaptively learns the significance of both spatial and temporal components in the data. This fusion mechanism enhances the model’s ability to generalize across diverse scenarios and operational conditions typically observed in Logical Range settings.
(3)
We conduct comprehensive experiments on real-world Logical Range datasets. The results demonstrate that our model outperforms state-of-the-art baselines in terms of accuracy, generalization, and robustness. Specifically, GCSG achieves an R 2 of 0.86, a mean absolute error (MAE) of 4.5, and a mean squared error (MSE) of 34, underscoring its practical utility in supporting more efficient and responsive resource scheduling across complex range systems.

2. Materials and Methods

2.1. Architecture for Logical Range

As introduced in Section 1, the Logical Range comprises a variety of networked task-processing devices, including theodolites, cameras, radar systems, and Time–Space–Position Information (TSPI) units. These devices are interconnected through a high-speed network to support real-time data processing, task coordination, and multimedia information flow (e.g., image/video display).
Figure 1 illustrates the overall structure of the Logical Range, where heterogeneous devices collaborate to execute computationally intensive tasks. These devices are responsible for data computation, command distribution, real-time acquisition, and visualization, resulting in dynamic and unpredictable workloads.
The resource usage of these devices is highly complex and variable, typically involving metrics such as CPU usage, memory consumption, network I/O (uplink/downlink), and disk activity. Among them, CPU usage is particularly critical, as most operations in Logical Range scenarios are computationally intensive and latency-sensitive. Therefore, this study focuses on CPU usage as the primary time-series signal for load forecasting and task scheduling optimization.
As illustrated in Figure 2, the CPU usage data from all task-processing devices are recorded over time to form a multivariate time series, denoted as
x = { x i : i = 1 , 2 , , T } ,
where x i represents the aggregated CPU usage at time step i. The sequence x exhibits strong temporal dependencies, and accurate forecasting of its future values is essential for reducing latency, avoiding resource bottlenecks, and improving experimental throughput in Logical Range systems.
To construct an effective forecasting dataset, we normalize x and apply a sliding time window strategy. At each time step i, a subsequence of length w,
x i w = [ x i w , x i w + 1 , , x i ]
is extracted as the input sample, and the forecasting target is set as the value x i + d at a future time step d ahead. The forecasting function is defined as
x i + d = f ( x i w : d ) ,
where f is the deep learning model proposed in this study and d is the forecasting step, with d 1 . The complete dataset can be constructed as
S = { , ( x i w , x i + d ) , ( x i + 1 w , x i + d + 1 ) , ( x i + 2 w , x i + d + 2 ) , }
and we denote each sample as s i w = ( x i w , x i + d ) . This sliding window mechanism enables continuous load forecasting and dynamic resource allocation. The processing of S is depicted in Figure 2.

2.2. Proposed Load Forecasting Methodology

The proposed forecasting model integrates the GAF, CNN, SENet, and GRU methods and is thus abbreviated as GCSG (GAF–CNN-SENet–GRU). The overall architecture of the GCSG load forecasting model for the Logical Range is shown in Figure 3.
The processing pipeline of the GCSG model comprises three stages: (1) spatial feature extraction via GAF and CNN-SENet; (2) temporal feature extraction using GRU; (3) attention-based feature fusion.

2.2.1. Spatial Feature Extraction

Spatial feature extraction involves transforming the dataset from time-series data into image data via GAF encoding.
Subsequently, a CNN extracts features from the resulting images, with feature optimization through channel attention mechanisms such as SENet [20].
  • Image conversion.
GAF encoding is a visualization and feature extraction technique for time-series analysis and signal processing. It converts one-dimensional time-series data into two-dimensional images while preserving their temporal characteristics. In this study, the GAF approach is used to transform the input data x i w within each dataset s i w of Logical Range load data into two-dimensional images, expanding the data dimensionality. Figure 4 illustrates the conversion process.
The main steps of the GAF conversion process for Logical Range load data are as follows:
(1)
Extract the input data x i w from the dataset s i w , where w is the length of x i w and x i w = [ x i w , x i w + 1 , , x i ] ;
(2)
Convert x i w to polar coordinates using Equation (5):
θ i = arccos ( x i ) r i = t i T ,
where r i is the polar radius, t i is the timestamp of x i , T is the total timestamp length, and θ i is the polar angle. Here, r i preserves the temporal characteristics of the time-series data, while θ i retains the numerical variations.
(3)
Construct the GAF matrix g i using the cosine function (Equation (6)), with dimensions w × w :
g i = cos ( θ 1 + θ 1 ) cos ( θ 1 + θ w ) cos ( θ w + θ 1 ) cos ( θ w + θ w ) .
Compared to x i w , g i enhances the correlations between load data values across different timestamps in the Logical Range.
2.
Feature extraction.
The CNN is a deep learning architecture that is primarily used for tasks involving structured data, such as image recognition, classification, and object detection. A CNN operates by applying convolutional operations to extract features from images, followed by pooling operations to reduce the dimensionality of the extracted features. CNNs can automatically learn and capture local features within images and exhibit certain invariance to translations, scaling, and other transformations, making them highly effective for image data processing [21].
The channel attention mechanism enhances the interaction of information between channels in CNN feature maps, helping to learn the importance of different channels [22]. SENet, as one type of channel attention mechanism, adaptively adjusts the weights of individual channels to improve the network’s focus on essential features, thereby enhancing the CNN’s feature representation capability [23].
The GCSG model employs a CNN to extract features from g i and integrates SENet to adjust feature weights, thus improving the CNN’s perception of important features. The specific process is as follows:
(1)
Conduct successive convolution, activation, and pooling operations on g i to obtain feature maps p i = [ p i , 1 , p i , 2 , , p i , c ] , where the size of p i is h × l × c and c is the number of channels, with each channel containing a two-dimensional feature.
(2)
Compute the global statistics z i , j for each channel feature p i , j in p i , where j = 1 , 2 , , c , using Equation (7):
z i , j = 1 h × l a = 1 h b = 1 l p i , j ( a , b ) .
The global pooling operation aggregates the values of p i , j across its spatial dimensions to derive z i , j .
(3)
Compute the weight β i , j for each channel feature p i , j in p i to obtain the weight vector β i , where z i = [ z i , 1 , z i , 2 , , z i , c ] , using Equation (8):
β i = σ ( ω 2 δ ( ω 1 z i ) ) ,
where σ denotes the sigmoid function, δ represents the ReLU function, and ω 1 and ω 2 are full layer matrices.
(4)
Compute the weighted feature map m i using Equation (9):
m i = p i · β i .
(5)
Apply convolutional and transformation operations on m i to obtain the spatial feature output vector v i for x i w .

2.2.2. Temporal Feature Extraction

The GRU is a recurrent neural network architecture that incorporates gating mechanisms to address the problems of vanishing and exploding gradients faced when using traditional RNNs [24].
When processing time-series load data in the Logical Range, the GRU preserves temporal features through its hidden states. The GCSG model employs a two-layer stacked GRU to extract temporal features from x w i :
(1)
Input vector x i w from dataset s i w (where x i w = [ x i w , x i w + 1 , , x i ] ) with initialized zero hidden state into the first GRU layer.
(2)
Compute reset gate, update gate, candidate hidden state, and hidden state updates using x i w to obtain the hidden state vector u i .
(3)
Process u i and x i w through the second GRU layer to generate the hidden state vector o i .
(4)
Transform o i via a fully connected layer to derive the temporal features output vector k i .

2.3. Feature Fusion

An attention-based fusion method enhances the model’s focus on critical spatial and temporal features. The method weights the spatial feature v i and temporal feature k i of x i w according to their contributions to load forecasting accuracy in the Logical Range. The weight computation follows Equation (10):
q i = softmax ( v i k i T χ k i ) k i ,
where χ k i is the dimension weight for k i , and q i is the attention-fused feature vector.
A fully connected layer transforms q i ’s dimensions as shown in Equation (11):
y i = W q i ,
where W is the weight matrix and y i is the final forecasting result.

3. Results

3.1. Experimental Setup

To evaluate the effectiveness of the GCSG model, the experimental data used in this study were collected from the multi-region joint simulation test system of the Digital Twin Logical Range, jointly developed by the Chinese Flight Test Establishment and Xi’an University of Technology. This system primarily processes aircraft flight status data obtained from devices such as GPS, theodolites, cameras, and radars during flight tests. It integrates six high-performance servers to form a central processing node cluster that is responsible for data processing, digital twin model rendering, and command forwarding.
Scripts were developed to collect and store various load metrics from each node in the central processing cluster. The data collection spanned 144 h, with a sampling interval of 90 s, resulting in 5760 records per node (Logical Range task processing device). The dataset contains timestamps, Logical Range IDs, device IDs, and CPU load measurements (0–100%).
To ensure consistency in model training and convergence, all input features were normalized using standardization (i.e., zero mean and unit variance) based on the statistics of the training set. This preprocessing step improves the numerical stability of the model and ensures fair comparison across different methods.
To ensure the accuracy and reproducibility of the experimental results, all experiments were conducted under consistent software and hardware conditions. The detailed hardware and software configurations are summarized in Table 1.

3.2. Performance Evaluation Metrics

To ensure a comprehensive and reliable evaluation of the GCSG model’s performance, we adopt a multi-metric strategy comprising the coefficient of determination ( R 2 , see Equation (12)), mean absolute error (MAE, see Equation (13)), and mean squared error (MSE, see Equation (14)). Specifically, R 2 measures the proportion of variance explained by the model, while the MAE and MSE quantify the prediction errors from absolute and squared perspectives, respectively. This combination of metrics enables a more holistic assessment of the model’s accuracy and robustness, avoiding overreliance on any single evaluation indicator.
  • The coefficient of determination ( R 2 ) measures the proportion of variance in the predicted Logical Range load data explained by the model. Higher values indicate better performance:
    R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ i ) 2 .
  • The mean absolute error (MAE) is computed as the average absolute difference between predicted and actual load forecasting values:
    MAE = 1 n i = 1 n y i y ^ i .
  • The mean squared error (MSE) quantifies average squared prediction errors, penalizing larger deviations:
    MSE = 1 n i = 1 n ( y i y ^ i ) 2 .
In the above formulas, y ^ i denotes the predicted value, y i denotes the true (actual) value, and y ¯ is the mean of the true values.
These metrics were used to assess the model’s accuracy for time-series load forecasting in the Logical Range.

3.3. Performance Analysis

To evaluate the performance of the proposed GCSG model in the Logical Range scenario, we conducted comparative experiments with several widely used deep learning models, including CNN, LSTM, GRU, LSTM-ED [25], BiLSTM [26], LSTM-GRU [27], and ConvLSTM [28] models. For a fair comparison, each baseline model was individually fine-tuned to achieve its best performance through grid search. All models were trained and evaluated using the same dataset splits and experimental conditions.
The specific hyperparameter settings for each model, including hidden size, learning rate, number of training epochs, batch size, dropout rate, number of layers, and optimizer, are summarized in Table 2.
(1)
Input window size analysis.
Figure 5 displays the performance metrics of different models under varying input window sizes w using the prediction step d = 1 in the Logical Range. As shown in Figure 5a, the GCSG model achieved the highest R 2 values, consistently above 0.85 for w values between 16 and 40. The BiLSTM, ConvLSTM, LSTM-ED, and LSTM-GRU models (all LSTM variants) followed, with R 2 scores of 0.83–0.85. The GRU and LSTM models demonstrated comparable performance ( R 2 0.80 ) due to their single-model architectures. The CNN showed the lowest R 2 values (0.75–0.77), potentially due to its parameters and structural configuration.
In particular, Figure 5a demonstrates the fluctuation degree of R 2 scores across different w values for each model in the Logical Range, with the GCSG model exhibiting the most stable performance. This indicates the GCSG model’s superior robustness when adapting to varying w values.
Figure 5b,c present the MAE and MSE metrics of each model, respectively. The error trends consistently align with the R 2 results in Figure 5a. The GCSG model achieves the lowest errors, followed by the BiLSTM, ConvLSTM, LSTM-ED, and LSTM-GRU models, with marginally higher error rates. The GRU and LSTM models showed intermediate error levels, while the CNN yielded the highest number of errors.
In summary, Figure 5 confirms that the GCSG model delivers optimal load forecasting accuracy across different input window sizes w in the Logical Range, outperforming both combined models (BiLSTM, ConvLSTM, LSTM-ED, LSTM-GRU) and traditional single models (LSTM, GRU, CNN).
(2)
Prediction step analysis.
According to the results presented in Figure 5, an input window size w of 36 was chosen based on the average R 2 , MAE, and MSE values, and the prediction step d was varied from 1 to 29 in order to compare the prediction accuracy of the various models, as shown in Figure 6.
Figure 6 demonstrates that the GCSG model consistently achieved superior load forecasting accuracy across different prediction steps d in the Logical Range compared to the other models. As shown in Figure 6a, the combined models (ConvLSTM, LSTM-ED, and LSTM-GRU) exhibited higher R 2 values than single models (LSTM, GRU, and CNN). Among all combined models, the GCSG model maintained the highest R 2 values.
According to Figure 6b, the CNN exhibited the highest MAE values, while the LSTM and GRU models demonstrated lower MAE than the CNN but higher values than ConvLSTM, LSTM-ED, LSTM-GRU, and GCSG. When d 11 , BiLSTM’s MAE was lower than those of the LSTM and GRU. However, when d > 11 , its MAE became comparable to that of the LSTM. Additionally, the combined models (ConvLSTM, LSTM-ED, and LSTM-GRU) maintained similar MAE values across different d values, with LSTM-ED exhibiting marginally lower MAE than LSTM-GRU and ConvLSTM. The GCSG model consistently achieved the lowest MAE values across all prediction steps d.
Figure 6c reveals that the MSE of the GCSG model follows a nearly linear growth trend with increasing prediction step d, reflecting the expected accumulation of error. In contrast, the LSTM, GRU, LSTM-ED, ConvLSTM, BiLSTM, and LSTM-GRU models exhibited fluctuating MSE growth patterns. This demonstrates the GCSG model’s stable linear MSE d relationship in load forecasting applications.

3.3.1. Ablation Experiment

To rigorously evaluate the contribution of each component in the GCSG model for Logical Range load forecasting, we conducted systematic ablation studies using the following model variants:
  • GCSG: Full model (GAF+CNN+SENet+GRU).
  • GCG: Ablated SENet (GAF+CNN+GRU).
  • CSG: Ablated GAF (CNN+SENet+GRU).
  • CG: Baseline (CNN+GRU).
The experimental parameters maintained strict consistency across all model variants, with the detailed configuration provided in Table 3.
Table 4 and Table 5 present the load forecasting performance metrics ( R 2 , MAE, and MSE) for each model in the Logical Range, evaluated across varying input window sizes w and prediction step lengths d.
Table 4 demonstrates the experimental results for the four models with fixed d = 1 and different values of w. The four models (GCSG, GCG, CSG, and CG) exhibited stable performance across varying w values, with no significant fluctuations observed. Notably, when w = 36 , the R 2 , MSE, and MAE values most closely approximated the average metrics across all tested values of w.
Table 5 demonstrates that, as the prediction step d increases, the capability of the models for temporal dependency feature extraction in the Logical Range presents a decreasing trend. While all models exhibited declining R 2 values, the GCSG model consistently maintained superior load forecasting accuracy compared to the other models. The model performance ranking was GCSG > GCG > CSG > CG, with MAE and MSE metrics confirming this trend. These results validate that the combined GAF–CNN-SENet–GRU architecture in GCSG achieves optimal prediction performance.
The MAE and MSE values indicate that GCG outperforms CSG while approaching GCSG’s accuracy, underscoring the GAF’s greater impact on load forecasting accuracy compared to the SENet channel attention mechanism. GAF encoding effectively preserves temporal correlations while enhancing the relevance of inter-temporal values, enabling more comprehensive feature extraction. The GCSG model synergistically combines the GAF’s temporal preservation effect with SENet’s channel importance weighting.
These experimental results conclusively demonstrate the GCSG model’s superior accuracy across all prediction steps, proving that the simultaneous integration of SENet and GAF significantly enhances the resulting model’s load forecasting performance.

3.3.2. Generalization Performance

To evaluate the generalization capability of the GCSG model across different Logical Range training datasets, we conducted experiments using ten distinct data partitioning schemes with varying random seeds. Each scheme maintained an input window size of w = 36 and prediction steps d ranging from 1 to 29 with an interval of 2. The experimental results are presented in Figure 7.
Analysis of the datasets partitioned with different random seeds in the Logical Range revealed that the GCSG model can maintain consistent load forecasting performance. Figure 7a demonstrates stable R 2 values around 0.85, while Figure 7b shows MAE values consistently near 4.4. Similarly, Figure 7c confirms the stabilization of the MSE at around 35. Compared to other models, GCSG exhibited superior stability across all three metrics without significant fluctuations, demonstrating its robust generalization capability across different data partitions.

4. Discussion

The results demonstrate that the GCSG model consistently outperformed other models in predicting time-series load data, achieving higher R-squared values and lower MAE and MSE across different window sizes and prediction steps. This suggests that the integration of GAF, CNN, SENet, and GRU in the GCSG model contributes significantly to its predictive accuracy. These findings align with prior studies emphasizing the importance of incorporating attention mechanisms and sequential learning in time-series forecasting tasks.
A comparison with the existing literature reveals both similarities and differences. While our study corroborated previous findings regarding the effectiveness of LSTM-based models in time-series prediction, the novel combination of GAF and SENet in the GCSG model provides additional improvements in capturing temporal dependencies and feature extraction.
We acknowledge several limitations in our study. The choice of hyperparameters and model configurations could have influenced the results, and the generalizability of the GCSG model to other domains requires further investigation.
In future research, exploring the application of the GCSG model in different domains and incorporating domain-specific features could enhance its robustness and applicability. Further investigations focused on the optimization of hyperparameters and evaluating model interpretability will be valuable for advancing the understanding of deep learning techniques in time-series analysis.

5. Conclusions

This study presented GCSG, an innovative hybrid deep learning model designed for precise load forecasting in Logical Range environments. The integration of Gramian Angular Field transformation with CNN-SENet spatial feature extraction and GRU-based temporal modeling was shown to be particularly effective in handling the complex, uneven workload distributions that are characteristic of Logical Range testing scenarios. Experimental validation demonstrated that GCSG achieves superior performance, with an R 2 of 0.86, MAE of 4.5, and MSE of 34.5, representing significant improvements of 2.2%, 25.7%, and 40.4%, respectively, over conventional approaches (including BiLSTM, LSTM-ED, and other baselines).
The model exhibited remarkable stability across varying operational conditions, maintaining consistent accuracy with input window sizes ranging from 16 to 40 and prediction steps spanning 1 to 29. This robustness was further verified through testing on multiple data partitioning schemes using different random seeds, confirming GCSG’s reliable generalization capability.
Looking ahead, two promising research directions emerge: extension to multi-device collaborative forecasting scenarios and optimization for edge computing deployment in order to enhance computational efficiency. These developments would further strengthen the model’s applicability in large-scale Logical Range implementations while addressing potential latency constraints in time-sensitive applications.

Author Contributions

Conceptualization, H.C.; methodology, H.C.; writing—original draft preparation, H.C.; formal analysis, Z.D.; data curation, Z.D.; supervision, H.C. and Z.D. H.C. and Z.D. reviewed the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All materials and data used are available from Xi’an University of Technology, China. The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, H.; Dang, Z.; Hei, X.; Wang, K. Design and application of logical range framework based on digital twin. Appl. Sci. 2023, 13, 6589. [Google Scholar] [CrossRef]
  2. Huan, M. Development of Logical Range Planning Furthermore, Verification Tool for Joint Test Platform. Master’s Thesis, Harbin Institute of Technology, Harbin, China, 2018. [Google Scholar]
  3. Li, D.; Li, L.; Jin, L.; Huang, G.; Wu, Q. Research of load forecasting and elastic resources scheduling of openstack platform based on time series. J. Chongqing Univ. Posts Telecommun. Nat. Sci. Ed. 2016, 28, 560–566. [Google Scholar]
  4. Khan, A.; Yan, X.; Tao, S.; Anerousis, N. Workload characterization and prediction in the cloud: A multiple time series approach. In Proceedings of the 2012 IEEE Network Operations and Management Symposium, Maui, HI, USA, 16–20 April 2012; pp. 1287–1294. [Google Scholar]
  5. He, J.; Hong, S.; Zhang, C.; Liu, Y.; Deng, F.; Yu, J. A method to cloud computing resources requirement prediction on saas application. In Proceedings of the 2021 International Conference on Machine Learning and Intelligent Systems Engineering (MLISE), Chongqing, China, 9–11 July 2021; pp. 107–116. [Google Scholar]
  6. Shyam, G.K.; Manvi, S.S. Virtual resource prediction in cloud environment: A bayesian approach. J. Netw. Comput. Appl. 2016, 65, 144–154. [Google Scholar] [CrossRef]
  7. Sun, W.; Zhang, H.; Palazoglu, A.; Singh, A.; Zhang, W.; Liu, S. Prediction of 24-hour-average PM2.5 concentrations using a hidden markov model with different emission distributions in northern california. Sci. Total Environ. 2013, 443, 93–103. [Google Scholar] [CrossRef] [PubMed]
  8. Dinda, P.A.; O’Hallaron, D.R. Host load prediction using linear models. Clust. Comput. 2000, 3, 265–280. [Google Scholar] [CrossRef]
  9. Liu, X.; Xie, X.; Guo, Q. Research on cloud computing load forecasting based on lstm-arima combined model. In Proceedings of the 2022 Tenth International Conference on Advanced Cloud and Big Data (CBD), Guilin, China, 4–5 November 2022; pp. 19–23. [Google Scholar]
  10. Mehdi, H.; Pooranian, Z.; Vinueza Naranjo, P.G. Cloud traffic prediction based on fuzzy arima model with low dependence on historical data. Trans. Emerg. Telecommun. Technol. 2022, 33, e3731. [Google Scholar] [CrossRef]
  11. Xie, Y.; Jin, M.; Zou, Z.; Xu, G.; Feng, D.; Liu, W.; Long, D. Real-time prediction of docker container resource load based on a hybrid model of arima and triple exponential smoothing. IEEE Trans. Cloud Comput. 2020, 10, 1386–1401. [Google Scholar] [CrossRef]
  12. Sharma, A.K.; Punj, P.; Kumar, N.; Das, A.K.; Kumar, A. Lifetime prediction of a hydraulic pump using arima model. Arab. J. Sci. Eng. 2024, 49, 1713–1725. [Google Scholar] [CrossRef]
  13. Prevost, J.J.; Nagothu, K.; Kelley, B.; Jamshidi, M. Prediction of cloud data center networks loads using stochastic and neural models. In Proceedings of the 2011 6th International Conference on System of Systems Engineering, Albuquerque, NM, USA, 27–30 June 2011; pp. 276–281. [Google Scholar]
  14. Karim, M.E.; Maswood, M.M.S.; Das, S.; Alharbi, A.G. Bhyprec: A novel bi-lstm based hybrid recurrent neural network model to predict the cpu workload of cloud virtual machine. IEEE Access 2021, 9, 131476–131495. [Google Scholar] [CrossRef]
  15. Shu, W.; Zeng, F.; Ling, Z.; Liu, J.; Lu, T.; Chen, G. Resource demand prediction of cloud workloads using an attention-based gru model. In Proceedings of the 2021 17th International Conference on Mobility, Sensing and Networking (MSN), Exeter, UK, 13–15 December 2021; pp. 428–437. [Google Scholar]
  16. Bi, J.; Ma, H.; Yuan, H.; Zhang, J. Accurate prediction of workloads and resources with multi-head attention and hybrid lstm for cloud data centers. IEEE Trans. Sustain. Comput. 2023, 8, 375–384. [Google Scholar] [CrossRef]
  17. Patel, E.; Kushwaha, D.S. A hybrid cnn-lstm model for predicting server load in cloud computing. J. Supercomput. 2022, 78, 1–30. [Google Scholar] [CrossRef]
  18. Kumar, J.; Goomer, R.; Singh, A.K. Long short term memory recurrent neural network (lstm-rnn) based workload forecasting model for cloud datacenters. Procedia Comput. Sci. 2018, 125, 676–682. [Google Scholar] [CrossRef]
  19. Rajagukguk, R.A.; Ramadhan, R.A.; Lee, H.-J. A review on deep learning models for forecasting time series data of solar irradiance and photovoltaic power. Energies 2020, 13, 6623. [Google Scholar] [CrossRef]
  20. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  21. Ilesanmi, A.E.; Ilesanmi, T.O. Methods for image denoising using convolutional neural network: A review. Complex Intell. Syst. 2021, 7, 2179–2198. [Google Scholar] [CrossRef]
  22. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  23. Huang, J.; Ren, L.; Zhou, X.; Yan, K. An improved neural network based on senet for sleep stage classification. IEEE J. Biomed. Health Inform. 2022, 26, 4948–4956. [Google Scholar] [CrossRef] [PubMed]
  24. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  25. Nguyen, H.M.; Kalra, G.; Kim, D. Host load prediction in cloud computing using long short-term memory encoder–decoder. J. Supercomput. 2019, 75, 7592–7605. [Google Scholar] [CrossRef]
  26. Zhang, M.; Wu, D.; Xue, R. Hourly prediction of PM2.5 concentration in beijing based on bi-lstm neural network. Multimed. Tools Appl. 2021, 80, 24455–24468. [Google Scholar] [CrossRef]
  27. Cho, M.; Kim, C.; Jung, K.; Jung, H. Water level prediction model applying a long short-term memory (lstm)–gated recurrent unit (gru) method for flood prediction. Water 2022, 14, 2221. [Google Scholar] [CrossRef]
  28. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.-Y. Convolutional lstm network: A machine learning approach for precipitation nowcasting. Adv. Neural Inf. Process. Syst. 2015, 28, 1–9. [Google Scholar]
Figure 1. Architecture of Logical Range Environment.
Figure 1. Architecture of Logical Range Environment.
Applsci 15 05628 g001
Figure 2. Sliding window-based processing of device load data in Logical Range.
Figure 2. Sliding window-based processing of device load data in Logical Range.
Applsci 15 05628 g002
Figure 3. Processing pipeline of the GCSG model.
Figure 3. Processing pipeline of the GCSG model.
Applsci 15 05628 g003
Figure 4. Gramian Angular Field (GAF) -based conversion of Logical Range load data.
Figure 4. Gramian Angular Field (GAF) -based conversion of Logical Range load data.
Applsci 15 05628 g004
Figure 5. Comparison of models under different input window sizes.
Figure 5. Comparison of models under different input window sizes.
Applsci 15 05628 g005
Figure 6. Comparison of models under different prediction steps.
Figure 6. Comparison of models under different prediction steps.
Applsci 15 05628 g006
Figure 7. Comparison of models under different datasets.
Figure 7. Comparison of models under different datasets.
Applsci 15 05628 g007
Table 1. Experimental setup.
Table 1. Experimental setup.
CategorySpecification
Operating systemWindows 10 (64-bit)
Programming languagePython 3.9
Deep learning frameworkPyTorch 2.0.0+ cu118
Processor12th Gen Intel(R) Core(TM) i9-12900K 3.20 GHz
GPUNVIDIA GeForce RTX 3090
Memory64 (GB)
Table 2. Model parameter configurations.
Table 2. Model parameter configurations.
ModelHidden SizeLREpochsBatch SizeDropoutLayersOptimizer
CNN1000.001100500.31Adam
LSTM100.150500.31Adam
GRU640.00150500.31Adam
LSTM-ED640.000150500.21Adam
BiLSTM100.0150500.011Adam
LSTM-GRU640.000150500.41Adam
ConvLSTM640.00150500.011Adam
GCSG320.00011001000.012Adam
Table 3. Experimental configuration for ablation study.
Table 3. Experimental configuration for ablation study.
Parameter CategoryValue/Range
Input window size (w) 16 w 40 (optimal w = 36 )
Prediction steps (d) d { 1 , 3 , , 29 }
Batch size100
Training epochs2000
Learning rate0.0001
OptimizerAdam
Table 4. Comparison of R 2 , MAE, and MSE for different input window sizes (w).
Table 4. Comparison of R 2 , MAE, and MSE for different input window sizes (w).
w R 2 MAEMSE
CGGCGCSGGCSGCGGCGCSGGCSGCGGCGCSGGCSG
160.8380.8480.8400.8604.5384.5134.6324.44835.22734.15635.59133.382
200.8380.8510.8420.8624.5704.4384.5364.39935.01433.11134.70532.064
240.8340.8460.8420.8594.5904.4414.5954.46135.47033.69335.33233.152
280.8230.8420.8390.8564.8354.5354.7284.48938.58835.15336.79734.250
320.8350.8390.8400.8584.5804.5594.6314.41235.42835.45636.31433.615
360.8310.8430.8390.8564.6254.5034.6714.46836.23134.44835.81533.804
400.8270.8330.8370.8534.6574.6554.6894.49937.33737.13637.42334.500
Table 5. Comparison of R 2 , MAE, and MSE for different prediction steps (d).
Table 5. Comparison of R 2 , MAE, and MSE for different prediction steps (d).
d R 2 MAEMSE
CGGCGCSGGCSGCGGCGCSGGCSGCGGCGCSGGCSG
10.8310.8430.8380.8554.9004.5564.7224.45038.28235.63137.52933.827
30.7840.8010.7920.8035.5205.3465.4355.30351.47546.54649.79345.259
50.7680.7820.7780.7915.6905.5185.6555.43853.30550.53352.34549.243
70.7550.7700.7630.7735.8225.5985.7445.54855.90152.71754.33451.084
90.7460.7630.7530.7705.9665.7055.9105.62758.97553.92857.09652.407
110.7430.7580.7520.7646.1385.8196.0345.78360.00056.05058.23855.400
130.7380.7520.7460.7596.2586.0456.1715.89463.84359.92961.34658.356
150.7310.7510.7410.7546.3006.1096.2225.99164.94161.57563.05759.696
170.7210.7460.7330.7516.3896.1846.2816.12865.95062.05464.92760.328
190.7090.7400.7170.7466.5376.2996.4506.22768.92762.94867.55960.916
210.7040.7300.7130.7406.6396.3216.5626.26570.98163.91868.21362.744
230.6920.7260.7010.7306.6866.3556.5996.28372.93365.96671.21364.096
250.6890.7220.6960.7286.7116.4066.6376.30773.82966.74872.58365.703
270.6870.7140.6950.7236.8326.4906.7496.40774.89467.99373.31266.731
290.6680.6940.6720.7037.1626.6987.0176.63479.97574.37978.63773.266
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, H.; Dang, Z. A Hybrid Deep Learning-Based Load Forecasting Model for Logical Range. Appl. Sci. 2025, 15, 5628. https://doi.org/10.3390/app15105628

AMA Style

Chen H, Dang Z. A Hybrid Deep Learning-Based Load Forecasting Model for Logical Range. Applied Sciences. 2025; 15(10):5628. https://doi.org/10.3390/app15105628

Chicago/Turabian Style

Chen, Hao, and Zheng Dang. 2025. "A Hybrid Deep Learning-Based Load Forecasting Model for Logical Range" Applied Sciences 15, no. 10: 5628. https://doi.org/10.3390/app15105628

APA Style

Chen, H., & Dang, Z. (2025). A Hybrid Deep Learning-Based Load Forecasting Model for Logical Range. Applied Sciences, 15(10), 5628. https://doi.org/10.3390/app15105628

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop