Next Article in Journal
Large Language Models for Machine Learning Design Assistance: Prompt-Driven Algorithm Selection and Optimization in Diverse Supervised Learning Tasks
Previous Article in Journal
Shock Absorption Control of Sand-Layer Isolation Liquid Storage Structure with Soft Steel
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PRPOS: A Periodicity-Aware Resource Prediction Framework for Online Services

by
Yi Liang
*,
Hongwen Zhou
,
Tianxu Li
and
Haotian Shen
College of Computer Science, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(20), 10967; https://doi.org/10.3390/app152010967
Submission received: 18 September 2025 / Revised: 8 October 2025 / Accepted: 10 October 2025 / Published: 13 October 2025
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

Accurate prediction of resource utilization is essential for efficient cloud resource management and Quality-of-Service (QoS) assurance in online services. However, most existing methods neglect to explicitly model inherent periodic patterns in resource usage—particularly those characterized by extended period lengths, consistent trend shapes with significant magnitude variations across periods—which limits their predictive accuracy. To address this gap, we propose PRPOS (Periodicity-aware Resource Prediction for Online Services), a novel periodicity-aware prediction framework specifically designed for online service workloads. PRPOS operates in two cohesive phases: It first employs a robust period detection mechanism that effectively handles magnitude variations and noise to identify dominant periods; then, a dual scale based on a Gated Recurrent Units (GRU) predictor explicitly incorporates the identified periodicity to concurrently model fine-grained in-period dynamics and coarse-grained cross-period trends. Extensive evaluation on the Alibaba Cluster Trace v2018 demonstrates that PRPOS consistently outperforms state-of-the-art approaches, achieving average improvements of 45.3% in Mean Absolute Percentage Error (MAPE) and 44.3% in Root Mean Squared Error (RMSE). The demonstrated performance enables the application of PRPOS to cloud resource orchestration for online services, allowing for proactive resource provisioning that enhances both efficiency and reliability.

1. Introduction

With the rapid development of cloud computing technology, an increasing number of online services, such as web services and stream processing, are being deployed on cloud platforms to harness the unlimited availability of virtualized resources and a pay-as-you-go cost model [1]. Efficient resource provisioning is crucial to guarantee the QoS of online services as well the cloud resource utilizations [2]. However, as the incoming service requests are fluctuating, the resource demands of online services vary constantly over time. Provisioning resources in a dynamic and cost-effective manner presents a significant challenge for cloud operators.
Accurate resource prediction is the premise of dynamic resource provisioning to online services [3]. In cloud scenarios, the resource prediction of online services mainly refers to using their historical resource usage to estimate their future resource utilization. With an accurate resource prediction, cloud operators can make reasonable resource provisioning decisions so as to improve service quality and reduce the waste of cloud resources.
As online services are typically long-running, their resource utilization patterns naturally lend themselves to being conceptualized as time series. Periodicity encapsulates the recurrent pattern of variations observable in such time series. As the scale of services expands and the regularity of users’ service access patterns intensifies, the periodic characteristic gains even greater significance in the resource usage of online services. Our statistical analysis, anchored in the Alibaba cluster trace, reveals that over 90% of online services exhibit distinctive periodic characteristics in their resource usage [4].
Previous studies in resource predictions have not yet to focus on these periodic features of online services. Even when attempts were made to harness the temporal dynamics of online service resource utilization, they frequently assumed that future resource demands are solely correlated with the most recent historical data points. These studies have primarily relied on regression techniques, such as Autoregressive Integrated Moving Average (ARIMA), or machine learning approaches, like Convolutional Neural Networks (CNNs) and attention mechanisms, to implicitly explore correlations between resource usage at different time points [4,5,6,7,8,9,10,11]. However, our statistical analysis has unveiled that online service resource utilization exhibits long-term periodic patterns. For instance, in the Alibaba cluster trace, the period length of resource usage for approximately 50% of online services exceeds 800 consecutive data points. However, current approaches to resource prediction typically utilize no more than 30 of the most recent data points [12,13,14]. Merely relying on the most proximal historical data points for resource predictions neglects the opportunity to capture cross-period correlations in resource utilization, thereby compromising the accuracy of resource prediction models.
Where promising, periodicity-based resource prediction for online services confronts two fundamental challenges. The first arises in period detection, where dynamic operating environments produce resource usage patterns that simultaneously display trend consistency and magnitude heterogeneity across cycles. Although morphological similarity across cycles confirms consistent periodic behavior, significant variations in both absolute magnitudes (e.g., 15% differences in peak Central Processing Unit (CPU) utilization in Alibaba cluster trace) and oscillatory ranges (e.g., 10-percentage-point amplitude fluctuations in Alibaba cluster trace) undermine the effectiveness of conventional detection methods. Frequency-domain approaches like Fourier transform fail to capture long-term periods due to spectral leakage [15,16,17], while time-domain methods, like auto-correlation analysis, are oversensitive to amplitude variations and suffer from peak submergence [18,19]. Although advanced time–frequency joint methods improve accuracy through multi-level wavelet decomposition, their high computational complexity fails to meet the lightweight real-time detection requirements of online services [20,21].
The second challenge lies in the inherent shortcomings of conventional predictive modeling. Most existing approaches, including advanced deep learning methods, employ a single-scale architecture to process resource usage data—a complex mix of long-term periodic patterns and short-term noise [22,23,24]. This unified representation fails to disentangle multi-scale temporal dependencies, causing models (even those with attention mechanisms) to struggle in distinguishing structurally relevant historical phases from incidental fluctuations. Consequently, without explicit mechanisms to capture periodicity, they often assign misplaced attention to recent outliers or noise while underutilizing stable, recurring periodic patterns, ultimately leading to sub-optimal prediction accuracy and robustness in online service resource prediction.
Aiming at addressing the above issues, we propose PRPOS, a novel resource prediction method for online services based on periodicity. PRPOS follows a two-phase framework. In the first phase, termed period detection, we target long-period patterns characterized by consistent trends yet significant magnitude variations. A lightweight detection mechanism is designed, combining normalized and weighted Hodrick–Prescott (HP) [25] filtering with a rank-based Auto-Correlation Function (ACF) [26] to smooth trend variations while preserving temporal order, thereby effectively suppressing magnitude fluctuations and generating a candidate period. Subsequently, a period refinement strategy based on cosine similarity is introduced to validate and adjust the candidate by measuring morphological consistency across different periodic segments, ultimately producing a high-precision period value.
In the second phase, referred to as resource prediction, we propose a periodicity-aware prediction model based on Gated Recurrent Units (GRUs), explicitly designed to leverage the periodic patterns identified earlier. The model comprises three specialized modules—in-period-GRU, cross-period-GRU, and aggre-FCN—which jointly capture both short-term dynamics and long-term periodic characteristics in online service resource usage. Using the pre-detected period length, the in-period-GRU encodes the most recent period to model short-term dynamics, while the cross-period-GRU integrates historical segments aligned with the same phase across multiple cycles to extract stable periodic patterns. The outputs from both GRU modules are then integrated by aggre-FCN, a fully connected network that performs adaptive feature fusion to produce the final prediction. This multi-scale modeling approach effectively exploits periodic structures in resource usage data, enhancing prediction accuracy and robustness in dynamic cloud environments.
The main contribution of this paper can be summarized as follows:
  • A proposal for incorporating periodicity characteristics into resource usage prediction for online services in cloud data centers.
  • A novel two-phased online service resource prediction framework, i.e., PRPOS. It accurately identifies dominant periods through a hybrid detection mechanism capable of handling significant magnitude variations across cycles. It also introduces a dual-scale GRU-based predictor that explicitly utilizes the discovered periodicity to simultaneously capture fine-grained in-period patterns and coarse-grained cross-period trends.
  • A comprehensive evaluation of PRPOS against other state-of-art baselines. Extensive experiments based on real-world and synthetic datasets confirm its significant improvements in prediction accuracy and robustness, solidifying its practical value for resource prediction in cloud environments.
The rest of this paper is organized as follows. We discuss the related works of this research in Section 2. We provide a detailed description of PRPOS in Section 3. Performance evaluation results are presented in Section 4. Discussion of our work is provided in Section 5. Finally, we conclude our work in Section 6.

2. Related Works

2.1. Resource Prediction

To maximize the utilization of shared resources, accurate resource prediction has become critical in data centers. It not only helps reduce resource waste effectively but also ensures QoS. Existing resource prediction approaches can be broadly categorized into three types: statistical methods, machine learning-based methods, and deep learning-based methods.
Statistical methods: Calheiros et al. [5] developed an ARIMA-based module for cloud workload prediction. Bi et al. [6] proposed a hybrid method that combines wavelet decomposition and ARIMA to predict the number of arriving tasks at the next time interval in data centers. Chen et al. [27] used Ensemble Empirical Mode Decomposition (EEMD) to decompose the time series, predicted each component separately with ARIMA, and combined the results to estimate host utilization. Sharifian et al. [28] applied a three-level wavelet decomposition to the original series, predicted trend and low-frequency components with SVR, and modeled high-frequency components with a Generalized AutoRegressive Conditional Heteroskedasticity(GARCH) model. However, these statistical methods rely heavily on prior knowledge of data characteristics such as distribution and trends, demanding considerable domain expertise. They also generally perform poorly on nonlinear or non-stationary data.
Machine learning-based methods: Such methods do not require extensive prior knowledge and are therefore widely used in handling nonlinear and non-stationary data problems [29,30,31]. Moreno-Vozmediano et al. [7] used a Support Vector Machine (SVM) regression model, whose parameters were selected by an analytical method, to predict the workload of the web server based on historical observations. Baig et al. [32] proposed a novel method to adaptively and automatically identify the most appropriate model to accurately predict data centers’ resource utilization. Gao et al. [33] introduced a clustering-based workload prediction method, which first clusters all the tasks into several categories and then trains a prediction model for each category respectively. Wen et al. [34] presented a Deep Belief Network (DBN) and Particle Swarm Optimization (PSO)-based CPU usage prediction algorithm named DP-CUPA and aimed to provide more accurate CPU usage prediction in data centers. Chen et al. [35] proposed a new Memory Wavelet Neural Network (MWNN) model to achieve both high accuracy and low time consumption in the prediction of real-time data streams. Xiao et al. [8] proposed Enhanced Extreme Gradient Boosting (Enhanced-XGB), an XGBoost-based prediction model, which integrates a convolutional neural network to train the weights of each tree. The final prediction is obtained through a weighted sum of the individual tree predictions plus a bias term. The common drawback of such methods lies in their failure to fully leverage the dynamic dependencies embedded in historical data along the temporal dimension.
Deep learning-based methods: Such methods can effectively capture complex patterns in historical data and achieve higher prediction accuracy. Zhang et al. [9] used Recurrent Neural Networks (RNN) for workload prediction, while Song et al. [10] applied Long Short-term Memory (LSTM) for multi-step-ahead workload predictions. H. M. Nguyen et al. [36] introduced an LSTM Encoder–Decoder model to enhance memory capability. Thonglek et al. [37] designed an LSTM-based model to improve resource allocation decisions for batch jobs. Shen et al. [38] proposed a host load prediction model using bidirectional LSTM (BiLSTM), and Bi et al. [12] combined bidirectional and grid LSTM networks for high-quality predictions of workload and resource series. Surbab et al. [13] developed a deep learning approach with adaptive window size selection to capture recent resource utilization trends. Ouhame et al. [39] employed a CNN-LSTM model to predict multivariate workload (CPU, memory, and network usage). Shu et al. [14] designed an attention-based GRU model to efficiently extract valuable information from large datasets. Chen et al. [40] proposed Resource Prediction based on Temporal Convolutional Networks (RPTCN), a Temporal Convolutional Network (TCN) to capture the long-term temporal dependencies for resource usage prediction. J. Dogani et al. [41] combined CNN and GRU to improve multi-step host workload forecasting. J. Bi et al. [11] integrated multi-head attention with hybrid LSTM for accurate workload and resource prediction. S. Garg et al. [22] developed a deep learning model incorporating BIRCH clustering with LSTM, Transformer, and Informer architectures to predict resource utilization in heterogeneous clouds. H. Yuan et al. [23] introduced VSBG, a prediction approach that combines Variational Mode Decomposition (VMD), Savitzky Golay (SG) filter, BiLSTM, and Grid LSTM, for precise workload and resource predicting in data centers. F. Ullah et al. [42] designed a BiLSTM-based framework for predicting nonlinear multivariate workload and resource utilization. Lastly, S. Tripathi et al. [24] proposed MWRP, a novel architecture integrating Transformer, CNN, and LSTM to improve environmental resource utilization prediction in cloud computing. S. Kashyap et al. [43] introduced a Variable-Size Sliding Window (VSSW) and a Model Selector Decision Support System (MSDSS), this is an approach for multi-model methods, incorporating ARIMA, Multiple Linear Regression (MLR), Gaussian Process Regression (GPR), Multi-Layer Perceptron (MLP), Support Vector Machine Regression (SVMR), and LSTM. However, this approach has two notable limitations: It relies on manual threshold tuning for model selection and lacks a mechanism for explicit periodicity modeling. J. bi et al. [44] proposed a forecasting method based on the Integration of a SG filter, a Frequency-Enhanced Decomposed Transformer (FEDformer) model, and a Frequency-Enhanced Channel Attention Mechanism (FECAM), named FISFA. FISFA converts raw time-series data into the frequency domain to enhance robustness in long-sequence forecasting. Its proposed FEDformer module replaces the standard Transformer to capture multi-scale dependencies with lower computational complexity. However, the method still relies on a deep architecture that incurs significant costs, and its implicit frequency-domain approach fails to explicitly model temporal periodicity.
In summary, existing deep learning-based prediction methods primarily focus on implicitly capturing temporal dependencies in long sequences of resource usage, while overlooking inherent periodic patterns. This failure to explicitly leverage such characteristics for selecting highly relevant historical data not only reduces prediction accuracy but also increases model training complexity.

2.2. Period Detection

Existing periodicity detection algorithms can be broadly classified into two categories: (1) time-domain methods, which identify periodic patterns by analyzing the signal’s self-similarity using the ACF [18,19,45]; and (2) frequency-domain methods, which detect underlying periodicities by transforming the time series into the frequency domain via the Discrete Fourier Transform (DFT) [15,16,17,46], producing a periodogram that represents energy distribution across frequencies. However, ACF-based approaches are sensitive to outliers and noise, often resulting in false alarms, while periodograms suffer from spectral leakage [47], leading to inaccurate or coarse period estimates—especially for long periods.
To mitigate these issues, a number of hybrid methodologies operating in both time and frequency domains have been developed. For instance, AUTOPERIOD [47,48] first extracts a set of candidate periods from the periodogram and subsequently validates them by examining corresponding peaks in the ACF, under the assumption that legitimate periods manifest as prominent ACF hills. Spectral and Average Autocorrelation Zero Distance Density (SAZED) [49], as an ensemble approach, combines spectral density estimation from the frequency domain with autocorrelation zero analysis in the time domain, reinforced through iterative autocorrelation refinement. Another extension [50] employs a sliding window mechanism along with a Kalman filter to adapt to non-stationary time series, though its performance is not universal and it remains vulnerable to outlier corruption. ePeriodicity [51] is designed to explicitly account for uncertainties and noise in periodic signals, delivering provable robustness against incomplete observations, albeit only applicable to binary sequence data. RobustPeriod [20,21] is the state-of-art period detection work. It is a robust multiple periodicity detection framework that decouples intertwined periodicities via Maximal Overlap Discrete Wavelet Transform (MODWT) in the time–frequency domain, and it leverages a robust Huber-periodogram and Huber-autocorrelation function for detection and validation. Q. Wu et al. [52] introduced a periodic pyramid structure that integrates self-attention with multi-scale periodic modeling to capture complex periodic dependencies in time series, achieving strong performance in forecasting and classification tasks. However, it suffers from high computational cost and limited generalization to weakly periodic or non-periodic data.
In the context of online service resource usage in data centers, periodic behaviors often exhibit extended cycle lengths, consistent trend pattern across cycles as well as non-stationary magnitude variations and contaminations such as noise and outliers. Despite advances in hybrid time-frequency methods, current period detection techniques still fall short in fully addressing these complex characteristics, particularly the high computational complexity, underscoring the need for more robust and efficient solutions tailored to this specific domain.

3. Materials and Methods

3.1. Overview of PRPOS

As illustrated in Figure 1, PRPOS adopts a two-phase framework designed for resource prediction in online services. The framework begins by collecting raw resource usage sequence data—such as CPU utilization of online services—from data center infrastructures. In the first phase, targeting the common characteristics of trend consistency and magnitude variations across multiple periods in resource usage, a period detection method combining normalized HP filtering [25], rank-based ACF [53,54], and cosine similarity [55] is employed. This approach accurately identifies the dominant periodic pattern in the resource sequences through candidate period generation and period refinement steps. In the second phase, leveraging the detected periodic structure, PRPOS constructs a prediction model capable of integrating multi-period temporal features. Built around GRUs and Fully Connected Networks (FCNs) [56], the model captures short-term dependencies from recent data (within the current period) and long-term evolutionary patterns from multiple historical periods. Finally, the multi-scale temporal representations are merged through a feature aggregation network to produce accurate predictions of resource usage.

3.2. Period Detection

Period detection in online service resource usage comprises two sequential phases: candidate period detection and period refinement, shown in Figure 2. The candidate period detection phase preserves periodicity through trend decomposition: The original sequence is separated into trend and residual components, with periodic features extracted from residuals to derive preliminary candidates. The refinement phase measures trend similarity between sub-sequences to fine-tune candidates, suppressing spurious patterns caused by magnitude variations and noise while enhancing true periodicity.

3.2.1. Candidate Period Detection

Candidate detection combines HP filtering for cyclical component extraction and rank ACF for period detection.
  • Normalized-Weighted HP Filtering-Based Cyclic Component Extraction
The HP filter decomposes the original resource usage sequence Y into trend ( τ t ) and cyclical ( c t ) components through the canonical formulation y t = τ t + c t , where c t provides purified periodic constituents for subsequent period detection [57]. To enhance robustness against magnitude heterogeneity—i.e., variations in overall resource usage levels across periods—while preserving inter-period trend similarity, we introduce a normalized–weighted adaptation of the HP filter.
More specifically, the proposed method initially performs magnitude normalization as follows:
Y ˜ t = Y t μ Y σ Y
where μ Y represents the mean of the sequence, and σ Y denotes the standard deviation of the sequence.
Subsequently, for the normalized sequence, weights are calculated based on the magnitude of each normalized value using the following formula:
ω t = 1 1 + Y ˜ t
Finally, the HP filter’s objective function is modified as follows:
c t ˜ ^ = argmin c ˜ 1 2 t = 0 T 1 c ˜ t 2 + λ t = 1 T 1 ω t Y ˜ t 1 c ˜ t 1 2 Y ˜ t c ˜ t + Y ˜ t + 1 c ˜ t + 1 2
By normalizing the resource usage sequence to eliminate scale dependence and integrating magnitude-dependent weights ( ω t ) into the smoothing constraints, this formulation explicitly addresses magnitude disparities [58]. Here, λ is tuned via residual statistical matching. The approach simultaneously suppresses the dominance of cycles with high absolute usage levels and preserves inter-cycle trend similarity, a dual mechanism critical for robust period detection in sequences with variable magnitude.
2.
Candidate Period Detection with Rank-Based ACF
Given the long-period characteristics of online service resource usage, we employ time-domain methods for period detection. While traditional ACF is commonly used [18,19], it fails to reliably identify periods in these sequences due to magnitude variations between cycles—even when they share similar trend patterns. To overcome this limitation, we adopt rank-based ACF, which first transforms the HP-filtered cyclical components into normalized ranks, preserving waveform ordering while eliminating magnitude effects.
For the rank transformation of HP-filtered cyclical components c t , t { 1 , 2 , , T } , we define the rank of any element c t as follows:
r t = i = 1 T I I c i < c t + 1 2 i = 1 T I I c i = c t + 1
where I I ( · ) denotes the indicator function [59]. The second term handles tied values by assigning average ranks.
Subsequently, we normalize the ranks to the interval [−1, 1] via
n r t = 2 ( r t 0.5 ) T 1
This approach calculates auto-correlation on rank-transformed values rather than raw magnitudes, which completely eliminates inter-cycle magnitude differences while preserving morphological features. By isolating periodicity information from magnitude variations, the results depend solely on waveform similarity, significantly improving robustness in long-period detection for online services.
Based on the normalized rank transformation, we employ Algorithm 1 to compute subsequent periods in online service resource usage.
Algorithm 1 Candidate Period Identification
Input: 
Normalized rank sequence of online service resource usage: N R
Output: 
Candidate period: p
  1:
A C F S e q generateACF ( N R ) fill // Generate ACF sequence
  2:
L o c a l P e a k s findLocPeak ( A C F S e q ) fill // Find local peaks
  3:
p median ( distance ( L o c a l P e a k s ) ) fill // Choose the median distance as period
  4:
return p
Statistical analysis of Alibaba cluster trace reveals that online service resource usage typically exhibits consistent daily periodic patterns, despite large fluctuations in magnitude. Traditional ACF is highly sensitive to magnitude variations and noise spikes, often requiring error-prone manual threshold tuning. In contrast, Algorithm 1 automatically identifies significant peaks in the auto-correlation of rank-transformed data and determines the period using median peak spacing. This approach eliminates subjective thresholds and reliably detects periods even with strong inter-cycle variations.

3.2.2. Candidate Period Refinement

While our candidate period detection mechanism effectively identifies the candidate period, the preliminary result only indicates possible cycle locations through rank-based ACF peaks without verifying waveform stability in actual resource usage sequences. This necessitates a refinement phase to validate the periodicity through waveform pattern analysis.
The primary principle underlying the period refinement is to partition the original resource usage sequence of online services into several equally sized subsequences, based on the candidate period length, and measure the similarity between these partitioned subsequences. By fine-tuning the candidate period length, we select the one that attains the utmost similarity among the subsequences as the ultimate dominant period.
Considering the value range variations across time periods in online service’s resource usage sequence, we choose cosine similarity as the subsequence similarity measurement. Cosine similarity is commonly employed to gauge the resemblance between two sequences, primarily emphasizing the coherence in their trends rather than their absolute values. A higher cosine similarity value indicates a greater resemblance between the two sequences. For the candidate period, trendSimilarity is defined as the average pairwise cosine similarity among all partitioned subsequences, expressed as
t r e n d S i m i l a r y = k = 1 , l = 1 m c o s i n e s i m i l a r i t y ( X k , X l ) m ,   k l
c o s i n e s i m i l a r i t y ( X k , X l ) = i = 1 n x i k · x i l i = 1 n ( x i k ) 2 i = 1 n ( x i l ) 2
where n represents the candidate period length; X k and X l respectively represent two different resource usage subsequences; x i k , x i l denote the value of the corresponding points of these two subsequences, i = 1 , 2 n . The corresponding similarity of each fine-tuned candidate period is expressed as the average of similarity scores among all pairs of subsequences.
The tuning of the candidate period is processed as Algorithm 2.
Algorithm 2 Dominant Period Refinement
Input: 
resource usage sequence of online service: R S ,
step size of period refinement: Δ t ,
terminal condition of period refinement: s n ;
Output: 
the ultimate dominant period: f i n a l C P
  1:
ascCP 0 , decCP 0 , finalCP 0 fill // Initialization
  2:
ascCP findascCP ( R S , C P , Δ t , s n ) fill // Find candidate period ≥ CP
  3:
decCP finddecCP ( R S , C P , Δ t , s n ) fill // Find candidate period ≤ CP
  4:
finalCP compareSimilarity ( ascCP , decCP )
  5:
return finalCP
The refinement of the candidate period is based on two underlying assumptions: First, that the candidate period, obtained through the processing of HP and ACF, approximates the eventual dominant period within a few refinement iterations; second, that as the candidate period approaches the ultimate dominant period, the corresponding subsequences display an increasing level of similarity. The refinement process is carried out in two directions, either increasing or decreasing the candidate period by a step size of Δ t . The input s n defines the terminal condition for period refinement. Based on the second assumption, the bidirectional refinement continues until the subsequence similarity shows a continuous decrease over sn consecutive instances (line #2, #3). Once this refinement process is completed, the final dominant period will be chosen from the identified candidate dominant periods in both directions with the higher trend similarity (line #4).

3.3. Resource Prediction

Based on the detected period, PRPOS selectively chooses the relevant historical data to explore the patterns of resource usage variation within and between periods of online services, facilitating accurate resource prediction. In PRPOS, we employ the Gated Recurrent Unit (GRU) neural network as a fundamental component to capture the patterns of variation in resource usage time-series data. These patterns are subsequently combined to generate the final resource prediction.
  • Historical Data Selection
Based on the periodic patterns in online service resource usage, we observe that historical data typically contains two components which facilitate the prediction of future demand.
As depicted in Figure 3, the first component, referred to as in-period-data, consists of the resource usage subsequence closest to the prediction time within the most recent period. This component captures short-term dependencies between predicted usage and historical observations. The second component, termed cross-period-data, comprises resource usage subsequences from the same or temporally adjacent points within each historical period relative to the prediction point. This component is essential for two reasons: first, resource usage at corresponding time positions across different periods shows strong similarity; second, incorporating adjacent time points helps represent variation patterns within the value range over periodic intervals. Importantly, this cross-period-data is crucial for modeling long-term dependencies in resource usage behavior.
B.
Model design
Figure 4 demonstrates the resource prediction model in PRPOS.
GRU as Building Block. In designing PRPOS, we deliberately selected Gated Recurrent Unit (GRU) as the fundamental building block for temporal modeling, in preference to more recent alternatives such as Transformer-based architectures. This decision was motivated by several domain-specific considerations. Resource usage in online services (e.g., CPU utilization) is predominantly influenced by immediate past states and, despite the occasional presence of outliers or abrupt changes, typically exhibits smooth and sequential variations rather than arbitrary long-range dependencies. Although Transformers excel at capturing global interactions through self-attention, their quadratic inference complexity makes them less suitable for rapidly updating predictions in resource-constrained, latency-sensitive environments such as online prediction systems. Furthermore, the ability of Transformers to leverage long sequences does not necessarily translate into accuracy gains for online service resource prediction, where recent points and homologous periodic segments carry the most relevant signal.
As shown in Figure 5, with their gated recurrent mechanism, GRUs provide a balanced solution: They mitigate the vanishing gradient problem of simple Recurrent Neural Networks (RNNs) and efficiently model medium-range dependencies while maintaining linear computational complexity and a modest parameter count [9,14]. This makes GRUs particularly well-suited for building lightweight yet expressive temporal feature extractors that can be integrated into larger prediction workflows without introducing unnecessary latency or resource consumption.
Model Architecture. The prediction model in PRPOS is mainly composed of three modules: in-period-GRU, cross-period-GRU, and aggre-FCN.
The in-period-GRU module processes the in-period-data to capture short-term temporal dependencies in resource usage. Specifically, the input is represented as a vector of length w, corresponding to the most recent w time steps within the current period. The vector is processed through a network of GRU units, each handling one time step. These units share parameters, allowing the module to capture sequential dependencies efficiently. The module outputs an m-dimensional feature vector representing the encoded short-term context, where m is a trainable hidden size hyperparameter.
The cross-period-GRU module is designed to capture long-term temporal dependencies from historical resource usage patterns. Its input, the cross-period-data, is structured as a matrix of size k × h , where k denotes the number of historical periods considered and h is the length of the aligned usage subsequence within each period. Each row of the matrix corresponds to a contiguous resource usage subsequence drawn from a historical period. These subsequences are aligned based on the target prediction time—that is, each row starts at the same phase position within its respective period as the current prediction point. The rows are arranged in reverse chronological order, with the most recent period placed first and the earliest period last. Each row is processed independently by a shared GRU unit—that is, the same GRU cell with identical parameters is applied to every row, enabling efficient and consistent feature extraction across all historical sequences. The module ultimately produces an n-dimensional output vector, where n is a learnable hidden size hyperparameter, encoding the long-term temporal features across multiple periods.
The latent representations of both short-term and long-term temporal dependencies—that is, the output vectors from the in-period-GRU and the cross-period-GRU—are concatenated into a unified feature vector. This combined representation serves as the input to the aggre-FCN module. The aggre-FCN consists of a fully connected network that learns to adaptively weight and fuse the short-term and long-term features. Through non-linear transformation and dimensionality reduction, this module effectively synthesizes the multi-scale temporal information to produce the final resource usage prediction.
Model Training. Let the resource usage time series be denoted as R = { r t } for 0 t n . Training samples are constructed for each prediction point using an endpoint-based partitioning scheme: We start at time t and extract backward to form contiguous subsequences of length P (the detected period). This approach guarantees that the most recent period—ending exactly at the prediction point—contains a complete set of data points, thereby avoiding truncated or incomplete inputs that would occur if we partitioned from the start of the sequence.
From these period-length subsequences, we derive the two input components. First, the in-period-data comprises the most recent w time steps ending at t. Second, the cross-period-data consists of k segments of length h, each aligned to end at time t across k historical periods. A sample is valid only if at least k historical periods are available. The overall dataset is split chronologically into training and testing sets in a 7:3 ratio. This preserves the temporal order of events and allows us to evaluate the model’s ability to generalize to future time periods, providing a more realistic assessment of its deployment performance. The model is trained to minimize the Mean Squared Error (MSE) [60]:
m i n i = 0 m 1 ( y i y i ) 2 m
In addition, we use the Adam optimizer with learning rate decay, ReLU activations [61], and employ dropout and L 2 regularization to mitigate overfitting.

4. Results

In this section, we demonstrate the superior performance of our proposed PRPOS method in resource prediction and period detection using real-world and synthic data center traces. The experimental results are presented graphically in the main text, with the complete numerical data provided in tabular form in Appendix B.

4.1. Experiment Settings

  • Datasets
To evaluate the resource prediction capability of PRPOS, we utilize the Alibaba cluster trace v2018 dataset. Released in 2018, this dataset offers comprehensive insights into data center operations by recording the resource usage of over 5000 online services during an eight-day period. Among its six constituent tables, we specifically employ c o n t a i n e r _ m e t a , which contains service metadata, including the status field used to identify normally functioning services, and c o n t a i n e r _ u s a g e which includes CPU, memory, and disk measurements recorded at 30-s intervals. After selecting 40 services with normal status from c o n t a i n e r _ m e t a , we aggregate their resource usage sequences by c o n t a i n e r _ i d in c o n t a i n e r _ u s a g e . We focus on CPU utilization percentage as the primary prediction metric due to its representative nature. For each service, we extract timestamp-sorted CPU utilization percentages to form individual time series, which together constitute our experimental dataset. Each service trains a dedicated resource prediction model using the data partitioning approach described in the "Model Training" subsection in part B of Section 3.3.
To comprehensively evaluate PRPOS’s period detection under scenarios with both trend consistency and magnitude heterogeneity, we construct a synthetic dataset based on statistical characteristics from the Alibaba cluster trace v2018. This dataset enforces trend consistency by maintaining identical resource usage change patterns across consecutive cycles. Periodic patterns are modeled as standard sine waves with period lengths measured in number of 30-s data intervals. The lengths span five ranges—[800–1000), [1000–1200), [1200–1400), [1400–1600), and [1600–1820] intervals—systematically covering the dominant period length distribution in the dataset. For each of the five configurations, we generate 100 CPU usage percentage sequences, each containing 8 cycles. Magnitude heterogeneity is introduced into these sequences through two key dimensions: baseline levels, representing the steady-state resource occupancy and ranging from 5% to 50%, and fluctuation intensity, characterized by ±15% inter-cycle variations. Additionally, we randomly inject Gaussian noise ( σ n 2 = 0.1 / 1 / 2 ) and outliers ( θ = 0.01 / 0.1 / 0.2 ) into each sequence.
Our experiments are conducted on one server equipped with one 4-core 3.3 GHz Intel Core i5-6600 processor, 32 GB memory, and a 2TB Hard Drive HDD.
B.
Baselines
The experimental baselines were constructed by adapting existing competitive approaches for cloud resource prediction, each employing statistical, machine learning, or deep learning architectures.
  • ARIMA [6]: Combines wavelet decomposition and ARIMA for short-term trend prediction but struggles with nonlinear and complex patterns.
  • XGBoost [8]: Uses CNN to optimize tree weights, boosting accuracy but increasing computational complexity.
  • CNN [62]: Captures local temporal patterns via sliding kernels but fails to model long-term periodic dependencies due to fixed receptive fields.
  • LSTM [63]: Robust to noise and temporal dependencies but suffers from high computational complexity and data hunger for intricate patterns.
  • BiLSTM [64]: Enhances context modeling with bidirectional processing but inherits LSTM’s high cost and interpretability issues.
  • GRU [65]: More parameter-efficient than LSTM but limited in parallelization and long-range dependency modeling.
  • GRU+Attention [14]: Attention mechanisms alleviate GRU’s long-range issues but introduce full-sequence processing overhead and noise sensitivity.
  • RPTCN [40]: Combines dilated convolutions and attention for long-range dependencies but struggles with periodic phase shifts and irrelevant feature noise.
  • VSSW-MSDSS [43]: Dynamically selects among statistical, ML, and DL models via variable-size sliding windows but relies on manual threshold tuning and lacks explicit periodicity modeling.
  • FISFA [44]: Leverages SG filter, a FEDformer, and FECAM for spectral patterns, but incurs heavy computation and implicit periodicity handling.
In addition, we evaluate the accuracy and computational efficiency of our period detection method in PRPOS, comparing it to widely adopted frequency-domain and time-domain baselines.
  • findFrequency [66]: This is based on using the maximum value in the frequency spectrum to estimate period length.
  • SAZED [49]: This approach includes the SAZEDopt and SAZEDmaj variants, which combine spectral and autocorrelation analysis for parameter-free seasonality detection with high efficiency. However, it assumes stable amplitudes and exhibits sensitivity to outliers, owing to its reliance on raw autocorrelation.
  • RobustPeriod [20,21]: This leverages wavelet transforms and robust Huber optimization to detect multiple periodicities. It suffers from high computational costs and the phase-amplitude decoupling, making it less ideal for magnitude-varying resource patterns.
C.
Evaluation Metrics
To quantify and compare the prediction performance of PRPOS with the ten baselines, we use two performance metrics—MAPE [67] and RMSE [68]—as follows:
M A P E ( y i , y i ) = 1 n i = 1 n y i y i y i
R M S E ( y i , y i ) = 1 n i = 1 n ( y i y i ) 2
where y i and y i are the observed value and ground truth of instance i, and n is the number of all sample instances.
In period detection evaluation, we use Accuracy to measure their performance.
A c c u r a c y ( y i , y i ) = 1 n i = 1 n y i y i
D.
Parameter Settings
In this section, we present the core parameter settings in PRPOS. The detailed parameter settings for each baseline model are provided in Appendix A (Table A1 and Table A2).
The configurable parameters in PRPOS are categorized into three functional groups corresponding to the stages of period detection, sample data selection, and resource prediction modeling. It is noteworthy that the parameters for the prediction modeling stage are specifically the parameters of the deep learning model, as shown in Table 1.
In the period detection stage, the HP filter smoothing parameter ( λ = 14,400) adopts the classical setting to extract near-daily periodic trends [25]. The rank-based ACF is computed with two key parameters: the window length L and the maximum lag L . In our experiments, we set L = 3000 and L = 4000 so as to fully encompass the characteristic near-daily periodicity (6.5–18 h in Alibaba trace) and its potential harmonics of online service CPU usage [26]. The cosine similarity threshold τ _ c is set to 0.8 to enforce strict waveform consistency against the inherent burstiness and noise in online service CPU usage. During period refinement, the step size Δ t is set to the minimum sampling interval (30 s) in the trace data to achieve the finest adjustment granularity, while the terminal condition sn is set to 5 to serve as a robust stopping criterion, effectively distinguishing between short-lived noise spikes and a genuine, persistent shift in the periodic pattern of CPU consumption.
In the sample data selection stage, parameters are set to model the continuous trends and mitigate transient noise in online service’s CPU usage. The input data size N is set to 60 to capture essential short-term context. The sampled window size per cycle W is set to 10 to provide the resolution needed to represent pattern morphology. The number of historical cycles K is set to 5 to offer a robust view of long-term, repeating trends, allowing the model to average out non-repeating noise.
In the resource prediction modeling stage, the model structure is defined by several key parameters. The fully connected network has a depth L fc = 2 with neuron numbers N fc 1 = 128, N fc 2 = 64, balancing capacity and stability. The GRU hidden unit number d h = 64 is chosen for effective temporal modeling. The core design is the dual-scale GRU, where the in-period-GRU hidden layer depth L ip = 1 captures high-frequency dynamics, and the cross-period-GRU hidden layer depth L cp = 2 models complex long-term trends. The learning rate α = 0.005 ensures stable convergence.

4.2. Results and Analysis

  • Overall Performance Evaluation
To evaluate the resource prediction performance of PRPOS under diverse utilization patterns, we classified the CPU usage sequences of online services from Alibaba cluster trace v2018 dataset based on their average cosine similarity of periodic patterns obtained through PRPOS’s period detection method. These sequences were grouped into four categories according to similarity ranges—[0.8, 0.85] (Cos-Low), [0.85, 0.9] (Cos-Medium-Low), [0.9, 0.95] (Cos-Medium-High), and [0.95, 1] (Cos-High)—each reflecting a distinct periodic characteristic. From each category, ten sequences were randomly selected for experimentation. To ensure a fair comparison, all baseline methods use the 60 most recent data points as input, matching the total historical data volume of PRPOS, which integrates 10 points from the current cycle and 5 historical cycles (10 points each).
Figure 6 illustrates the overall performance of PRPOS and baseline methods. PRPOS consistently outperforms all baselines across all four test groups, achieving the best results in online service resource prediction. Specifically, it attains an average improvement of 45.3% and a maximum of 80% in MAPE, along with an average improvement of 44.3% and a maximum of 81.4% in RMSE. PRPOS achieves superior performance by explicitly modeling periodicity in resource usage. By intelligently incorporating both short-term dependencies (from the current cycle) and long-term dependencies (from historical cycles), PRPOS effectively captures both immediate variations and recurring patterns. This dual-temporal approach proves more accurate than methods that either ignore periodic features or rely solely on transformer-based dependency modeling.
We further observe that the performance improvement of PRPOS becomes more pronounced as the trend similarity across cycles increases. As illustrated in Figure 7, which shows the relative improvement in both MAPE and RMSE over baseline methods, PRPOS achieves significantly greater accuracy gains in groups with higher cosine similarity. For example, compared to the GRU-based model, as similarity increases from [0.8, 0.85] to [0.95, 1], PRPOS attains relative reductions in MAPE ranging from 32.3% to 38.5%, and in RMSE from 27.8% to 30.8%. This trend highlights the advantage of periodic pattern modeling: Higher inter-cycle trend consistency implies more stable and predictable resource usage, thereby improving prediction accuracy.
To evaluate the impact of input length on model performance, we maintained PRPOS’s fixed 60-data-point input architecture (comprising the 10 most recent observations plus 5 × 10 phase-aligned historical points) while progressively increasing the sequence lengths of baseline deep learning models to assess their capacity for leveraging extended historical contexts. Specifically, we extended the input length to 72, 96, 128, and maximum feasible lengths (denoted as “ M a x ”) for all selected baselines except VSSW-MSDSS. Considering the computational constraint during model training, the maximum lengths were configured as 200 for CNN-based model, 256 for RNN-style models and RPTCN, and 512 for FISFA. We employ the Cos-Medium-Low and Cos-Medium-High service groups for this set of the experiments.
As shown in Figure 8, prediction accuracy does not improve monotonically with longer historical windows. PRPOS consistently outperforms baselines even under the same historical data scale. This is because (1) overly short windows in baseline models fail to capture short-term patterns, while (2) excessively long windows introduce noise and outliers that impair prediction. In contrast, PRPOS maintains robust performance by selectively retaining the most relevant historical data and explicitly modeling inter-cycle trends.
Finally, to validate the robustness of PRPOS, we evaluated its predictive performance on the Cos-Low and Cos-High service groups by varying two key parameters: the sampled window size per cycle and the number of historical cycles used for prediction.
In our experiments, we use the configuration of 10 samples per cycle and 5 historical cycles, which has been employed in our preceding experiments, as the baseline. As illustrated in Figure 9 and Figure 10, the model demonstrates notable robustness. Both MAPE and RMSE remained stable as the sample window size per cycle increased from 8 to 64. A noticeable performance degradation occurred only when the window size was drastically reduced to 4, which resulted in a MAPE increase of up to 20% compared to the baseline. For all other configurations, the deviations in MAPE and RMSE were confined within 3.75% and 2.28%, respectively. Similarly, varying the number of historical cycles from 3 to 7 induced marginal deviations within 7.5% for MAPE and 6.63% for RMSE. However, reducing the cycles to 2 caused a sharp performance decline, increasing MAPE and RMSE by approximately 28.75% and 44.57%, respectively. These results confirm that PRPOS exhibits strong robustness across a wide range of parameter settings, provided they are chosen within appropriate bounds.
Furthermore, compared to the baseline configuration, increasing either parameter did not yield significant performance gains. This suggests that the default setup offers a favorable trade-off between prediction accuracy and computational efficiency.
B.
Period Detection Performance
In this section, we evaluate PRPOS’s novel period detection method, specifically optimized for time series data characterized by consistent temporal trends and heterogeneous magnitude patterns. The experimental dataset and baseline methods are described detail in part A of Section 4.1.
Figure 11 summarizes the accuracy of all methods across different period lengths and under varying degrees of magnitude heterogeneity. PRPOS consistently achieves the highest period detection accuracy in all scenarios. By applying normalized HP filtering and rank-based ACF to mitigate inter-cycle magnitude variations, PRPOS significantly outperforms both SAZED and findFrequency across different period lengths, with an average accuracy improvement of 109.6% and a maximum improvement of 240%.
We further analyze two selected subsets from the 500 synthetic resource usage sequences: a light-variation subset (baseline: 5–10%, fluctuation intensity: 1–10% of the baseline) and a heavy-variation subset (baseline: >10%, fluctuation intensity > 10% of the baseline). Our experimental results demonstrate that PRPOS’s period detection method achieves accuracy improvements of 179.7% on average(up to 304.8%) over SAZED and findFrequency on these subsets. Notably, in experiments with the heavy-variation group compared to the light-variation group, PRPOS achieves higher accuracy improvement, validating the effectiveness of normalization and rank transformation operations.
Among the four baseline methods, RobustPeriod achieves the best detection performance, matching PRPOS in accuracy thanks to its MODWT decomposition, Huber robust estimation, and joint time–frequency validation. However, computational cost comparisons reveal significant differences. As shown in Table 2, PRPOS maintains sub-second execution times across all period lengths, while RobustPeriod’s runtime grows sharply with longer periods—reaching up to 3788× slower than PRPOS for periods in [1600, 1820]. This efficiency gap stems from their underlying mechanisms: PRPOS leverages low-complexity normalized HP filtering and rank-based ACF, which exploit monotonic cosine similarity to terminate corrections early, whereas RobustPeriod incurs higher overhead from DWT-based frequency decomposition and DFT-based heuristic searches. Notably, findFrequency incurs the lowest computational cost among all methods but suffers from a 2.4-fold reduction in accuracy, attributable to spectral leakage resulting from its simplistic frequency-domain approach.
C.
Ablation Study
To thoroughly assess the contributions of individual components in the PRPOS prediction model, we perform ablation studies comparing the complete model with its variants, with results detailed in Figure 12. All ablated versions show measurable accuracy degradation compared to the full PRPOS model, which achieves superior performance through its dual capability to perform (1) robust period detection via distribution-resistant normalization and period correction to handle magnitude variations and (2) comprehensive dependency modeling that simultaneously captures both short-term in-period relationships and long-term cross-period patterns across historical sequences.
Effectiveness of the period detection method: We conduct two ablation studies on the three key components in our period detection method. First, replacing normalized HP filtering and rank-based ACF with conventional HP filtering and standard ACF (w/o HA) results in an average increase of 17.17% in MAPE and 12.2% in RMSE across all datasets, confirming the importance of distribution-resistant normalization in reducing period detection noise. Removing the cosine similarity-based refinement component (w/o PR) causes a milder degradation, with MAPE increasing by 8.6% and RMSE by 5.9%, indicating its role in providing additional noise suppression beyond the baseline detector.
Effectiveness of the resource prediction model: To evaluate the effectiveness of our resource prediction model, which consists of in-period and cross-period components, we construct two ablation variants. The w/o IP variant is constructed by removing in-period from the model while retaining cross-period’s structure and inputs; results show that this leads to a 31.7% increase in MAPE and a 28.9% rise in RMSE, underscoring the importance of short-term historical patterns. The w/o CP variant removes the cross-period component and extends in-period’s input to the latest 60-step history; this configuration fails to capture long-term periodic features—even with matched input length—causing a further 55.4% increase in MAPE and a 46.1% rise in RMSE.
Effectiveness of Sequential Modeling Units: To validate the effectiveness of using GRUs for sequential modeling in both the in-period and cross-period components, we constructed two ablation variants—I/CP-GRU→CNN and I/CP-GRU→Attention—replacing the GRU units with convolution layers and a standard self-attention module (with positional encoding), respectively. Compared to the complete model, the CNN variant shows an average increase of 18.7% in MAPE and 7.3% in RMSE, confirming that GRU outperforms CNN in capturing temporal trends and short-term dependencies. However, replacing GRU with a self-attention mechanism also degrades performance (MAPE: 9.9%, RMSE: 3.8%). This is because our periodicity-based approach splits fixed-length sequences into historical cycles and selects phase-aligned short sub-sequences, negating the self-attention mechanism’s long-sequence weighting. Instead, GRU’s strength in local short-term dependencies prevails. Additionally, GRU’s gating mechanisms adaptively filter noise/outliers for robustness to imperfect data, whereas the attention mechanism is more susceptible to such interference.

5. Discussion

The experimental results in Section 4 affirm the efficacy of PRPOS. This section discusses the core insights behind its success, directly addresses its generalization potential, and outlines its inherent boundaries to guide future works.

5.1. The Value of Explicit Periodicity Modeling

The superior performance of PRPOS stems from its fundamental premise: Periodicity is a primary, structural characteristic of online services that warrants explicit modeling. While deep learning baselines implicitly learn temporal dependencies, their lack of architectural inductive bias for periodicity makes them susceptible to noise and less efficient at leveraging long-range, stable patterns. PRPOS compels the model to distinguish between short-term fluctuations and cyclical trends through its dedicated detection and dual-scale prediction architecture. This approach proves particularly powerful in cloud environments, where identifying the true signal amidst operational noise is crucial for robust predictions.

5.2. Generalization, Boundaries, and Future Paths

Our evaluation focused on CPU utilization because it most saliently exhibits the strong, consistent periodicity that PRPOS is designed to capture. A preliminary analysis of the Alibaba trace revealed that memory and I/O usage for the sampled online services often lacks such clear periodic patterns, instead demonstrating more stochastic and aperiodic characteristics within this dataset. Furthermore, we selected an 8-day evaluation window because it typically encompasses multiple full cycles of the dominant periodicity in online services. This duration thus provides a robust testbed for validating our method’s core principle: the ability to identify and leverage stable periodic patterns.
Consequently, we posit that the framework’s generalization to other resource types depends primarily on the presence of a dominant periodic trend in the target resource. For resources that do exhibit such morphological patterns, our method remains applicable because its detection mechanism is agnostic to absolute values. The current boundaries of PRPOS lie in its single-period architecture and offline detection scheme. Future work will therefore focus on handling multi-periodicity and achieving online adaptation. Developing such an adaptive mechanism is key to addressing longer time series, as it would enable the model to continuously track and adjust to long-term concept drift, where underlying periodicity may gradually evolve.

6. Conclusions

In this paper, we presented PRPOS, a novel periodicity-aware framework for accurate resource prediction in online services. To address the critical yet under-exploited challenge of long-term periodic patterns with consistent trends but significant magnitude variations, PRPOS introduces a two-phase methodology: a robust period detection mechanism combining normalized-weighted HP filtering, rank-based ACF and cosine similarity measurement, followed by a dual-scale GRU predictor that explicitly leverages the detected period to model both in-period and cross-period dependencies.
Extensive evaluations on the Alibaba cluster trace demonstrate that PRPOS achieves an improvement over state-of-the-art baselines, with an average reduction of 45.3% in MAPE and 44.3% in RMSE. These quantitative results solidly validate the efficacy of explicitly integrating periodicity into the resource prediction pipeline.
The practical value of PRPOS lies in its ability to provide cloud operators with more accurate and robust predictions for dynamic resource provisioning. By enabling more informed autoscaling and scheduling decisions, our framework paves the way for enhanced QoS assurance and cost reduction in cloud data centers. In future work, we will extend PRPOS to handle multi-periodic patterns and integrate its predictions into real-world scheduling systems.

Author Contributions

Conceptualization, Y.L. and H.Z.; Data curation, Y.L. and H.Z.; Formal analysis, Y.L. and H.Z.; Funding acquisition, Y.L.; Investigation, Y.L., H.Z., and T.L.; Methodology, Y.L. and H.Z.; Resources, H.Z. and H.S.; Software, H.Z. and H.S.; Validation, Y.L. and H.Z.; Visualization, H.Z.; Writing—original draft, Y.L. and H.Z.; Writing—review and editing, Y.L., H.Z., T.L., and H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant Number: 62276011).

Data Availability Statement

The dataset used for the experiment is publicly available at the following URL: https://github.com/alibaba/clusterdata/blob/v2018/cluster-trace-v2018/trace_2018.md (accessed on 1 July 2025).

Acknowledgments

We gratefully acknowledge Jin Yi for his valuable assistance with writing—review and editing—for his very quick reactions to the many questions asked by the authors, and for generously providing the computing resources necessary to conduct the experiments in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ACFAutocorrelation Function
ARIMAAutoregressive Integrated Moving Average
BiLSTMBi-directional Long Short-Term Memory
CNNConvolutional Neural Network
CPUCentral Processing Unit
DBNDeep Belief Network
DP-CUPADBN and PSO based CPU Usage Prediction Algorithm
FCNFully Connected Network
GPRGaussian Process Regression
GRUGated Recurrent Unit
HPHodrick–Prescott
LSTMLong Short-Term Memory
MAPEMean Absolute Percentage Error
MLPMulti-Layer Perceptron
MLRMultiple Linear Regression
PRPOSPeriodicity-based Resource Prediction of Online Services
PSOParticle Swarm Optimization
QoSQuality-of-Service
RMSERoot Mean Squared Error
RPTCNResource Prediction based on Temporal Convolutional Networks
SAZEDSpectral and Average Autocorrelation Zero Distance Density
SGSavitzky Golay
SVM                 Support Vector Machine
SVMRSupport Vector Machine Regression
VMDVariational Mode Decomposition
VSSW-MSDSSVariable-Size Sliding Window-Model Selector Decision Support System
XGBExtreme Gradient Boosting

Appendix A

Detailed parameter configurations for each baseline method are provided in Appendix A. Since VSSW-MSDSS employs multiple models, Table A2 is dedicated exclusively to presenting the parameter settings for this method.
Table A1. Parameter settings in baseline methods.
Table A1. Parameter settings in baseline methods.
ModelNameValue
LSTMHistorical observation data window size60
Number of hidden units128
Batch size128
Learning rate0.005
BiLSTMHistorical observation data window size60
Number of hidden units128
Batch size128
Learning rate0.001
CNNHistorical observation data window size60
Number of convolutional layers2
Stride1
Kernel size3
Learning rate0.005
RPTCNHistorical observation data window size60
Number of convolutional layers3
Stride1
Kernel size3
Learning rate0.001
Dilation rates of the three convolutional layers1, 2, 4
GRUHistorical observation data window size60
Number of hidden units128
Batch size64
Learning rate0.002
GRU+AttentionHistorical observation data window size60
Number of hidden units60
Batch size50
Learning rate0.001
Number of attention head1
Attention output dimension60
FISFADimension of model60
Learning rate0.001
Layer number in encoder1
Layer number in decoder2
Batch size16
Early stopping patience9
Table A2. Parameter settings in VSSW-MSDSS.
Table A2. Parameter settings in VSSW-MSDSS.
ModelNameValue
ARIMAp—The order of the autoregressive model3
d—The degree of differencing1
q—The order of the moving-average model2
GPRKernel functionRadial Basis Function
Length scale1.5
Noise level0.1
MLRRegularization parameter[0.001, 10.0]
MLPHidden layers3
Neurons per layer32, 64, 128
Learning rate0.001
Batch size64
SVMRKernel functionRadial Basis Function
Regularization parameter1.0
Gamma0.1
LSTMNumber of hidden units128
Batch size64
Learning rate0.001
MSDSSEvaluation indicatorsTrue Positive Rate (TPR),
True Positive Rate (FPR),
True Negative Rate (TNR),
False Negative Rate (FNR),
Accuracy
Classification methodsDecision tree classifier,
KNN,
Naive Bayes

Appendix B

The tables in Appendix B present the detailed numerical data tables corresponding to all figures featured in the main text. A guide linking these tables to their corresponding figures is provided in Table A3.
Table A3. Guide linking appendix tables to their corresponding figures.
Table A3. Guide linking appendix tables to their corresponding figures.
FigureAppendix Table
Figure 6Table A4 and Table A5
Figure 7Table A6 and Table A7
Figure 8Table A8 and Table A9
Figure 9Table A10
Figure 10Table A11
Figure 11Table A12
Figure 12Table A13 and Table A14
Table A4. MAPE of PRPOS and baseline methods.
Table A4. MAPE of PRPOS and baseline methods.
ModelCos-LowCos-Medium-LowCos-Medium-HighCos-High
ARIMA0.4410.3290.3090.341
XGBoost0.2420.2280.2290.211
CNN0.180.1670.1830.172
BiLSTM0.1350.1380.1330.141
LSTM0.1440.1520.1540.143
GRU0.130.1320.1340.13
GRU+Attention0.1210.1240.1140.116
RPTCN0.1180.120.1080.115
VSSW-MSDSS0.2310.30.2790.288
FISFA0.110.1150.1090.108
PRPOS0.0880.0870.0840.08
Note: For each group, the lowest MAPE is in bold and the highest is underlined.
Table A5. RMSE of PRPOS and baseline methods.
Table A5. RMSE of PRPOS and baseline methods.
ModelCos-LowCos-Medium-LowCos-Medium-HighCos-High
ARIMA4.454.24.564.52
XGBoost2.482.122.283.04
CNN1.911.781.991.81
BiLSTM1.421.371.391.48
LSTM1.511.591.381.87
GRU1.331.391.291.33
GRU+Attention1.221.251.251.28
RPTCN1.21.221.21.21
VSSW-MSDSS2.42.82.693.24
FISFA1.151.21.191.3
PRPOS0.960.930.850.92
Note: For each group, the lowest RMSE is in bold and the highest is underlined.
Table A6. MAPE improvement of PRPOS over baseline models.
Table A6. MAPE improvement of PRPOS over baseline models.
ModelCos-LowCos-Medium-LowCos-Medium-HighCos-High
ARIMA80.0573.5672.8276.54
XGBoost63.6461.8463.3262.09
CNN51.1147.9054.1053.49
BiLSTM34.8136.9636.8443.26
LSTM38.8942.7645.4544.05
GRU32.3134.0937.3138.46
GRU+Attention27.2729.8326.3231.03
RPTCN25.4227.5022.2230.43
VSSW-MSDSS61.9071.069.8972.22
FISFA20.024.3422.9325.92
Note: For each group, the highest MAPE improvement is in bold.
Table A7. RMSE improvement of PRPOS over baseline models.
Table A7. RMSE improvement of PRPOS over baseline models.
ModelCos-LowCos-Medium-LowCos-Medium-HighCos-High
ARIMA78.4377.8681.3679.65
XGBoost61.2956.1362.7269.74
CNN49.7447.7557.2949.17
BiLSTM32.3932.1238.8537.84
LSTM36.4241.5138.4150.80
GRU27.8233.0934.1130.83
GRU+Attention21.3125.6032.0028.13
RPTCN20.0023.7729.1723.97
VSSW-MSDSS60.0066.7968.4071.60
FISFA16.5222.5028.5729.23
Note: For each group, the highest RMSE improvement is in bold.
Table A8. MAPE performance under different input data size for Cos-Medium-Low group.
Table A8. MAPE performance under different input data size for Cos-Medium-Low group.
Model607296128Max
CNN0.180.1920.1890.2110.231
RPTCN0.1180.1070.1120.120.106
BiLSTM0.140.1210.1280.1230.138
LSTM0.1440.1420.1320.1380.147
GRU0.1350.1220.1250.1270.142
GRU+Attention0.1210.1160.1070.1040.111
VSSW-MSDSS0.2310.2250.2290.2360.224
FISFA0.110.1020.0980.0890.084
PRPOS0.0880.0880.0880.0880.088
Note: For each group, the lowest MAPE is in bold and the highest is underlined.
Table A9. MAPE performance under different input data size for Cos-Medium-High group.
Table A9. MAPE performance under different input data size for Cos-Medium-High group.
Model607296128Max
CNN0.1720.1640.1760.1810.189
RPTCN0.1150.1080.1030.1050.084
BiLSTM0.1410.1310.1250.1280.133
LSTM0.1430.1370.1320.1380.14
GRU0.1380.1270.1280.1310.14
GRU+Attention0.1160.110.1020.1080.096
VSSW-MSDSS0.2880.2810.2780.2850.274
FISFA0.1080.0960.0920.0870.082
PRPOS0.080.080.080.080.08
Note: For each group, the lowest MAPE is in bold and the highest is underlined.
Table A10. Performance variations of PRPOS under different sampled window size per cycle.
Table A10. Performance variations of PRPOS under different sampled window size per cycle.
Sampled Window SizeCos-LowCos-High
MAPERMSEMAPERMSE
40.1021.220.0961.15
80.0910.980.0820.936
100.0880.960.080.92
160.08610.9480.0770.901
320.08810.970.0790.92
640.08520.9650.0820.941
Table A11. Performance variations of PRPOS under different number of historical cycles.
Table A11. Performance variations of PRPOS under different number of historical cycles.
Historical Cycles NumberCos-LowCos-High
MAPERMSEMAPERMSE
20.1061.3510.1031.33
30.0941.080.0860.981
50.0880.960.080.92
60.0860.9310.080.93
70.0850.9720.0780.912
Table A12. Period detection accuracy of PRPOS and baseline methods.
Table A12. Period detection accuracy of PRPOS and baseline methods.
ModelPeriod LengthsLevels of Noise
400800120016001820Light-VariationHeavy-Variation
findFreque50.1248.0642.4632.1425.0635.2121.03
SAZEDmaj64.0754.6447.3538.1132.5736.8629.00
SAZEDopt67.9757.5350.0740.7838.0338.4235.00
RobustPer89.8688.4687.2375.3674.1285.7878.00
PRPOS93.9893.1392.3288.3285.2190.3585.12
Note: For each group, the highest accuracy is in bold and the lowest is underlined.
Table A13. MAPE of PRPOS and ablation variants.
Table A13. MAPE of PRPOS and ablation variants.
ModelCos-LowCos-Medium-LowCos-Medium-HighCos-High
w/o HA0.10220.10110.09760.0961
w/o PR0.0950.09390.09070.0885
w/o IP0.10670.10930.11430.1151
w/o CP0.130.1320.1340.13
I/CP-GRU→CNN0.10360.10240.09890.0974
I/CP-GRU→Attention0.09210.09280.09240.0946
PRPOS0.0880.0870.0840.08
Note: For each group, the lowest MAPE is in bold.
Table A14. RMSE of PRPOS and ablation variants.
Table A14. RMSE of PRPOS and ablation variants.
ModelCos-LowCos-Medium-LowCos-Medium-HighCos-High
w/o HA1.06751.03420.94521.0598
w/o PR0.99570.97870.90850.9915
w/o IP1.22461.18641.08431.2227
w/o CP1.331.391.291.33
I/CP-GRU→CNN1.02370.99170.90641.0056
I/CP-GRU→Attention0.97350.96430.89320.9661
PRPOS0.960.930.850.92
Note: For each group, the lowest RMSE is in bold.

References

  1. Verma, A.; Pedrosa, L.; Korupolu, M.; Oppenheimer, D.; Tune, E.; Wilkes, J. Large-scale cluster management at Google with Borg. In Proceedings of the Tenth European Conference on Computer Systems, Bordeaux, France, 21–24 April 2015; pp. 1–17. [Google Scholar] [CrossRef]
  2. Buyya, R.; Garg, S.K.; Calheiros, R.N. SLA-oriented resource provisioning for cloud computing: Challenges, architecture, and solutions. In Proceedings of the 2011 International Conference on Cloud and Service Computing, Hong Kong, China, 12–14 December 2011; pp. 1–10. [Google Scholar] [CrossRef]
  3. Chen, Z.Y.; Sun, M.; Han, X.X. Prediction-driven collaborative emergency medical resource allocation with deep learning and optimization. J. Oper. Res. Soc. 2023, 74, 590–603. [Google Scholar] [CrossRef]
  4. Alibaba Inc. Cluster Data Collected from Production Clusters in Alibaba for Cluster Management Research. Available online: https://github.com/alibaba/clusterdata (accessed on 1 July 2025).
  5. Calheiros, R.N.; Masoumi, E.; Ranjan, R. Workload prediction using ARIMA model and its impact on cloud applications’ QoS. IEEE Trans. Cloud Comput. 2015, 3, 449–458. [Google Scholar] [CrossRef]
  6. Bi, J.; Zhang, L.; Yuan, H. Hybrid task prediction based on wavelet decomposition and ARIMA model in cloud data center. In Proceedings of the 2018 15th International Conference on Networking, Sensing and Control (ICNSC), Chicago, IL, USA, 27–29 March 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–6. [Google Scholar] [CrossRef]
  7. Moreno-Vozmediano, R.; Montero, R.S.; Huedo, E. Efficient resource provisioning for elastic cloud services based on machine learning techniques. J. Cloud Comput. 2019, 8, 5. [Google Scholar] [CrossRef]
  8. Xiao, C.; Huang, J.; Wu, W. Enhanced-XGB: An online service resource demand forecasting method for colocation data centers. In Proceedings of the 2021 21st International Symposium on Cluster, Cloud and Internet Computing (CCGrid), Melbourne, Australia, 10–13 May 2021; IEEE/ACM: New York, NY, USA, 2021; pp. 436–444. [Google Scholar] [CrossRef]
  9. Zhang, W.; Li, B.; Zhao, D. Workload prediction for cloud cluster using a recurrent neural network. In Proceedings of the 2016 International Conference on Identification, Information and Knowledge in the Internet of Things (IIKI), Beijing, China, 20–22 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 104–109. [Google Scholar] [CrossRef]
  10. Song, B.; Yu, Y.; Zhou, Y. Host load prediction with long short-term memory in cloud computing. J. Supercomput. 2018, 74, 6554–6568. [Google Scholar] [CrossRef]
  11. Bi, J.; Ma, H.; Yuan, H.; Zhang, J. Accurate prediction of workloads and resources with multi-head attention and hybrid LSTM for cloud data centers. IEEE Trans. Sustain. Comput. 2023, 8, 375–384. [Google Scholar] [CrossRef]
  12. Bi, J.; Li, S.; Yuan, H. Integrated deep learning method for workload and resource prediction in cloud systems. Neurocomputing 2020, 424, 35–48. [Google Scholar] [CrossRef]
  13. Surbab, C.; Wi, C.; Jlba, B. Adaptive sliding windows for improved estimation of data center resource utilization. Future Gener. Comput. Syst. 2020, 104, 212–224. [Google Scholar] [CrossRef]
  14. Shu, W.; Zeng, F.; Ling, Z. Resource demand prediction of cloud workloads using an attention-based GRU model. In Proceedings of the 2021 17th International Conference on Mobility, Sensing and Networking (MSN), Exeter, UK, 13–15 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 428–437. [Google Scholar] [CrossRef]
  15. Vlachos, M.; Meek, C.; Vagena, Z. Identifying similarities, periodicities and bursts for online search queries. In Proceedings of the 2004 SIGMOD International Conference on Management of Data (SIGMOD ’04), Paris, France, 13–18 June 2004; ACM: New York, NY, USA, 2004; pp. 131–142. [Google Scholar] [CrossRef]
  16. Bauer, A.; Züfle, M.; Herbst, N. Telescope: An Automatic Feature Extraction and Transformation Approach for Time Series Forecasting on a Level-Playing Field. In Proceedings of the 2020 36th International Conference on Data Engineering (ICDE), Dallas, TX, USA, 20–24 April 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1902–1905. [Google Scholar] [CrossRef]
  17. Drutsa, A.; Gusev, G.; Serdyukov, P. Periodicity in user engagement with a search engine and its application to online controlled experiments. ACM Trans. Web 2017, 11, 1–35. [Google Scholar] [CrossRef]
  18. Radinsky, K.; Svore, K.; Dumais, S. Modeling and predicting behavioral dynamics on the web. In Proceedings of the 21st International Conference on World Wide Web (WWW), Lyon, France, 16–20 April 2012; ACM: New York, NY, USA, 2012; pp. 599–608. [Google Scholar] [CrossRef]
  19. Wang, J.; Chen, T.; Huang, B. Cyclo-period estimation for discrete-time cyclo-stationary signals. IEEE Trans. Signal Process. 2006, 54, 83–94. [Google Scholar] [CrossRef]
  20. Wen, Q.; He, K.; Sun, L. RobustPeriod: Robust Time-Frequency Mining for Multiple Periodicity Detection. In Proceedings of the 2021 International Conference on Management of Data (SIGMOD), Virtual Event, 20–25 June 2021; ACM: New York, NY, USA, 2021; pp. 2328–2337. [Google Scholar] [CrossRef]
  21. Wen, Q.; Yang, L.; Sun, L. Robust Dominant Periodicity Detection for Time Series with Missing Data. In Proceedings of the 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–9 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–5. [Google Scholar] [CrossRef]
  22. Garg, S.; Ahuja, R.; Singh, R.; Perl, I. An effective deep learning architecture leveraging BIRCH clustering for resource usage prediction of heterogeneous machines in cloud data center. Cluster Comput. 2024, 27, 5699–5719. [Google Scholar] [CrossRef]
  23. Yuan, H.; Bi, J.; Li, S.; Zhang, J.; Zhou, M. An improved LSTM-based prediction approach for resources and workload in large-scale data centers. IEEE Internet Things J. 2024, 11, 22816–22829. [Google Scholar] [CrossRef]
  24. Tripathi, S.; Misra, R.; Singh, T.N. Multilayer multivariate forecasting network for precise resource utilization prediction in edge data centers. Future Gener. Comput. Syst. 2025, 166, 107692. [Google Scholar] [CrossRef]
  25. Hodrick, R.; Prescott, E.C. Postwar US business cycles: An empirical investigation. J. Money Credit Bank. 1997, 29, 1–16. [Google Scholar] [CrossRef]
  26. Zovko, I.I. Topics in Market Microstructure; Amsterdam University Press: Amsterdam, The Netherlands, 2008; ISBN 9789056295387. [Google Scholar]
  27. Chen, J.; Wang, Y. A hybrid method for short-term host utilization prediction in cloud computing. J. Electr. Comput. Eng. 2019, 2019, 2782349. [Google Scholar] [CrossRef]
  28. Sharifian, S.; Barati, M. An ensemble multiscale wavelet-GARCH hybrid SVR algorithm for mobile cloud computing workload prediction. Int. J. Mach. Learn. Cybern. 2019, 10, 3285–3300. [Google Scholar] [CrossRef]
  29. Noble, W.S. What is a support vector machine? Nat. Biotechnol. 2006, 24, 1565–1567. [Google Scholar] [CrossRef]
  30. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  31. Yang, H.J.; Hu, X. Wavelet neural network with improved genetic algorithm for traffic flow time series prediction. Optik 2016, 127, 8103–8110. [Google Scholar] [CrossRef]
  32. Baig, R.; Iqdal, W.; Berral, J. Adaptive prediction models for data center resources utilization estimation. IEEE Trans. Netw. Serv. Manag. 2019, 16, 1681–1693. [Google Scholar] [CrossRef]
  33. Gao, J.; Wang, H.; Shen, H. Machine learning based workload prediction in cloud computing. In Proceedings of the 2020 29th International Conference on Computer Communications and Networks (ICCCN), Honolulu, HI, USA, 3–6 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–9. [Google Scholar] [CrossRef]
  34. Wen, Y.; Wang, Y.; Liu, J. CPU usage prediction for cloud resource provisioning based on deep belief network and particle swarm optimization. Concurr. Comput. 2020, 32, e5370. [Google Scholar] [CrossRef]
  35. Chen, L.; Wang, W.; Yang, Y. An efficient dynamic neural network for predicting time series data stream. In Proceedings of the 2020 International Conference on Parallel & Distributed Processing with Applications, Big Data & Cloud Computing, Sustainable Computing & Communications, Social Computing & Networking (ISPA/BDCloud/SocialCom/SustainCom), Exeter, UK, 17–19 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 625–632. [Google Scholar] [CrossRef]
  36. Nguyen, H.M.; Kalra, G.; Kim, D. Host load prediction in cloud computing using long short-term memory encoder-decoder. J. Supercomput. 2019, 75, 7592–7605. [Google Scholar] [CrossRef]
  37. Thonglek, K.; Ichikawa, K.; Takahashi, K. Improving resource utilization in data centers using an LSTM-based prediction model. In Proceedings of the 2019 21st International Conference on Cluster Computing (CLUSTER), Albuquerque, NM, USA, 2–5 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–8. [Google Scholar] [CrossRef]
  38. Shen, H.; Hong, X. Host load prediction with bi-directional long short-term memory in cloud computing. arXiv 2020, arXiv:2001.05142. [Google Scholar] [CrossRef]
  39. Ouhame, S.; Hadi, Y.; Ullah, A. An efficient forecasting approach for resource utilization in cloud data center using CNN-LSTM model. Neural Comput. Appl. 2021, 33, 10043–10055. [Google Scholar] [CrossRef]
  40. Chen, W.; Lu, C.; Ye, K. RPTCN: Resource prediction for high-dynamic workloads in clouds based on deep learning. In Proceedings of the 2021 23rd International Conference on Cluster Computing (CLUSTER), Portland, OR, USA, 7–10 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 59–69. [Google Scholar] [CrossRef]
  41. Dogani, J.; Khunjush, F. Multivariate workload and resource prediction in cloud computing using CNN and GRU by attention mechanism. J. Supercomput. 2023, 79, 3437–3470. [Google Scholar] [CrossRef]
  42. Ullah, F.; Bilal, M.; Yoon, S.-K. Intelligent time-series forecasting framework for non-linear dynamic workload and resource prediction in cloud. Comput. Netw. 2023, 225, 109653. [Google Scholar] [CrossRef]
  43. Kashyap, S.; Singh, A.; Gill, S.S. Machine learning-centric prediction and decision based resource management in cloud computing environments. Cluster Comput 2025, 28, 130. [Google Scholar] [CrossRef]
  44. Bi, J.; Ma, H.; Yuan, H.; Buyya, R.; Yang, J.; Zhang, J.; Zhou, M. Multivariate resource usage prediction with frequency-enhanced and attention-assisted transformer in cloud computing systems. IEEE Internet Things J. 2024, 11, 26419–26429. [Google Scholar] [CrossRef]
  45. Yang, Z.; Hou, L.; Zhao, X. Robust Autocorrelation for Period Detection in Time Series. In Proceedings of the 17th International Conference on Agents and Artificial Intelligence (ICAART), Porto, Portugal, 23–25 February 2025; Volume 2, pp. 36–44. [Google Scholar]
  46. Behrendt, M.; de Angelis, M.; Comerford, L.; Zhang, Y.; Beer, M. Projecting interval uncertainty through the discrete Fourier transform: An application to time signals with poor precision. Mech. Syst. Signal Process. 2022, 172, 108920. [Google Scholar] [CrossRef]
  47. Vlachos, M.; Yu, P.; Castelli, V. On periodicity detection and structural periodic similarity. In Proceedings of the 2005 SIAM International Conference on Data Mining (SDM), Newport Beach, CA, USA, 21–23 April 2005; SIAM: Philadelphia, PA, USA, 2005; pp. 449–460. [Google Scholar] [CrossRef]
  48. Mitsa, T. Temporal Data Mining; Chapman and Hall/CRC: Boca Raton, FL, USA, 2010; pp. 154–196. [Google Scholar]
  49. Toller, M.; Santos, T.; Kern, R. SAZED: Parameter-free domain-agnostic season length estimation in time series data. Data Min. Knowl. Discov. 2019, 33, 1775–1798. [Google Scholar] [CrossRef]
  50. Parthasarathy, S.; Mehta, S.; Srinivasan, S. Robust periodicity detection algorithms. In Proceedings of the 2006 15th International Conference on Information and Knowledge Management (CIKM), Arlington, VA, USA, 6–11 November 2006; ACM: New York, NY, USA, 2006; pp. 874–875. [Google Scholar] [CrossRef]
  51. Li, Z.; Wang, J.; Han, J. Mining event periodicity from incomplete observations. IEEE Trans. Knowl. Data Eng. 2015, 27, 1219–1232. [Google Scholar] [CrossRef]
  52. Wu, Q.; Yao, G.; Feng, Z.; Shuyuan, Y. Peri-midformer: Periodic pyramid transformer for time series analysis. Neural Inf. Process. Syst. 2024, 2024 37, 13035–13073. [Google Scholar] [CrossRef]
  53. Luan, J.; Wang, H.; Wang, K.; Zhang, B. Robust distributed estimation and variable selection for massive datasets via rank regression. Ann. Inst. Stat. Math. 2022, 74, 435–450. [Google Scholar] [CrossRef]
  54. Wang, N.; Chau, S.C.K. Practically Efficient Secure Computation of Rank-based Statistics Over Distributed Datasets. arXiv 2023, arXiv:2302.08121. [Google Scholar] [CrossRef]
  55. Tan, P.N.; Steinbach, M.; Kumar, V. Introduction to Data Mining; Pearson Education India: New Delhi, India, 2005; pp. 76–78. [Google Scholar]
  56. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar] [CrossRef]
  57. Kim, H. Hodrick-Prescott Filter. 12 March 2004. Available online: http://webhome.auburn.edu/~hzk0001/hpfilter.pdf (accessed on 12 June 2025).
  58. Box, G.E.P.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis, Forecasting and Control; Holden-Day, Inc.: San Francisco, CA, USA, 1990. [Google Scholar]
  59. Ross, S.M. A First Course in Probability; Pearson: Harlow, UK, 2020. [Google Scholar]
  60. Hyvarinen, A.; Hurri, J.; Hoyer, P.O. Independent Component Analysis. Natural Image Statistics: A Probabilistic Approach to Early Computational Vision; Springer: London, UK, 2001; pp. 151–175. [Google Scholar]
  61. Agarap, A.F. Deep learning using rectified linear units (relu). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  62. Kavianpour, P.; Kavianpour, M.; Jahani, E.; Ramezani, A. A CNN-BiLSTM model with attention mechanism for earthquake prediction. J. Supercomput. 2023, 79, 19194–19226. [Google Scholar] [CrossRef]
  63. Tran, N.; Nguyen, T.; Nguyen, B.M. A multivariate fuzzy time series resource forecast model for clouds using LSTM and data correlation analysis. In Proceedings of the 2018 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Rio de Janeiro, Brazil, 8–13 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 636–645. [Google Scholar] [CrossRef]
  64. Dang-Quang, N.M.; Yoo, M. Multivariate deep learning model for workload prediction in cloud computing. In Proceedings of the 2021 International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 20–22 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 858–862. [Google Scholar] [CrossRef]
  65. Xu, M.; Song, C.; Wu, H. esDNN: Deep neural network based multivariate workload prediction in cloud computing environments. ACM Trans. Internet Technol. 2022, 22, 1–24. [Google Scholar] [CrossRef]
  66. Hyndman, R.J.; Athanasopoulos, G.; Bergmeir, C.; Caceres, G.; Chhay, L.; O’Hara-Wild, M.; Petropoulos, F.; Razbash, S. Package ‘Forecast’; R Foundation for Statistical Computing: Vienna, Austria, 2019. [Google Scholar]
  67. De Myttenaere, A.; Golden, B.; Le Grand, B.; Rossi, F. Mean absolute percentage error for regression models. Neurocomputing 2016, 192, 38–48. [Google Scholar] [CrossRef]
  68. Chai, T.; Draxler, R.R. Root mean square error (RMSE) or mean absolute error (MAE)?–Arguments against avoiding RMSE in the literature. Geosci. Model Dev. 2014, 7, 1247–1250. [Google Scholar] [CrossRef]
Figure 1. Framework of PRPOS.
Figure 1. Framework of PRPOS.
Applsci 15 10967 g001
Figure 2. Framework of period detection.
Figure 2. Framework of period detection.
Applsci 15 10967 g002
Figure 3. Example of Historical Data Selection in PRPOS.
Figure 3. Example of Historical Data Selection in PRPOS.
Applsci 15 10967 g003
Figure 4. Resource prediction model in PRPOS.
Figure 4. Resource prediction model in PRPOS.
Applsci 15 10967 g004
Figure 5. GRU Architecture.
Figure 5. GRU Architecture.
Applsci 15 10967 g005
Figure 6. Overall performance comparison of PRPOS and baseline methods.
Figure 6. Overall performance comparison of PRPOS and baseline methods.
Applsci 15 10967 g006
Figure 7. Performance improvement of PRPOS over baseline models.
Figure 7. Performance improvement of PRPOS over baseline models.
Applsci 15 10967 g007
Figure 8. Comparison of MAPE under different input data size.
Figure 8. Comparison of MAPE under different input data size.
Applsci 15 10967 g008
Figure 9. Performance variations of PRPOS under different sampled window size per cycle.
Figure 9. Performance variations of PRPOS under different sampled window size per cycle.
Applsci 15 10967 g009
Figure 10. Performance variations of PRPOS under different number of historical cycles.
Figure 10. Performance variations of PRPOS under different number of historical cycles.
Applsci 15 10967 g010
Figure 11. Comparison of period detection accuracy of PRPOS and baseline methods.
Figure 11. Comparison of period detection accuracy of PRPOS and baseline methods.
Applsci 15 10967 g011
Figure 12. Comparison of the prediction accuracy of PRPOS and ablation variants.
Figure 12. Comparison of the prediction accuracy of PRPOS and ablation variants.
Applsci 15 10967 g012
Table 1. Parameter settings in PRPOS.
Table 1. Parameter settings in PRPOS.
 NameValue
Period DetectionSmoothing parameter in HP filtering ( λ )14,400
Rank-based ACF window length (L)3000
Rank-based ACF lags (G)4000
Cosine similarity threshold ( τ _ c )0.8
Step size of period refinement ( Δ t )30
Terminal condition of period refinement ( s n )5
Sample Data
Selection
Input data size (N)60
Sampled window size per cycle (W)10
Number of historical cycles (K)5
Resource Prediction
Modeling
Full connection layer depth ( L f c )2
Full connection layer neuron number ( N f c ) N f c 1 = 128, N f c 2 = 64
GRU hidden unit number ( d h )64
in-period-GRU hidden layer depth ( L i p )1
cross-period-GRU hidden layer depth ( L c p )2
Learning rate ( α )0.005
Table 2. Comparison of computational efficiency for period detection methods.
Table 2. Comparison of computational efficiency for period detection methods.
MethodsPeriod Size
[800–1000)[1000–1200)[1200–1400)[1400–1600)[1600–1820]
PRPOS0.0440.0820.1350.1540.206
RobustPeriod69.485201.89409.806563.839780.353
SAZEDopt0.0870.1540.2240.3210.358
SAZEDmaj0.0430.0720.1090.1570.181
findFrequency0.000460.001410.001730.002500.00297
Note: For each group, the lowest computational cost is in bold and the highest is underlined.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liang, Y.; Zhou, H.; Li, T.; Shen, H. PRPOS: A Periodicity-Aware Resource Prediction Framework for Online Services. Appl. Sci. 2025, 15, 10967. https://doi.org/10.3390/app152010967

AMA Style

Liang Y, Zhou H, Li T, Shen H. PRPOS: A Periodicity-Aware Resource Prediction Framework for Online Services. Applied Sciences. 2025; 15(20):10967. https://doi.org/10.3390/app152010967

Chicago/Turabian Style

Liang, Yi, Hongwen Zhou, Tianxu Li, and Haotian Shen. 2025. "PRPOS: A Periodicity-Aware Resource Prediction Framework for Online Services" Applied Sciences 15, no. 20: 10967. https://doi.org/10.3390/app152010967

APA Style

Liang, Y., Zhou, H., Li, T., & Shen, H. (2025). PRPOS: A Periodicity-Aware Resource Prediction Framework for Online Services. Applied Sciences, 15(20), 10967. https://doi.org/10.3390/app152010967

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop