Next Article in Journal
A Design of a Leaf-Shaped Biomimetic Flexible Wideband Antenna
Previous Article in Journal
Research on the Vibration Sensor Based on Microfiber Loop Resonator
Previous Article in Special Issue
Color Normalization Through a Simulated Color Checker Using Generative Adversarial Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MSP-EDA: Multivariate Time Series Forecasting Based on Multiscale Patches and External Data Augmentation

1
State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
2
School of Data Science and Engineering, East China Normal University, Shanghai 200062, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(13), 2618; https://doi.org/10.3390/electronics14132618
Submission received: 23 May 2025 / Revised: 21 June 2025 / Accepted: 27 June 2025 / Published: 28 June 2025
(This article belongs to the Special Issue Machine Learning in Data Analytics and Prediction)

Abstract

Accurate multivariate time series forecasting remains a major challenge in various real-world applications, primarily due to the limitations of existing models in capturing multiscale temporal dependencies and effectively integrating external data. To address these issues, we propose MSP-EDA, a novel multivariate time series forecasting framework that integrates multiscale patching and external data enhancement. Specifically, MSP-EDA utilizes the Discrete Fourier Transform (DFT) to extract dominant global periodic patterns and employs an adaptive Continuous Wavelet Transform (CWT) to capture scale-sensitive local variations. In addition, multiscale patches are constructed to capture temporal patterns at different resolutions, and a specialized encoder is designed for each scale. Each encoder incorporates temporal attention, channel correlation attention, and cross-attention with external data to capture intra-scale temporal dependencies, inter-variable correlations, and external influences, respectively. To fuse information from different temporal scales, we introduce a trainable global token that serves as a variable-wise aggregator across scales. Extensive experiments on four public benchmark datasets and three real-world vector database datasets that we collect demonstrate that MSP-EDA consistently outperforms state-of-the-art methods, achieving particularly notable improvements on vector database workloads. Ablation studies further confirm the effectiveness of each module and the adaptability of MSP-EDA to complex forecasting scenarios involving external dependencies.

1. Introduction

Multivariate time series (MTS) forecasting aims to predict future values of one or more target variables by leveraging multiple correlated and temporally evolving features. It has been widely applied in domains such as economics [1], traffic [2,3], healthcare [4,5], and workload forecasting [6,7,8]. Among these, workload forecasting often involves highly dynamic and non-stationary patterns, exhibiting multiscale fluctuations driven by diverse query types, variable data volumes, and burst access behaviors. For example, the workloads of vector databases typically reflect this complexity, making accurate forecasting essential for adaptive resource scheduling, performance tuning, and system scalability under volatile conditions [8,9,10].
Existing MTS forecasting methods can be broadly categorized into statistical learning and deep learning approaches. Traditional methods, such as Vector AutoRegression (VAR) [11] and extended ARIMA models, are effective for stationary, low-noise datasets but struggle with complex patterns. In recent years, deep learning models have dominated due to their ability to model nonlinear dynamics and handle noisy real-world data. Early works based on RNNs and CNNs, such as SegRNN [12] and SCINet [13], have demonstrated promising performance. More recently, Transformer-based models [14,15,16] have gained attention for their superior ability to capture long-term dependencies, although further design improvements are needed to fully exploit their potential in MTS forecasting.
Accurate time series forecasting in real-world scenarios demands the ability to capture complex and diverse temporal patterns. However, achieving this remains a significant challenge due to several key issues.
First, real-world time series often exhibit multi-modal behaviors, characterized by a mixture of long-term trends, seasonal fluctuations, and abrupt short-term changes. Effectively modeling these intertwined patterns is critical but non-trivial. Most existing approaches either overlook or only extract simple temporal patterns [17,18], which makes it difficult to capture the diverse dynamics inherent in time series data [19,20]. Related studies [21,22,23] leverage Fourier Transform to extract frequency-domain representations, capturing global periodic patterns effectively. However, these methods are generally limited by their reliance on fixed basis functions or fixed receptive fields, which makes it difficult to adapt to localized, non-stationary temporal patterns. In particular, they tend to ignore the fine-grained, transient frequency changes that are common in real-world time series, especially under conditions of noise, abrupt events, or irregular cycles. Moreover, their frequency modeling tends to be either global or statically defined, making it hard to dynamically adjust to scale-varying or time-varying features.
Second, different temporal resolutions often capture distinct and complementary information. For example, the utilization of CPU and memory resources in cloud computing reveals unique temporal patterns spanning daily, monthly, and seasonal scales [6]. Moreover, as illustrated in Figure 1, multivariate time series data exhibit strong temporal dependencies and complex inter-variable relationships [24]. To better represent these characteristics, the figure depicts sequences at three segmentation scales (shown by background colors), each containing two correlated variables. This design emphasizes the need to capture both multiscale temporal patterns and cross-variable dynamics, which are essential for accurate modeling. Effectively capturing both sequential dynamics and cross-feature interactions is critical for enhancing both accuracy and generalization ability of forecasting models [25,26]. However, many existing models treat time series at a single temporal scale, thereby failing to capture the inherent multiscale nature of real-world data [27,28,29]. Additionally, most attention-based forecasting models [30] either focus solely on temporal correlations or apply shallow inter-variable attention, lacking an integrated mechanism for capturing interactions across both time and variable dimensions at multiple scales. Furthermore, they do not provide a unified way to dynamically fuse information across resolutions, which is essential for reconciling long-term trends with short-term fluctuations.
Third, integrating external data can significantly enhance forecasting performance [15]. Approaches such as ARIMAX [31] capture the relationships between exogenous and endogenous variables, along with autoregression on the endogenous variables. TiDE [32] further suggests that forecasting models might have access to the future values of exogenous variables during the prediction of endogenous variables. However, most existing methods assume well-aligned and complete data streams, which is often unrealistic in practical scenarios [15]. In real-world applications, time series data are frequently affected by missing values, time delays, and uneven sampling, making it challenging to model the dynamic interactions between exogenous and endogenous variables accurately.
In summary, existing methods face several key limitations. Many rely on fixed or global frequency bases, which limits their ability to capture localized and time-varying patterns in non-stationary time series. Others fail to effectively model the interactions between multiscale temporal structures and inter-variable dependencies, which are crucial for understanding complex dynamics in real-world data. Additionally, most approaches assume complete and well-aligned external data, making them less robust to missing values, time delays, and irregular sampling commonly encountered in practical scenarios.
In this article, we propose MSP-EDA, a novel forecasting model that integrates multiscale patch and external data augmentation to address the aforementioned challenges. First, we combine Discrete Fourier Transform (DFT) with an adaptive multiscale wavelet transform to simultaneously extract global periodic trends and localized time–frequency variations. Unlike existing models that rely on fixed-frequency bases, our approach enables flexible and fine-grained temporal-frequency decomposition, which is essential for modeling non-stationary time series. Second, we introduce a multiscale patching strategy that segments time series into variable-length patches at different resolutions. Each patch is processed by a dedicated encoder equipped with a temporal attention module for sequential dependencies and a channel attention module for inter-variable interactions. This design enables the model to jointly capture temporal dynamics and variable relationships across multiple scales, which are often modeled separately in prior works. Third, a dedicated embedding module fuses timestamp-related external information with internal representations, and a cross-attention mechanism selectively integrates external variables based on contextual relevance. This dynamic integration mechanism offers greater robustness compared to models that assume clean and fully aligned external data. Finally, a global token aggregates multi-resolution outputs from all encoder layers, allowing the model to adaptively fuse information across temporal scales and enhance forecasting accuracy.
Experiments on several datasets show that the proposed method consistently outperforms existing approaches, achieving superior or near-optimal performance. In particular, on three vector database datasets, MSP-EDA significantly outperforms the best baseline methods in terms of MSE. Compared to the second-best TimeXer model [15], MSP-EDA improves MSE by 1.4% on the Dvdb dataset, 2.1% on the Mvdb, and 1.9% on the Wvdb, respectively, highlighting its effectiveness in forecasting for real-world applications.
Our main contributions can be summarized as follows:
  • We design an adaptive multiscale wavelet-based representation module that automatically adjusts scale decomposition based on the characteristics of input data, enabling the model to extract rich and relevant time–frequency features.
  • We introduce a cross-attention that dynamically fuses external data, enhancing the ability to adapt to complex and changing input conditions.
  • We propose a novel MSP-EDA framework, which unifies multiscale patch representation and external data integration to capture complex temporal patterns and intra-variable dependencies.
  • We perform extensive experiments on seven benchmark datasets to demonstrate that MSP-EDA consistently outperforms state-of-the-art baselines, confirming its effectiveness and generalizability across diverse forecasting scenarios.
The remainder of this paper is organized as follows. We first review the related literature in Section 2. Then, we elaborate on MSP-EDA and its modules in Section 3. Finally, we show the experimental results in Section 4 and conclude the whole paper in Section 5.

2. Related Work

2.1. Multivariate Time Series Forecasting

Existing methods for MTS forecasting can be categorized into statistical learning and neural network-based approaches [24]. Statistical methods, such as ARIMA [33] and VAR [11], capture linear dependencies and are well-suited for stationary data. In contrast, deep learning models like TimesNet [34], DeepAR [35] treat time series as sequences of vectors and utilize CNNs or RNNs to capture temporal dependencies. Furthermore, Transformer-based architectures, including Informer [16], FEDformer [18], MedFormer [5], iTransformer [14], and Crossformer [25] are more effective at modeling nonlinear patterns and handling complex, noisy data. MLP-based methods, including MSD-Mixer [26], TiDE [32], TimeMixer [36], and LSTF-Linear [37] employ a simple architecture with relatively few parameters and have also demonstrated good forecasting accuracy. However, many of these models rely on fixed-length time segments, which limits their ability to capture the full complexity of time series data, as they fail to account for patterns occurring at different temporal scales, such as long-term trends and short-term fluctuations.

2.2. Multiscale Modeling for MTS Forecasting

Multiscale modeling has shown remarkable success in enhancing correlation learning and feature extraction in domains such as computer vision [38] and multi-modal learning [39]. However, its application to time series forecasting remains relatively underexplored. Some recent efforts have attempted to incorporate multiscale representations into time series models [40,41]. MTST [42] employs a multi-branch architecture to model various temporal patterns at different resolutions, addressing the challenge of capturing both short-term high-frequency patterns and long-term seasonal trends. Pathformer [27] introduces a multiscale Transformer with adaptive pathways that integrates temporal resolution and distance to capture diverse temporal patterns. However, while these methods achieve multiscale partitioning, they generally assume independence across channels, potentially limiting their ability to capture inter-variable relationships and fully exploit the dependencies between different data features.

2.3. MTS Forecasting with External Data

Incorporating external data into time series forecasting can significantly improve predictive performance by providing additional insight beyond the intrinsic temporal patterns of the target variables [15]. Classical models like ARIMAX [33] and SARIMAX [43] explicitly model both autoregressive dependencies and the relationships between internal and external variables. In deep learning, models such as the Temporal Fusion Transformer (TFT) [44] introduce a unified attention-based mechanism to learn variable importance dynamically over time, handling both historical and known future inputs. Similarly, NBEATSx [45] and TiDE [32] incorporate future external inputs along with historical inputs from the target variables. TimeXer [15] focuses on cross-time dependencies via inter-time attention, allowing it to capture long-term dependencies between distant time steps. However, these methods often rely on strict temporal alignment, overlook localized or multiscale temporal variations, and lack explicit mechanisms for fusing external signals across different time resolutions.
Table 1 presents a comparison of various forecasting methods based on three critical components: decomposition, multiscale modeling, and external data integration. Each of these components plays a pivotal role in capturing the complex patterns and dynamics within time series data. Decomposition enhances the ability to represent the data by extracting frequency-domain and more detailed time–frequency features, which improve overall data representation. Multiscale modeling captures temporal dependencies at multiple resolutions while considering the inter-channel relationships, allowing for more effective capture of complex data features. The integration of external data strengthens predictive performance by incorporating additional contextual information, enriching the ability to address dynamic and diverse time series. MSP-EDA combines all three components, offering a comprehensive approach that more effectively captures the rich and varied information within time series data, thereby outperforming existing methods.

3. Methodology

Time series forecasting aims to predict future values of a sequence based on its historical observations. Formally, given a multivariate time series input X = [ x 1 , x 2 , , x T ] R D × T , where D is the number of variables and T is the length of the historical window, the objective is to forecast the future values Y = [ y T + 1 , , y T + H ] R D × H over a prediction horizon H.
The key challenge in this task lies in effectively modeling complex temporal dynamics, including long-term periodic patterns, short-term variations, and potential external influences. To this end, we propose MSP-EDA, a novel model that integrates multiscale patch representations with external data enhancement to improve forecasting accuracy and robustness. In the remainder of this section, we first present an overview of the model architecture in Section 3.1. Then, we detail each module individually: data pattern enhancement in Section 3.2, multiscale patch processing in Section 3.3, and multiscale fusion in Section 3.4.

3.1. Overview

Our approach begins by enhancing the input time series through frequency- and time–frequency-based analysis to capture both global periodic trends and local variations. The enhanced data are then decomposed into patches at multiple temporal scales, where each scale is processed by dedicated encoders that incorporate external information. Finally, a global token is used to fuse multiscale representations for accurate forecasting. As illustrated in Figure 2, our model consists of three main components:
  • Data Pattern Enhancement: Extracts both global periodic trends and localized time–frequency features using signal processing techniques, specifically leveraging the Discrete Fourier Transform (DFT) to identify dominant global frequency components and the Continuous Wavelet Transform (CWT) to capture localized, non-stationary patterns across time and scale.
  • Multiscale Patch Processing: Decomposes the time series into multiple temporal resolutions to capture multiscale dependencies while simultaneously modeling temporal dynamics and inter-variable correlations. In addition, tailored embedding strategies are designed to effectively integrate external data into each scale for enhanced contextual understanding.
  • Multiscale Fusion: Combines representations from different temporal scales by introducing a global variable token, which is fused with the outputs from each encoder layer to generate the final prediction. This process effectively integrates multiscale information and external data, ensuring a comprehensive representation for accurate forecasting.

3.2. Data Pattern Enhancement

In time series forecasting, enhancing raw data representations is crucial for capturing complex temporal characteristics, such as global periodic trends and local variations. To this end, we design a dedicated enhancement module to extract both global and local time–frequency patterns. Specifically, we employ the Discrete Fourier Transform (DFT) to extract long-term periodical patterns, while the Continuous Wavelet Transform (CWT) is utilized to capture fine-grained, localized temporal dynamics.
To capture both global and local temporal characteristics, we design an enhancement module that integrates Discrete Fourier Transform (DFT) and Continuous Wavelet Transform (CWT). DFT provides a compact representation of long-term periodic trends by decomposing signals into global frequency components, while CWT enables localized time–frequency analysis through scaled and shifted wavelets. This combination allows the model to extract multiscale features more effectively than fixed-kernel convolutions or learnable filters, which often struggle with non-stationary or transient patterns common in real-world time series.
Periodic feature extraction involves transforming input data from the time domain to the frequency domain to capture frequency-based patterns. We employ DFT, denoted as DFT ( · ) , which decomposes a signal into a set of sinusoidal components with different frequencies, thereby revealing dominant periodic trends while suppressing noise. The extracted frequency-domain features are then transformed back to the time domain using the inverse transform IDFT ( · ) for further processing. The process is represented as follows:
X freq = IDFT TopK amp DFT ( x ) ,
where TopK amp ( · ) represents the frequency-domain filtering operation that retains the top-k frequency components with the highest frequency.
Local time–frequency features are extracted using the CWT, which projects the input signal onto a family of scaled and shifted wavelet functions to capture localized patterns across different temporal resolutions. To enhance adaptability, we parameterize the multiscale settings of the wavelet transform as trainable variables. In particular, the time-scaling parameters of the wavelet basis functions are learned during training, allowing the model to adaptively select optimal time–frequency decomposition scales based on the characteristics of the input data. The k-th adaptive wavelet kernel function ψ k ( t ) is represented as follows:
ψ k ( t ) = 1 Z k · C θ ψ t σ k ,
where t denotes the time variable and the normalization factor Z k ensures a consistent energy across wavelets computed as the L 2 norm. The operator C θ [ · ] represents linear interpolation, which resizes the wavelet to match the input sequence length. The scale parameter is defined as σ k = exp ( β k ) · exp ( γ ) , where β k controls the scale of the k-th wavelet and γ is a global scaling factor. Both β k and γ are learnable and optimized during training.
Unlike fixed-scale wavelet transforms, where scales are predefined and typically spaced logarithmically, the adaptive method learns the scale parameters β k and γ during training. Initially, both parameters are set to zero, allowing the network to learn the appropriate scales from the data. Throughout the training process, these parameters are adjusted based on the specific characteristics of the input sequence, such as frequency components and signal variations. The parameter β k allows each wavelet kernel to adapt its scale to focus on different frequency components, enabling the network to capture features on multiple scales. The global scaling factor γ adjusts the overall scale across all wavelets, optimizing the ability to capture broad or fine-grained patterns in the data. This dynamic adaptation enables the network to respond more effectively to non-stationary signals, improving the extraction of time–frequency features.

3.3. Multiscale Patch Processing

To effectively model complex temporal patterns and cross-variable interactions in multivariate time series, we propose a multiscale patch processing module consisting of a multiscale embedding and a unified encoder. The embedding divides the input sequences into patches of varying lengths to capture temporal dependencies at multiple resolutions, enabling the model to jointly learn short-term variations and long-term trends. The encoder applies temporal attention, channel attention, and cross-attention with timestamp-derived features to extract sequential, inter-variable, and external dependencies. This design allows the model to disentangle heterogeneous patterns across time and variables, improving the adaptability to real-world, multiscale forecasting scenarios.
The multiscale patching strategy divides the input time series into multiple sets of non-overlapping subsequences with varying lengths, enabling the model to capture patterns across short- and long-term temporal scales. Formally, we define a set of M temporal scales S = { S 1 , S 2 , , S M } , where each S i defines a specific patch size. Given an input sequence X R D × L , each scale S i splits X into P = L / S i patches ( X 1 , X 2 , , X P ) , where X S i j R D × S i . For each patch, a temporal embedding is applied along the time dimension, producing X t j R D × d m , where d m denotes the embedding dimension. This multi-resolution embedding enables robust feature extraction across different temporal granularities.
After multiscale embedding, the resulting representations from each scale are passed to a corresponding encoder. Each encoder includes two main components: (1) a dual-attention block that combines temporal attention, capturing intra-series dependencies over time, and channel attention, modeling inter-variable correlations; and (2) a cross-attention module that incorporates timestamp-derived features, such as hour of day, day of week, or other periodic temporal patterns, to enhance the temporal context of learned representations. By integrating these three attention mechanisms, the encoder is able to learn comprehensive and time-aware representations, which are crucial for accurate and robust time series forecasting under diverse temporal conditions. The detailed processing flow of this module is illustrated in Figure 3.
To model temporal dependencies across different time resolutions, we apply attention between patches rather than within each patch. Each patch, constructed at a specific temporal scale, is treated as a compressed temporal unit that retains key local dynamics. This design reduces redundancy and focuses attention on higher-level temporal transitions across patches. Given a set of sub-sequences { X 1 , X 2 , , X P } partitioned with patch size S i , we denote the embedded input as X t R C × P × d m , where P is the number of patches and d m is the embedding dimension. A multi-head self-attention mechanism is then used to model inter-patch dependencies, enabling the model to capture both short- and long-range temporal patterns across multiple scales. The attention is computed by first initializing the query and key as follows:
Q inter i = W Q i X t i , K inter i = W K i X t i , 1 i H ,
where W Q i , W K i are the learnable projection matrices for the query and key, respectively. The attention output for each head is computed as follows:
Attn inter i = Softmax Q inter i K inter i d m V inter i ,
where Softmax ( · ) normalizes the attention weights. After applying inter-patch attention, each patch is updated based on its dependencies with other patches, and the resulting attention outputs of all heads are concatenated to form the final representation, denoted as Attn inter R P × d m , i.e.,
Attn inter = Concat Attn inter 1 , , Attn inter H W O ,
where W O R H × d h × d m is the output projection, and H is the number of attention heads. This temporal attention enables the model to effectively learn long-range temporal patterns across the entire sequence by focusing on relationships between compressed patch-level representations at multiple time scales.
To effectively model inter-variable dependencies in multivariate time series, we introduce a channel attention module that captures the importance of each feature dimension. Instead of treating all input channels equally, this mechanism adaptively reweighs feature representations by learning channel-wise interactions, allowing the model to focus on the most informative variables while suppressing irrelevant or redundant ones. This design enhances the model’s capacity to extract meaningful cross-feature relationships, which are essential for accurate and robust forecasting [24,30]. Given the input representation X R C × P × d m , where C is the number of channels, P represents the number of patches, and d m is the embedding dimension, we first permute the tensor to reshape it to X R d m × P × C . This reorganization allows us to focus on the feature dimension, making it easier to apply attention mechanisms along the channel axis. A linear projection is then applied along the channel dimension C to compute the attention weights, which are used to refine the feature representations. The detailed processing steps are as follows:
Q c = W q X , K c = W k X , V c = W v X ,
where Q c , K c , and V c denote the query, key, and value, respectively, and W q , W k , and W v represent the learnable parameter matrices of Q c , K c , and V c . Then, we compute the attention weights of each channel as follows:
Attn c ( Q c , K c , V c ) = Softmax Q c K c T d m V c ,
Finally, the attention-weighted output is combined with the input through a residual operation to produce the final output.
External time-related features provide essential prior knowledge that helps forecast models disambiguate similar input patterns that occur at different times. For example, behaviors at 8 a.m. and 8 p.m. may follow different dynamics, even if other features are similar. To incorporate this inductive bias, we extract temporal features such as weekday, date, hour, and minute from raw timestamps during preprocessing. These features are embedded alongside the original input sequence, allowing the model to learn position-aware and periodic patterns that are difficult to infer from the original signal alone. This design enhances the model’s temporal awareness, particularly under scenarios with recurring seasonal trends or irregular sampling intervals. The processing steps are as follows:
Ex _ Embedding = ValueEmbedding Concat ( X , X time ) ,
where ValueEmbedding ( · ) denotes the embedding operation function, Concat ( · ) denotes the concat operation, X R C × L represents the origin input, and X time R T × L represents the timestamp data. Instead of applying patch division directly on these external features, which could significantly increase computational overhead and introduce redundant noise, we adopt a more efficient strategy. Each external feature sequence is directly embedded as a corresponding variable token. These embedded tokens are integrated into the model alongside the primary input channels, enabling the attention mechanisms to jointly model both intrinsic dynamics and exogenous temporal patterns. This approach allows the model to better align periodic variations with the internal temporal structure of the data, improving its capacity to make temporal-aware predictions.
After the above processing, the embedded data is represented as Ex _ Embedding R ( C + T ) × D , and we denote it as E R ( C + T ) × D for simplicity. By embedding external data as variable-level tokens, cross-attention is employed to effectively aggregate temporal semantics from the external information. Specifically, the input time series is used as the query (Q), while the embedded external variables serve as the keys (K) and the values (V), establishing correlations between the two modalities. This operation is formulated as follows:
Q = X s P R C × D K = V = E R ( C + T ) × D ,
where X s P denotes the representation of the last temporal patch P, and S refers to the patch size. The reason for using only the last patch rather than all patches is that the last patch, being the most recent, contains the highest information density and has the greatest impact on the final prediction. Unlike TimeXer [15], which introduces a learnable token to bridge endogenous temporal patches and exogenous variate-level representations via cross-attention, our approach maintains temporal locality by directly utilizing the most recent patch as the query. This design preserves the interpretability of the attention mechanism and aligns better with the causal structure inherent in time series forecasting.
The proposed approach effectively combines multiple attention mechanisms to enhance the ability to capture complex temporal dependencies and inter-feature relationships. By employing temporal attention, the model captures sequential dependencies across patches, allowing for the modeling of long-range temporal dynamics while mitigating noise from intra-patch processing. Channel attention further improves the capability to capture inter-feature interactions by focusing on channel-level dependencies, enhancing feature representation across multiple channels. Furthermore, cross-attention with external data integrates contextual temporal information, such as weekday, hour, and minute, to further enrich the understanding of periodic patterns and temporal semantics. By leveraging these multiscale attention strategies, the model gains a comprehensive and nuanced understanding of the time series data, improving both forecasting accuracy and robustness.

3.4. Multiscale Fusion

To integrate information across multiple temporal resolutions, we introduce a trainable global token G R 1 × C × 1 × d m that performs variable-wise, cross-scale aggregation. Unlike simple concatenation or pooling, this learnable token serves as a dynamic query to adaptively summarize multiscale representations based on the relevance to each variable. It captures global contextual information across different patch scales, enabling the model to weigh contributions from both short-term and long-term patterns.
The token is initialized from a standard normal distribution N ( 0 , 1 ) and expanded along the batch dimension during training, yielding G R B × C × 1 × d m . Given the encoder outputs F i R C × P i × d m from each scale i, we concatenate F i and G along the temporal axis. This design maintains positional relationships between scale-specific features and the global token, allowing the model to explicitly learn cross-scale attention patterns.
Compared to conventional fusion strategies such as averaging or hierarchical stacking, our global token-based approach provides a flexible and learnable mechanism to model fine-grained interactions across scales and variables, improving adaptability to varying temporal dynamics. Specifically, we have
H = Concat ( F 1 , , F N , G ) R C × ( i = 1 N P i + 1 ) × d m ,
where F i represents the output of the i-th patch length. After that, we employ the multi-head attention to compute the attention weights of different scales. The final result of all attention heads is spliced and then transformed back to the d m dimension by a linear operation, formulated as follows:
O = W A Concat ( A 1 , A 2 , , A h ) ,
where W A ( · ) denotes the linear operation, and A i denotes the result of the attention calculated for the i-th head.
For the final output representation O R C × ( P i + 1 ) × d m , a prediction head is applied to map it to the desired forecast window. To do this, we first flatten the last two dimensions to obtain a unified vector representation, enabling the model to jointly consider both patch-level and temporal information. This flattened representation is then passed through a multi-layer perceptron (MLP) for nonlinear transformation, followed by a dropout layer to prevent overfitting and enhance generalization. The process is defined as follows:
Y ^ = Dropout ( MLP ( Flatten ( O ) ) ) , Y ^ R C × pred _ len .
The prediction module bridges the high-dimensional multiscale representation and the final forecasting objective, effectively distilling rich temporal and cross-feature information into accurate future predictions.

4. Experiments

4.1. Experimental Setup

We conduct extensive experiments to evaluate the performance of our approach on long-term forecasting, including seven real-world benchmark datasets, four of which are well-known time series datasets and the remaining three of which are real-world workload datasets we collect from vector databases. Furthermore, we have conducted detailed comparisons with seven advanced baselines. The evaluation metrics used are Mean Squared Error (MSE) and Mean Absolute Error (MAE), defined as follows:
MSE = 1 n i = 1 n ( y i y ^ i ) 2 ,
MAE = 1 n i = 1 n | y i y ^ i | .
We conduct comprehensive experiments on seven datasets, covering both widely used public benchmarks and real-world vector database workloads. Detailed statistics for these datasets are provided in Table 2. The public ETTh1, ETTh2, and ETTm2 [16] datasets from power-transformer temperature monitoring and the Weather [17] dataset feature relatively long periodicity and smooth temporal variations, making them appropriate for evaluating forecasting performance. In contrast, our three proprietary vector-database workload datasets (Dvdb, Mvdb, and Wvdb) exhibit pronounced dynamic behavior with frequent abrupt shifts and spike patterns, reflecting realistic, high-variability workload scenarios. This diversity ensures that MSP-EDA is rigorously tested not only on gently varying, well-studied series but also on challenging, non-stationary workloads, thereby validating its effectiveness across both stable and highly dynamic forecasting environments.
We compared with seven state-of-the-art time series forecasting models, including four Transformer-based models (TimeXer [15], PatchTST [29], and FEDformer [18]), one CNN-based model (TimesNet [34]), two MLP-based models (TimeMixer [37] and TiDE [32]). We also include a decomposition-then-prediction method HRTCP [47], which first applies signal decomposition before performing forecasting.

4.2. Overall Performance

As shown in Table 3, we provide a comprehensive comparison of eight forecasting models across multiple prediction horizons on seven datasets. For each method, the results are obtained by running the model three times and reporting the average performance to ensure robustness and fairness. The best results are highlighted in red, while the second-best results are underlined in blue. From the table, it is evident that our proposed method consistently outperforms other approaches under various settings, demonstrating superior accuracy and generalizability.
On the three real-world vector database datasets (Dvdb, Mvdb, and Wvdb), our model significantly outperforms strong baselines like TimeXer, achieving relative MSE improvements of 1.4%, 2.1%, and 1.9%, respectively. This validates its ability to capture complex temporal patterns that involve high-dimensional metrics and dynamic workloads. By combining multiscale temporal attention, channel-specific features, and external signal integration, our model handles noisy and highly dynamic scenarios effectively.
Moreover, MSP-EDA also performs competitively on public benchmarks, often outperforming or closely matching the best baselines. However, on ETTh2, the gain is limited due to weak inter-variable correlations. Models like PatchTST, assuming channel-wise independence, perform better in such cases by avoiding spurious dependencies. In contrast, our model’s explicit modeling of cross-variable interactions may introduce noise, slightly affecting accuracy.
In addition, we include a comparison with HRTCP, a decomposition-then-prediction method that explicitly separates frequency components before modeling. While this approach can effectively isolate signals of different frequencies, its ability to capture inter-variable dependencies in multivariate settings is limited. As a result, its overall performance lags behind our proposed model on most datasets. Nonetheless, with targeted enhancements, such decomposition-based approaches may hold greater potential in future multivariate forecasting scenarios.
Despite the overall strong performance, MSP-EDA exhibits certain limitations on specific datasets. On both ETTh1 and Weather datasets, the model achieves very competitive results in terms of MSE, obtaining the best performance on ETTh1 and either the best or second-best on Weather. However, in both cases, the MAE scores are relatively less competitive. This indicates that while the model effectively reduces large errors, it is less sensitive to smaller deviations or local fluctuations. Such a pattern suggests a trade-off in the current design, where optimizing for MSE may reduce precision in finer-grained accuracy.
As shown in Figure 4, the proposed model exhibits varying training times across different datasets, primarily due to differences in parameter configurations. The maximum training time is approximately two hours, while the longest testing time remains within one minute, which is considered acceptable for practical applications. Furthermore, since the test sets in real-world deployments are typically smaller, the actual testing time is expected to be even shorter. These results demonstrate the model’s feasibility and efficiency for deployment in real-time or resource-constrained environments.

4.3. Ablation Study

We conducted a comprehensive ablation study on three datasets: Dvdb, Mvdb, and Wvdb, to evaluate the individual contributions of key components in our model. The baseline model (denoted as origin) incorporates all modules, while six ablation variants are constructed by removing one component at a time: DFT-based transformation (dft), continuous wavelet transform (cwt), cross-attention (cross), temporal attention (time), channel attention (channel), and global token fusion (global). The results, reported in terms of mean squared error (MSE) with a prediction length of 96, are illustrated in Figure 5.
The findings show that disabling any single module leads to an increase in MSE across all datasets, confirming the effectiveness of each component. Among the variants, the origin model consistently achieves the best performance. Notably, the absence of the cwt and global modules results in the most substantial degradation, particularly on the Dvdb dataset. This highlights the importance of multiscale temporal feature extraction and the integration of global contextual information. The cross and time attention modules also contribute significantly, as their removal causes moderate increases in MSE, indicating their role in capturing temporal and inter-series dependencies. The impact of the channel attention module appears to be more dataset-dependent: while its removal has little effect on Dvdb and Mvdb, it leads to a noticeable degradation on Wvdb, suggesting that feature refinement across variables plays a more prominent role in that scenario.
Overall, the ablation study underscores the complementary nature of all the proposed modules. In particular, the incorporation of multiscale representations, global context, and attention mechanisms is critical to enhancing predictive accuracy and ensuring robust generalization across diverse multivariate time series datasets.
In addition, we further conduct a targeted study to evaluate the impact of wavelet design. This ablation study systematically evaluates the performance differences between fixed-scale and adaptive wavelet transforms in time series modeling. As illustrated in Figure 6, the adaptive wavelet transform, by dynamically selecting decomposition scales based on input data, demonstrates a clear advantage in capturing multiscale temporal features. This adaptivity leads to consistently better performance in terms of accuracy and robustness across all three workload datasets. Notably, the improvement is most significant on the Dvdb dataset, which exhibits pronounced non-stationarity and high-frequency fluctuations, allowing the adaptive mechanism to fully leverage its ability to capture variable-scale patterns. In comparison, the Mvdb and Wvdb datasets show relatively stable load patterns with limited fluctuations, which naturally constrains the potential performance gain. However, even in these more stationary scenarios, the adaptive approach still achieves marginal but consistent improvements over the fixed-scale variant. These results highlight the effectiveness and generalizability of adaptive wavelet mechanisms in enhancing model performance, especially under dynamic and heterogeneous workload conditions.

4.4. Hyperparameter Analysis

We evaluate the effect of two hyperparameters: multiscale patch combinations and d _ m o d e l of attention on the Dvdb, Mvdb, and Wvdb datasets. The prediction length is 96. The results are shown in Figure 7. Experimental results demonstrate that multiscale input combinations significantly enhance prediction performance compared to single-scale input. On the Dvdb and Wvdb datasets, finer-grained scales yield better results. For example, in single-scale settings, increasing the patch length from 4 to 8 and then to 16 allows the model to capture longer-term dependencies but leads to a reduced ability to model local patterns, resulting in increased prediction errors. Moreover, incorporating excessively long scales in multi-scale combinations may even degrade performance. This is because Dvdb and Wvdb exhibit high fluctuations under heavy workloads, making long-range features more susceptible to noise and less effective. In contrast, the Mvdb dataset demonstrates more stable behavior, maintaining strong performance even with larger-scale inputs.
This experiment investigates the impact of the model scale on forecasting performance by varying the hidden dimension size (d_model) within the range of 64 to 1024. The results are shown in Figure 8. The results demonstrate that increasing d_model generally enhances the learning capacity, leading to lower prediction errors across most datasets. However, this improvement comes with a trade-off: as d_model grows, the model incurs significantly higher computational and memory costs, raising concerns about scalability and efficiency in practical deployments. Specifically, while performance improves consistently from 64 to 128, a further increase to 256 causes a noticeable drop in accuracy on the Mvdb dataset, likely due to feature redundancy introduced by multiscale extraction, which can cause the model to overfit noisy or irrelevant patterns. Moreover, when d_model reaches 1024, both the Dvdb and Wvdb datasets exhibit degraded performance, suggesting that excessive model capacity can hinder generalization by capturing spurious fluctuations rather than meaningful trends. Although the Mvdb dataset still benefits from a larger model size, the marginal performance gains diminish as d_model increases, highlighting the need to balance model expressiveness with computational efficiency.

4.5. Visualization

To qualitatively evaluate the forecasting capabilities of different models, we visualize the final dimension of the prediction results on a representative subset of the Wvdb dataset, as shown in Figure 9. Compared with other methods, MSP-EDA demonstrates a clearer ability to follow the overall trend. Although minor fluctuations may not be accurately captured, MSP-EDA successfully identifies the critical downward trajectory, which is essential for accurate decision-making in time series forecasting [48]. In addition to the qualitative comparison, Figure 10 presents a quantitative evaluation of all models using MSE and MAE. As depicted, MSP-EDA consistently achieves the lowest error across both metrics, clearly validating its superior performance over existing baselines.

5. Conclusions

To address the challenges of capturing multiscale temporal dependencies and effectively integrating external data in time series forecasting, we propose a novel forecasting framework based on multiscale patch decomposition and external data augmentation. The model first applies the discrete Fourier transform (DFT) and adaptive multiscale continuous wavelet transform (CWT) to extract dominant periodic components and scale-sensitive local features from the input sequence. A multiscale encoder is then employed to capture temporal patterns at multiple resolutions, with each encoder comprising temporal attention, channel attention, and cross-attention mechanisms to model intra-scale dynamics, inter-variable dependencies, and the influence of external variables. To enhance global context understanding, a learnable global token is incorporated and fused with the multiscale representations. Experimental results across multiple real-world benchmarks demonstrate that the proposed method consistently achieves superior performance compared to existing approaches.
Although the proposed model achieves strong forecasting performance, its relatively complex architecture can hinder deployment in real-time or resource-constrained environments. Moreover, the integration of multiple modules and attention mechanisms reduces interpretability, making it difficult to trace and explain prediction outcomes. In future work, we will explore simplifying the model through techniques such as model pruning and knowledge distillation to reduce computational complexity without sacrificing accuracy. To enhance interpretability, we plan to incorporate attention visualization, feature attribution approaches like SHAP or Integrated Gradients, and modular analysis to better understand how the model makes forecasting decisions.

Author Contributions

Conceptualization, S.P. and H.L.; methodology, S.P. and H.L.; software, S.P.; validation, S.P.; formal analysis, S.P., W.S., P.C. and H.L.; investigation, S.P.; resources, S.P. and H.L.; data curation, S.P.; writing—original draft preparation, S.P.; writing—review and editing, S.P., W.S., P.C., H.X., D.M., M.C., Y.W. and H.L.; visualization, S.P.; supervision, H.L.; project administration, S.P., W.S., P.C., H.X., D.M., M.C., Y.W. and H.L.; funding acquisition, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China (No. 61562010), National Key Research and Development Program of China (No. 2023YFC3341205), Guizhou Provincial Major Scientific and Technological Program (No. [2024]003), Guizhou Provincial Program on Commercialization of Scientific and Technological Achievements (No. [2023]010, No. [2025]008), Research Projects of the Science and Technology Plan of Guizhou Province (No. [2023]276, No. [2022]261, No. [2022]271).

Data Availability Statement

The official access links for all publicly available datasets are provided in the manuscript. At the same time, all data and codes related to this study are available at https://github.com/ACMISLab/MSP-EDA (accessed on 22 May 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sezer, O.B.; Gudelek, M.U.; Ozbayoglu, A.M. Financial time series forecasting with deep learning: A systematic literature review: 2005–2019. Appl. Soft Comput. 2020, 90, 106181. [Google Scholar] [CrossRef]
  2. Lin, Y.; Wan, H.; Guo, S.; Lin, Y. Pre-training context and time aware location embeddings from spatial-temporal trajectories for user next location prediction. Proc. AAAI Conf. Artif. Intell. 2021, 35, 4241–4248. [Google Scholar] [CrossRef]
  3. Sun, C.; Ning, Y.; Shen, D.; Nie, T. Graph Neural Network-Based Short-Term Load Forecasting with Temporal Convolution. Data Sci. Eng. 2024, 9, 113–132. [Google Scholar] [CrossRef]
  4. Miao, X.; Wu, Y.; Wang, J.; Gao, Y.; Mao, X.; Yin, J. Generative semi-supervised learning for multivariate time series imputation. Proc. AAAI Conf. Artif. Intell. 2021, 35, 8983–8991. [Google Scholar] [CrossRef]
  5. Wang, Y.; Huang, N.; Li, T.; Yan, Y.; Zhang, X. Medformer: A Multi-Granularity Patching Transformer for Medical Time-Series Classification. Adv. Neural Inf. Process. Syst. 2024, 37, 36314–36341. [Google Scholar]
  6. Pan, Z.; Wang, Y.; Zhang, Y.; Yang, S.B.; Cheng, Y.; Chen, P.; Guo, C.; Wen, Q.; Tian, X.; Dou, Y.; et al. MagicScaler: Uncertainty-Aware, Predictive Autoscaling. Proc. VLDB Endow. 2023, 16, 3808–3821. [Google Scholar] [CrossRef]
  7. Wang, S.; Chu, Z.; Sun, Y.; Liu, Y.; Guo, Y.; Chen, Y.; Jian, H.; Ma, L.; Lu, X.; Zhou, J. Multiscale Representation Enhanced Temporal Flow Fusion Model for Long-Term Workload Forecasting. In Proceedings of the 33rd ACM International Conference on Information and Knowledge Management, New York, NY, USA, 21–25 October 2024; pp. 4948–4956. [Google Scholar] [CrossRef]
  8. Zhao, F.; Lin, W.; Lin, S.; Zhong, H.; Li, K. TFEGRU: Time-Frequency Enhanced Gated Recurrent Unit with Attention for Cloud Workload Prediction. IEEE Trans. Serv. Comput. 2024, 18, 467–478. [Google Scholar] [CrossRef]
  9. Gao, Y.; Huang, X.; Zhou, X.; Gao, X.; Li, G.; Chen, G. DBAugur: An Adversarial-based Trend Forecasting System for Diversified Workloads. In Proceedings of the 39th IEEE International Conference on Data Engineering, Anaheim, CA, USA, 3–7 April 2023; pp. 27–39. [Google Scholar] [CrossRef]
  10. Guo, Y.; Ge, J.; Guo, P.; Chai, Y.; Li, T.; Shi, M.; Tu, Y.; Ouyang, J. Pass: Predictive auto-scaling system for large-scale enterprise web applications. In Proceedings of the ACM Web Conference 2024, Singapore, 13–17 May 2024; pp. 2747–2758. [Google Scholar] [CrossRef]
  11. Kilian, L.; Lütkepohl, H. Structural Vector Autoregressive Analysis; Cambridge University Press: Cambridge, UK, 2017. [Google Scholar] [CrossRef]
  12. Lin, S.; Lin, W.; Wu, W.; Zhao, F.; Mo, R.; Zhang, H. SegRNN: Segment Recurrent Neural Network for Long-Term Time Series Forecasting. arXiv 2023, arXiv:2308.11200. Available online: https://arxiv.org/abs/2308.11200 (accessed on 22 May 2025).
  13. Liu, M.; Zeng, A.; Chen, M.; Xu, Z.; Lai, Q.; Ma, L.; Xu, Q. SCINet: Time Series Modeling and Forecasting with Sample Convolution and Interaction. Adv. Neural Inf. Process. Syst. 2022, 35, 5816–5828. [Google Scholar]
  14. Liu, Y.; Hu, T.; Zhang, H.; Wu, H.; Wang, S.; Ma, L.; Long, M. iTransformer: Inverted Transformers Are Effective for Time Series Forecasting. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024; Available online: https://openreview.net/forum?id=JePfAI8fah (accessed on 22 May 2025).
  15. Wang, Y.; Wu, H.; Dong, J.; Qin, G.; Zhang, H.; Liu, Y.; Qiu, Y.; Wang, J.; Long, M. TimeXer: Empowering Transformers for Time Series Forecasting with Exogenous Variables. Adv. Neural Inf. Process. Syst. 2024, 37, 469–498. [Google Scholar]
  16. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. Proc. AAAI Conf. Artif. Intell. 2021, 35, 11106–11115. [Google Scholar] [CrossRef]
  17. Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting. Adv. Neural Inf. Process. Syst. 2021, 34, 22419–22430. [Google Scholar]
  18. Zhou, T.; Ma, Z.; Wen, Q.; Wang, X.; Sun, L.; Jin, R. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In Proceedings of the 39th International Conference on Machine Learning, 17–23 July 2022; Volume 162, pp. 27268–27286. Available online: https://proceedings.mlr.press/v162/zhou22g/zhou22g.pdf (accessed on 22 May 2025).
  19. Bandara, K.; Hyndman, R.J.; Bergmeir, C. MSTL: A seasonal-trend decomposition algorithm for time series with multiple seasonal patterns. Int. J. Oper. Res. 2025, 52, 79–98. [Google Scholar] [CrossRef]
  20. Wen, Q.; Zhang, Z.; Li, Y.; Sun, L. Fast RobustSTL: Efficient and Robust Seasonal-Trend Decomposition for Time Series with Complex Patterns. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 6–10 June 2020; pp. 2203–2213. [Google Scholar] [CrossRef]
  21. Fan, W.; Yi, K.; Ye, H.; Ning, Z.; Zhang, Q.; An, N. Deep frequency derivative learning for non-stationary time series forecasting. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, Jeju, Republic of Korea, 3–9 August 2024; pp. 3944–3952. [Google Scholar] [CrossRef]
  22. Xu, Z.; Zeng, A.; Xu, Q. FITS: Modeling Time Series with $10k$ Parameters. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024; Available online: https://openreview.net/forum?id=bWcnvZ3qMb (accessed on 22 May 2025).
  23. Ye, W.; Deng, S.; Zou, Q.; Gui, N. Frequency Adaptive Normalization For Non-stationary Time Series Forecasting. Adv. Neural Inf. Process. Syst. 2024, 37, 31350–31379. [Google Scholar]
  24. Qiu, X.; Hu, J.; Zhou, L.; Wu, X.; Du, J.; Zhang, B.; Guo, C.; Zhou, A.; Jensen, C.S.; Sheng, Z.; et al. TFB: Towards Comprehensive and Fair Benchmarking of Time Series Forecasting Methods. Proc. VLDB Endow. 2024, 17, 2363–2377. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Yan, J. Crossformer: Transformer Utilizing Cross-Dimension Dependency for Multivariate Time Series Forecasting. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023; Available online: https://openreview.net/forum?id=vSVLM2j9eie (accessed on 22 May 2025).
  26. Zhong, S.; Song, S.; Zhuo, W.; Li, G.; Liu, Y.; Chan, S.H.G. A Multi-Scale Decomposition MLP-Mixer for Time Series Analysis. Proc. VLDB Endow. 2024, 17, 1723–1736. [Google Scholar] [CrossRef]
  27. Chen, P.; ZHANG, Y.; Cheng, Y.; Shu, Y.; Wang, Y.; Wen, Q.; Yang, B.; Guo, C. Pathformer: Multi-scale Transformers with Adaptive Pathways for Time Series Forecasting. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024; Available online: https://openreview.net/forum?id=lJkOCMP2aW (accessed on 22 May 2025).
  28. Lee, S.; Park, T.; Lee, K. Learning to Embed Time Series Patches Independently. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024; Available online: https://openreview.net/forum?id=WS7GuBDFa2 (accessed on 22 May 2025).
  29. Nie, Y.; Nguyen, N.H.; Sinthong, P.; Kalagnanam, J. A Time Series is Worth 64 Words: Long-term Forecasting with Transformers. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023; Available online: https://openreview.net/forum?id=Jbdc0vTOcol (accessed on 22 May 2025).
  30. Cai, W.; Liang, Y.; Liu, X.; Feng, J.; Wu, Y. MSGNet: Learning Multi-Scale Inter-Series Correlations for Multivariate Time Series Forecasting. Proc. AAAI Conf. Artif. Intell. 2024, 38, 11141–11149. [Google Scholar] [CrossRef]
  31. Williams, B.M. Multivariate vehicular traffic flow prediction: Evaluation of ARIMAX modeling. Transp. Res. Rec. 2001, 1776, 194–200. [Google Scholar] [CrossRef]
  32. Das, A.; Kong, W.; Leach, A.; Mathur, S.K.; Sen, R.; Yu, R. Long-term Forecasting with TiDE: Time-series Dense Encoder. Trans. Mach. Learn. Res. 2023. Available online: https://openreview.net/forum?id=pCbC3aQB5W (accessed on 22 May 2025).
  33. Shumway, R.H.; Stoffer, D.S.; Shumway, R.H.; Stoffer, D.S. ARIMA models. Time Series Analysis and Its Applications: With R Examples; Springer: Cham, Switzerland, 2017; pp. 75–163. [Google Scholar] [CrossRef]
  34. Wu, H.; Hu, T.; Liu, Y.; Zhou, H.; Wang, J.; Long, M. TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023; Available online: https://openreview.net/forum?id=ju_Uqw384Oq (accessed on 22 May 2025).
  35. Salinas, D.; Flunkert, V.; Gasthaus, J.; Januschowski, T. DeepAR: Probabilistic forecasting with autoregressive recurrent networks. Int. J. Forecast. 2020, 36, 1181–1191. [Google Scholar] [CrossRef]
  36. Wang, S.; Wu, H.; Shi, X.; Hu, T.; Luo, H.; Ma, L.; Zhang, J.Y.; ZHOU, J. TimeMixer: Decomposable Multiscale Mixing for Time Series Forecasting. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024; Available online: https://openreview.net/forum?id=7oLshfEIC2 (accessed on 22 May 2025).
  37. Zeng, A.; Chen, M.; Zhang, L.; Xu, Q. Are transformers effective for time series forecasting? Proc. AAAI Conf. Artif. Intell. 2023, 37, 11121–11128. [Google Scholar] [CrossRef]
  38. Li, Y.; Wu, C.Y.; Fan, H.; Mangalam, K.; Xiong, B.; Malik, J.; Feichtenhofer, C. MViTv2: Improved Multiscale Vision Transformers for Classification and Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4794–4804. [Google Scholar] [CrossRef]
  39. Wang, J.; Wu, Z.; Ouyang, W.; Han, X.; Chen, J.; Jiang, Y.G.; Li, S.N. M2TR: Multi-modal Multi-scale Transformers for Deepfake Detection. In Proceedings of the 2022 International Conference on Multimedia Retrieval, New York, NY, USA, 27–30 June 2022; pp. 615–623. [Google Scholar] [CrossRef]
  40. Challu, C.; Olivares, K.G.; Oreshkin, B.N.; Ramirez, F.G.; Canseco, M.M.; Dubrawski, A. NHITS: Neural Hierarchical Interpolation for Time Series Forecasting. Proc. AAAI Conf. Artif. Intell. 2023, 37, 6989–6997. [Google Scholar] [CrossRef]
  41. Shabani, M.A.; Abdi, A.H.; Meng, L.; Sylvain, T. Scaleformer: Iterative Multi-scale Refining Transformers for Time Series Forecasting. In Proceedings of the The Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023; Available online: https://openreview.net/forum?id=sCrnllCtjoE (accessed on 22 May 2025).
  42. Zhang, Y.; Ma, L.; Pal, S.; Zhang, Y.; Coates, M. Multi-resolution Time-Series Transformer for Long-term Forecasting. In Proceedings of the 27th International Conference on Artificial Intelligence and Statistics, Valencia, Spain, 2–4 May 2024; Volume 238, pp. 4222–4230. Available online: https://proceedings.mlr.press/v238/zhang24l/zhang24l.pdf (accessed on 22 May 2025).
  43. Arunraj, N.S.; Ahrens, D.; Fernandes, M. Application of SARIMAX model to forecast daily sales in food retail industry. Int. J. Oper. Res. Inf. Syst. 2016, 7, 1–21. [Google Scholar] [CrossRef]
  44. Lim, B.; Arık, S.O.; Loeff, N.; Pfister, T. Temporal Fusion Transformers for interpretable multi-horizon time series forecasting. Int. J. Forecast. 2021, 37, 1748–1764. [Google Scholar] [CrossRef]
  45. Olivares, K.G.; Challu, C.; Marcjasz, G.; Weron, R.; Dubrawski, A. Neural basis expansion analysis with exogenous variables: Forecasting electricity prices with NBEATSx. Int. J. Forecast. 2023, 39, 884–900. [Google Scholar] [CrossRef]
  46. Wang, H.; Peng, J.; Huang, F.; Wang, J.; Chen, J.; Xiao, Y. MICN: Multi-scale Local and Global Context Modeling for Long-term Series Forecasting. In Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023; Available online: https://openreview.net/forum?id=zt53IDUR1U (accessed on 22 May 2025).
  47. Su, Y.; Tan, M.; Teh, J. Short-Term Transmission Capacity Prediction of Hybrid Renewable Energy Systems Considering Dynamic Line Rating Based on Data-Driven Model. IEEE Trans. Ind. Appl. 2025, 61, 2410–2420. [Google Scholar] [CrossRef]
  48. Kim, J.; Kim, H.; Kim, H.; Lee, D.; Yoon, S. A Comprehensive Survey of Time Series Forecasting: Architectural Diversity and Open Challenges. arXiv 2024, arXiv:2411.05793. Available online: https://arxiv.org/abs/2411.05793 (accessed on 22 May 2025).
Figure 1. Example of multiscale time series data processing. Multivariate time series at three different segmentation scales are represented by distinct background colors. Each region shows two variables; red ellipses denote strong inter-variable correlations, and black arrows indicate temporal dependencies across time points.
Figure 1. Example of multiscale time series data processing. Multivariate time series at three different segmentation scales are represented by distinct background colors. Each region shows two variables; red ellipses denote strong inter-variable correlations, and black arrows indicate temporal dependencies across time points.
Electronics 14 02618 g001
Figure 2. The architecture of MSP-EDA, which consists of three components: data pattern enhancement (left), multiscale patch processing (including multiscale embedding and multi-layer encoders), and multiscale feature fusion.
Figure 2. The architecture of MSP-EDA, which consists of three components: data pattern enhancement (left), multiscale patch processing (including multiscale embedding and multi-layer encoders), and multiscale feature fusion.
Electronics 14 02618 g002
Figure 3. The structure of the encoder module, which consists of three attention mechanisms: temporal attention for modeling intra-series temporal dependencies, channel attention for capturing inter-variable correlations, and cross attention for integrating external data.
Figure 3. The structure of the encoder module, which consists of three attention mechanisms: temporal attention for modeling intra-series temporal dependencies, channel attention for capturing inter-variable correlations, and cross attention for integrating external data.
Electronics 14 02618 g003
Figure 4. Training and testing time cost statistics for each dataset.
Figure 4. Training and testing time cost statistics for each dataset.
Electronics 14 02618 g004
Figure 5. Ablation studies on the Dvdb, Mvdb, and Wvdb datasets to evaluate the individual contributions of different modules in MSP-EDA.
Figure 5. Ablation studies on the Dvdb, Mvdb, and Wvdb datasets to evaluate the individual contributions of different modules in MSP-EDA.
Electronics 14 02618 g005
Figure 6. Comparison of fixed-scale and adaptive-scale wavelet transforms in forecasting performance (MSE) across various forecasting lengths. Here, [ 4 , 8 , 16 ] , [ 4 , 8 ] , and [ 8 , 16 ] indicate the different fixed scales used for the wavelet transforms.
Figure 6. Comparison of fixed-scale and adaptive-scale wavelet transforms in forecasting performance (MSE) across various forecasting lengths. Here, [ 4 , 8 , 16 ] , [ 4 , 8 ] , and [ 8 , 16 ] indicate the different fixed scales used for the wavelet transforms.
Electronics 14 02618 g006
Figure 7. Evaluation of different patch length combinations in terms of on Dvdb, Mvdb, and Wvdb datasets.
Figure 7. Evaluation of different patch length combinations in terms of on Dvdb, Mvdb, and Wvdb datasets.
Electronics 14 02618 g007
Figure 8. Impact of attention dimension d_model on forecasting performance in terms of MSE on Mvdb, Wvdb, and Dvdb.
Figure 8. Impact of attention dimension d_model on forecasting performance in terms of MSE on Mvdb, Wvdb, and Dvdb.
Electronics 14 02618 g008
Figure 9. Prediction cases from Wvdb dataset by different models in the input-96-predict-96 setting.
Figure 9. Prediction cases from Wvdb dataset by different models in the input-96-predict-96 setting.
Electronics 14 02618 g009
Figure 10. MSE and MAE results for the instances in Figure 9.
Figure 10. MSE and MAE results for the instances in Figure 9.
Electronics 14 02618 g010
Table 1. Comparison of time series forecasting methods in the incorporation of key components.
Table 1. Comparison of time series forecasting methods in the incorporation of key components.
MethodDecompositionMultiscaleExternal Data
FEDformer [18]××
MICN [46]×
TiDE [32]×
MSGNet [30]×
MSD-Mixer [26]×
Pathformer [27]×
TimeXer [15]××
MSP-EDA (This work)
Table 2. Statistics of time series datasets in the experiments.
Table 2. Statistics of time series datasets in the experiments.
DatasetDimensionTimestampsGranularity
ETTh1717,4201 h
ETTh2717,4201 h
ETTm2769,68015 min
Weather2152,69610 min
Dvdb2228,2781 min
Mvdb1821,5701 min
Wvdb1518,4901 min
Table 3. Long-term forecasting results on seven benchmark datasets at different prediction lengths (96/192/336/720). The bold red and blue underlined fonts indicate the best and second-best results in each setting.
Table 3. Long-term forecasting results on seven benchmark datasets at different prediction lengths (96/192/336/720). The bold red and blue underlined fonts indicate the best and second-best results in each setting.
ModelMSP-EDATimeXerTimeMixerHRTCPTimesNetPatchTSTTiDEFEDformer
MetricMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAEMSEMAE
ETTh1960.37710.40340.38730.40570.38650.40110.38830.40320.39310.41450.39160.41030.38880.39680.38010.4196
1920.42770.43320.43990.42780.44250.43230.52130.48730.54830.49580.43710.42980.44010.42600.46940.4693
3360.46550.45650.47980.44830.47680.45560.55330.50130.55640.50620.47840.45240.48340.44860.52310.4908
7200.47920.48030.48760.47470.53340.49200.59100.53250.72570.60460.47700.47530.48610.47350.53650.5153
ETTh2960.29100.34030.29520.34250.30010.35190.29540.34380.32930.36950.29400.35090.29120.34040.34700.3894
1920.37450.40130.37030.39300.37680.39640.39370.40080.46570.45160.36850.39370.37650.39190.41540.4251
3360.41710.43260.42280.43400.44380.44500.44450.44670.50460.48520.41580.42980.42210.43010.46240.4697
7200.43820.45040.44740.45530.44820.45840.46130.46990.49250.48420.42720.44830.42320.44170.48190.4866
ETTm2960.17300.25970.17590.25790.17540.26010.18010.26130.18950.26860.18330.26710.18220.26500.20060.2849
1920.23830.30160.23640.29990.24480.30510.24800.30800.25470.31100.25110.31070.24870.30570.26390.3221
3360.29760.33950.29520.33850.30230.34750.31550.36020.31500.34920.31260.34870.30750.34300.32780.3628
7200.39980.39990.39650.39800.39510.40130.42180.41790.42110.40800.41240.40380.40770.39840.41690.4132
Weather960.16090.20890.15760.20490.16180.20880.17320.20860.16940.21870.17090.21220.19300.23430.21080.2900
1920.21050.25350.21150.25310.20700.25130.22110.25320.23900.27790.22870.26170.24000.27000.26400.3169
3360.26450.29220.26680.29310.29110.30950.29680.31560.28970.31100.28320.30030.29190.30620.31440.3470
7200.34360.34380.34360.34330.34300.34390.37810.37240.35900.35390.35920.34970.36520.35400.38150.3828
Dvdb960.97510.71300.98890.71370.98170.71540.98920.71941.07890.75460.99240.72141.0470.74551.07260.7686
1921.00070.72731.01910.73891.00800.72811.02730.74051.08080.75351.03590.74621.16580.79631.09640.7752
3361.02140.73651.03820.74121.02820.73391.05210.74531.09680.75981.05870.75661.18930.80581.11540.7833
7201.03600.74541.04900.74931.03620.74631.09130.77071.10340.76231.07680.76501.19830.81041.12190.7855
Mvdb960.21730.29680.22190.29780.23420.30600.23120.29970.24740.31010.23570.30080.25640.31310.24480.3143
1920.23910.31140.24030.31460.27810.33490.26380.31990.25470.32270.26760.33130.30720.35120.27360.3366
3360.27290.34650.27740.34520.32860.36940.32560.36880.28540.34900.31450.36520.35400.38370.32270.3724
7200.34610.39560.37520.40660.44790.43510.45030.43370.37670.41070.42710.42990.46590.44460.43640.4380
Wvdb960.90070.59130.91780.59780.91920.59890.92130.60070.93910.60410.90950.59740.95930.61410.94190.6128
1920.91110.60550.91960.60890.95090.62450.93240.60960.93470.61610.92390.61571.09700.66570.93860.6225
3360.92390.62740.93300.62880.98330.65300.96170.64580.95270.63630.94450.63921.12180.69020.95620.6450
7201.02330.69081.10540.70311.07230.70971.06170.71131.12700.71891.12490.70861.19950.74611.03710.7021
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, S.; Sun, W.; Chen, P.; Xu, H.; Ma, D.; Chen, M.; Wang, Y.; Li, H. MSP-EDA: Multivariate Time Series Forecasting Based on Multiscale Patches and External Data Augmentation. Electronics 2025, 14, 2618. https://doi.org/10.3390/electronics14132618

AMA Style

Peng S, Sun W, Chen P, Xu H, Ma D, Chen M, Wang Y, Li H. MSP-EDA: Multivariate Time Series Forecasting Based on Multiscale Patches and External Data Augmentation. Electronics. 2025; 14(13):2618. https://doi.org/10.3390/electronics14132618

Chicago/Turabian Style

Peng, Shunhua, Wu Sun, Panfeng Chen, Huarong Xu, Dan Ma, Mei Chen, Yanhao Wang, and Hui Li. 2025. "MSP-EDA: Multivariate Time Series Forecasting Based on Multiscale Patches and External Data Augmentation" Electronics 14, no. 13: 2618. https://doi.org/10.3390/electronics14132618

APA Style

Peng, S., Sun, W., Chen, P., Xu, H., Ma, D., Chen, M., Wang, Y., & Li, H. (2025). MSP-EDA: Multivariate Time Series Forecasting Based on Multiscale Patches and External Data Augmentation. Electronics, 14(13), 2618. https://doi.org/10.3390/electronics14132618

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop