Next Article in Journal
Multi-Objective MILP Models for Optimizing Makespan and Energy Consumption in Additive Manufacturing Systems
Previous Article in Journal
Influence of Marble Dust on Mechanical and Tribological Properties of Injection Molded Polypropylene Composites
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

A Parallel Processing Architecture for Long-Term Power Load Forecasting †

1
LMII—Faculty of Sciences and Technology, Hassan 1st University, P.O. Box 577, Settat 26000, Morocco
2
CELOG, Ecole Supérieure des Industries du Textile et de l’Habillement, Casablanca 20250, Morocco
*
Author to whom correspondence should be addressed.
Presented at the 1st International Conference on Smart Management in Industrial and Logistics Engineering (SMILE 2025), 16–19 April 2025, Casablanca, Morocco.
Eng. Proc. 2025, 97(1), 26; https://doi.org/10.3390/engproc2025097026
Published: 16 June 2025

Abstract

:
The increasing complexity of power grids and integration of renewable energy sources necessitate accurate power load forecasting across multiple time horizons. While existing methods have advanced significantly, they often struggle with consistent performance across different prediction ranges, leading to suboptimal resource allocation. We propose MP-RWKV (Multi-Path Recurrent Weighted Key–Value), an enhanced architecture that builds upon RWKV-TS and addresses these challenges through parallel processing paths for temporal modeling. Our model maintains robust performance across both short-term and long-term forecasting scenarios through its context state mechanism and position-aware attention. Evaluated on extensive power load data, MP-RWKV demonstrates superior performance over state-of-the-art baselines, including Transformer-based models and LSTM variants. The model achieves the lowest Mean Absolute Error (MAE) across prediction horizons ranging from 24 h to 432 h, showing particular strength in maintaining consistent accuracy where traditional models deteriorate. Notably, MP-RWKV successfully balances immediate temporal correlations with extended dependencies, offering promising implications for power grid management and sustainable energy systems. The model’s stable performance across varying prediction horizons makes it particularly suitable for real-world power load forecasting applications.

1. Introduction

The world energy landscape is transforming due to a gradual increase in electricity consumption over the years, which is forecast to reach historic levels in the next few decades [1]. The increased demand for renewable energy sources and smart grid technology has led to a greater need for reliable long-term power load forecasts. Accurate forecasting is essential for the efficient planning of resources, the development of infrastructure, and the maintenance of the stability and reliability of power systems [2].
Long-term load forecasting (LTLF) poses particular challenges compared to short-term forecasting. It requires models that can handle short-term fluctuations as well as long-term trends induced by economic growth, technological advancements and climate change [3]. Although traditional forecasting methods are good for their simplicity and interpretability, they are ineffective in handling the complexity and non-linearity associated with long-term power load patterns [4].
This indicates that significant developments have been made in power load forecasting methodologies. The initial approach was largely statistical, with the majority using autoregressive integrated moving average (ARIMA) models [5]. While these models are effective for stationary time series, they are highly limited when it comes to handling the non-stationarity and non-linearity of long-term power load data [6].
This completely changed the dynamics of forecasting. Improvements in AI techniques and machine learning have enhanced the architecture of neural networks, particularly RNNs and LSTM networks, which have a greater ability to capture time dependencies [7]. However, these models have been shown to degrade over larger time horizons due to issues such as vanishing gradients [8].
In recent years, attention-based models have emerged, inspired by advances in natural language processing. These models, as exemplified by the Transformer model, appear to have potential for modeling long-range dependencies [9]. However, they are sometimes insensitive to local patterns that are important in power load dynamics [10].
Despite these advances, a significant gap remains in successfully blending the strengths of various modeling techniques for LTLF. No matter how advanced they are, standalone models tend not to accommodate the entire range of factors that contribute to long-term power load patterns [11]. Recent advances in neural architectures have shown promise in addressing these limitations. Long-term load forecasting has been approached through hierarchical neural models with time integrators [12], while machine learning techniques have been applied to traffic classification and prediction tasks [13]. Additionally, recurrent neural networks have been successfully applied to long-term generating power forecasting for wind power systems [14].
To address this issue, we propose a model called MP-RWKV-TS. As an improvement on the efficient RWKV-TS framework [15], it combines the advantages of recurrent networks’ good fit in short-term correlations with transformer-based models’ good fit in long-range dependence capabilities.
Our work makes a number of significant contributions to the field of long-term power load forecasting:
  • We introduce MP-RWKV-TS, a combined model that extends the capabilities of RWKV-TS by utilizing parallel computation paths to identify both temporal patterns and cross-variable relationships.
  • We utilize an Adaptive Time-Mixing Operator (ATMO) to enhance the model’s temporal variable dependency strength across timescales.
  • Our approach incorporates a Cross-Variable Attention (CVA) mechanism that operates in parallel with temporal processing, enabling the modeling of complex interactions between factors affecting power load.
  • We conduct thorough and detailed experiments on real-world data from a North African urban power grid distribution network, demonstrating the advantages of MP-RWKV-TS in long-term forecasting scenarios.
  • We also introduce the interpretability aspects of the model to provide greater insight into the factors responsible for long-term power load dynamics.
The rest of this paper is organized as follows: Section 2 formulates the problem, Section 3 details our proposed MP-RWKV model, Section 4 presents the experimental results, and Section 5 concludes the paper.

2. Problem Formulation

Let D denote a dataset spanning N years containing both power load measurements and associated meteorological variables. We define the following:

2.1. Input Variables

For any time step t and dayd:

2.1.1. Weather Variables

Let W t , d R M denote the weather feature vector:
W t , d = [ Temp t , d , Hum t , d , Ws t , d ] T
where
  • Temp t , d R represents temperature;
  • Hum t , d R represents humidity;
  • Ws t , d R represents wind speed.

2.1.2. Power Load

Let P t , d R denote the power load at time step t of day d.

2.2. Temporal Modeling

Given a temporal context window τ , we construct the input feature vector x t , d as
x t , d = [ P t τ : t 1 , d ; W t τ : t 1 , d ] R ( τ × ( M + 1 ) )
where
  • P t τ : t 1 , d = [ P t τ , d , . . . , P t 1 , d ] represents historical load values;
  • W t τ : t 1 , d = [ W t τ , d , . . . , W t 1 , d ] represents historical weather conditions.

2.3. Prediction Objective

The forecasting task can be formulated as learning a function f:
P ^ t , d = f ( x t , d ; θ )
where
  • P ^ t , d is the predicted power load;
  • θ represents the model parameters;
  • f : R ( τ × ( M + 1 ) ) R is the prediction function.
The objective is to minimize the prediction error:
min θ L ( P t , d , P ^ t , d )
where L is an appropriate loss function measuring the discrepancy between predicted and actual power load values.

3. The Proposed Model

3.1. Theoretical Foundation

The MP-RWKV architecture extends the RWKV-TS model [15] by incorporating principles of parallel information processing inspired by cognitive science theories. While maintaining the core efficiency of RWKV-TS, our enhanced model introduces parallel processing paths to concurrently handle different aspects of temporal data. This approach allows the model to simultaneously process temporal patterns and cross-variable dependencies, similar to how biological systems process sequential and relational information.
Figure 1 illustrates the foundational RWKV-TS architecture that serves as the basis for our enhanced MP-RWKV model.

3.2. Input Processing and Representation

Consider a multivariate time series x R L × M , where L represents the sequence length and M denotes the number of variables. The input processing consists of three critical stages:

3.2.1. Instance Normalization

To ensure numerical stability and consistent gradient flow, each univariate sequence undergoes instance normalization:
x i = x i mean ( x i ) std ( x i )
where x i R L represents the i-th variable sequence.

3.2.2. Patch Segmentation

The normalized sequence is segmented into overlapping patches to capture local temporal structures:
N = L P S + 2
where P denotes patch length and S represents stride length. This segmentation creates patches x i p R N × P .

3.2.3. Token Projection

The projection phase maps the segmented patches into a high-dimensional feature space through a learned linear transformation. Given patches x i p R N × P , where N denotes the number of patches and P represents the patch length, the projection is formulated as follows:
x i t = x i p W p
where W p R P × D is a learnable projection matrix that transforms the input to dimension D. This transformation enables the model to learn rich feature representations while preserving temporal relationships within the sequence. The projected tokens x i t R N × D serve as the foundation for subsequent temporal and cross-variable processing in the multi-path architecture.

3.3. Parallel Path Processing

Consider a multivariate time series { x t } t = 1 T , where x t R D represents the input embedding at time step t. The enhanced architecture decomposes temporal processing into two complementary paths: a temporal path for sequential pattern extraction and a channel path for cross-variable dependency modeling. This decomposition enables specialized processing while maintaining parameter efficiency through strategic weight sharing.
As shown in Figure 2, our MP-RWKV architecture introduces parallel processing pathways that enable specialized handling of temporal patterns and cross-variable relationships.

3.3.1. Time-Mixing Path

The temporal path implements sequential pattern extraction through a hierarchical process consisting of adaptive mixing and position-aware attention mechanisms.

Adaptive Time-Mixing Operator (ATMO)

The ATMO computes dynamic interpolation coefficients to adaptively balance temporal dependencies:
μ t = σ ( W μ x t ) R 2 D
where W μ R 2 D × D defines the mixing transformation and σ ( · ) ensures the coefficients lie in ( 0 , 1 ) . The temporal representations are then computed through weighted interpolation:
[ k t T , v t T ] = W k v ( μ t [ 1 : D ] x t + ( 1 μ t [ 1 : D ] ) x t 1 )
where W k v R D × 2 D projects the interpolated state into key–value pairs. This formulation enables the model to dynamically adjust the influence of recent and historical information.

Time-Aware Attention

The temporal context is integrated through position-sensitive attention:
c t T = TimeAwareAttention ( k t T , v t T , t )
with the attention mechanism defined as
TimeAwareAttention ( k , v , t ) = softmax k T k d k + T t v
The temporal bias matrix T t = TemporalBias ( t ) R D × D encodes relative positional information, enabling the model to capture position-dependent patterns in the sequence.

3.3.2. Channel-Mixing Path

The channel path focuses on modeling relationships between variables through parameter-efficient transformations and gated attention mechanisms.

Cross-Variable Processing

The channel-specific representations leverage the latter half of the mixing coefficients:
[ k t C , v t C ] = W k v ( μ t [ D + 1 : ] x t )
This formulation shares the projection matrix W k v with the temporal path, promoting parameter efficiency while maintaining representational capacity.

Channel-Aware Attention

Cross-variable dependencies are modeled through a gated attention mechanism:
c t C = ChannelAwareAttention ( k t C , v t C , g t C )
defined as
ChannelAwareAttention ( k , v , g ) = softmax k T k d k g v
where g t C R D serves as an adaptive gating mechanism that modulates attention weights based on the relative importance of different variables. The scaling factor d k ensures stable gradient propagation during training.

3.4. Path Integration and Output Generation

The integration of temporal and channel paths is achieved through an adaptive fusion mechanism that dynamically balances their contributions while maintaining residual connections to the input signal.

Adaptive Path Fusion

Let o t T M R D and o t C M R D denote the outputs from temporal and channel paths, respectively. The fusion mechanism computes dynamic weights through
α t = σ ( W α [ o t T M ; o t C M ] ) R D
where W α R D × 2 D is a learnable projection matrix and [ · ; · ] denotes concatenation.
The paths are integrated through weighted combination with residual connection:
o t = α t o t T M + ( 1 α t ) o t C M + x t
This formulation ensures that the model can adaptively prioritize temporal or cross-variable patterns while maintaining access to the original input features.

3.5. Advancing upon RWKV-TS

Our MP-RWKV model builds upon the foundation established by RWKV-TS [15], introducing several enhancements that specifically address the challenges of long-term power load forecasting:
  • Parallel processing paths: While RWKV-TS processes all variables through a single computational path, MP-RWKV introduces parallel processing paths that enable specialized handling of temporal patterns and cross-variable relationships simultaneously. This allows the model to maintain the computational efficiency of RWKV-TS while expanding its representational capacity.
  • Cross-Variable Attention (CVA): MP-RWKV enhances the original model with a dedicated path for modeling relationships between different variables through a specialized attention mechanism. This is particularly crucial for power load forecasting, where interactions between meteorological factors and temporal indicators significantly impact prediction accuracy.
  • Adaptive path integration: Unlike the single-path processing in RWKV-TS, our model dynamically balances the contributions of temporal and cross-variable information based on the forecasting context, leading to more adaptive and robust predictions across varying horizons.
  • Enhanced gradient flow: By separating concerns into parallel paths, MP-RWKV mitigates potential gradient issues that can affect single-path recurrent models when dealing with long-term dependencies.
These enhancements maintain the computational efficiency that makes RWKV-TS attractive while extending its capability to handle the complex temporal and cross-variable relationships present in power load data across multiple time scales.

3.6. Training Objective

The model is optimized through a composite objective function that combines reconstruction accuracy with temporal consistency:
L ( θ ) = L M S E ( θ ) + λ L t e m p o r a l ( θ )
where θ denotes the model parameters and λ > 0 is a hyperparameter controlling the trade-off between objectives.
The mean-squared error term measures prediction accuracy:
L M S E ( θ ) = 1 n i = 1 n y i y ^ i 2 2
where y i and y ^ i represent the ground truth and predicted values, respectively.
The temporal regularization term promotes smoothness in the temporal representations:
L t e m p o r a l ( θ ) = o t T M o t 1 T M 2
This regularization encourages the model to learn continuous temporal patterns while allowing for abrupt changes when supported by the data.
The optimization problem can be formally stated as
θ * = a r g m i n θ L ( θ )
solved through stochastic gradient descent with adaptive moment estimation:
θ t + 1 = θ t η · m ^ t v ^ t + ϵ
where m ^ t and v ^ t are bias-corrected first- and second-moment estimates of the gradients.

3.7. Complexity Analysis

We analyze the computational and space complexity of MP-RWKV-TS in terms of sequence length L and feature dimension D. The analysis considers both the forward pass computation and memory requirements.
Table 1 presents the computational and space complexity analysis for different sequence models, where L represents sequence length and D represents feature dimension. While Transformers offer full parallelization, they suffer from quadratic complexity in sequence length. Recurrent models like RNN and LSTM have linear time complexity but cannot be parallelized. Our proposed MP-RWKV-TS model achieves linear time complexity while supporting partial parallelization, offering a balanced trade-off between efficiency and computational requirements.

4. Experiments

Extensive experimentation on diverse time series datasets was conducted to validate the effectiveness of MP-RWKV in temporal pattern recognition and forecasting tasks. The model is systematically evaluated against state-of-the-art architectures, including RWKV-TS, Transformer variants, and traditional forecasting methods, using three complementary performance metrics: mean-squared error (MSE) to quantify prediction accuracy with emphasis on larger deviations, Mean Absolute Error (MAE) to assess absolute prediction differences, and Mean Absolute Percentage Error (MAPE) to measure relative accuracy across different scales. Furthermore, comprehensive ablation studies were performed to analyze the individual contributions of key architectural components, including the Adaptive Time-Mixing Operator, Cross-Variable Attention mechanism, and multi-path integration strategy. This rigorous evaluation framework demonstrates the superiority of MP-RWKV in capturing complex temporal and cross-variable dependencies while providing insights into the model’s interpretability and the effectiveness of its architectural enhancements.

4.1. Dataset Description

4.1.1. Data Collection

The dataset comprises power consumption measurements from three distinct regions (Quads, Smir, and Boussafou) in the urban power distribution network of Tetouan, Morocco. Data collection was conducted throughout 2017, resulting in 52,416 samples with a granularity of 10-minute intervals.

4.1.2. Feature Composition

Each data point consists of nine features:

Meteorological Variables

  • Temperature ( Temp t , d R );
  • Relative Humidity ( Hum t , d R );
  • Wind Speed ( Ws t , d R ).

Temporal Indicators

  • Hour of Day ( h { 0 , 1 , , 23 } );
  • Day of Week ( d w { 1 , 2 , , 7 } );
  • Month ( m { 1 , 2 , , 12 } ).

Power Generation

  • Region 1 Power Load ( P t , d 1 R + );
  • Region 2 Power Load ( P t , d 2 R + );
  • Region 3 Power Load ( P t , d 3 R + ) - Target Variable.

4.1.3. Data Characteristics

The dataset exhibits the following properties:
  • Temporal coverage: 365 days;
  • Sampling frequency: 144 samples per day;
  • Total duration: 8736 h;
  • Missing values: <0.1%;
  • Data range: Normalized to [0,1] for each feature.
Figure 3 presents the statistical characteristics of weather factors and power consumption across the three monitored regions, providing insights into the data distribution and seasonal patterns.

4.2. Baseline Models for Comparison

We evaluate our proposed model against several state-of-the-art architectures specialized in time series forecasting:

Advanced Neural Architectures

  • RWKV-TS [15]: The foundation model upon which our enhanced MP-RWKV is built, featuring efficient recurrent computation with the ability to handle long sequences.
  • Transformer [16]: The canonical attention-based architecture that revolutionized sequential modeling. Recent studies have demonstrated its effectiveness in Long-Sequence Time Series Forecasting (LSTF).
  • Autoformer [17]: A Transformer-based model incorporating decomposition blocks to explicitly model seasonal–temporal patterns, specifically designed for long-horizon forecasting.
  • Informer [18]: An enhanced Transformer variant optimized for long-term sequence prediction, featuring improved computational efficiency through sparse attention mechanisms.
  • Reformer [19]: A memory-efficient variant utilizing locality-sensitive hashing for attention computation, achieving O ( L log L ) complexity while maintaining comparable performance.
  • LSTM [20]: A foundational recurrent architecture that addresses gradient instability through gating mechanisms, enabling effective modeling of long-term dependencies.

4.3. Evaluation Metrics

Model performance is evaluated using two standard metrics: Mean Absolute Error (MAE) and mean-squared error (MSE). Let { y i } i = 1 n denote the ground truth values and { y ^ i } i = 1 n denote the model predictions.

4.3.1. Mean Absolute Error

MAE measures the average magnitude of prediction errors:
MAE = 1 n i = 1 n | y ^ i y i |
where n represents the number of prediction points.

4.3.2. Mean-Squared Error

MSE quantifies the average squared deviation of predictions:
MSE = 1 n i = 1 n ( y ^ i y i ) 2
These metrics provide complementary perspectives on model performance:
  • MAE offers an interpretable measure of prediction accuracy in the original scale
  • MSE penalizes larger errors more heavily, making it particularly sensitive to outliers
For both metrics, lower values indicate better performance. The statistical significance of performance differences is assessed through paired t-tests with p < 0.05 .

4.4. Short-Term Forecasting Results

Table 2 presents the performance comparison of different models on short-term power load forecasting tasks using 48 h of input data to predict power loads at various horizons from 24 h to 120 h.
Figure 4 shows the MAE performance of the selected models using 48 h of data to predict power loads at various horizons. In short-term forecasting scenarios, the MP-RWKV model demonstrates superior performance compared to other models, indicating that preserving past information through context states is crucial for effective prediction. The RWKV-TS model, which also utilizes a similar mechanism, shows comparable performance, highlighting the effectiveness of this approach. Notably, our enhanced MP-RWKV consistently outperforms the base RWKV-TS model, demonstrating the value of parallel processing paths for handling temporal patterns and cross-variable relationships simultaneously.

4.5. Long-Term Forecasting Results

Table 3 presents the performance comparison for long-term forecasting scenarios, where models use 8 h of input data to predict power loads at extended horizons from 144 h to 432 h.
Figure 5 presents the long-term forecasting results showing MAE performance when using 8 h of data to predict power loads at horizons ranging from 144 h to 432 h. In these extended prediction scenarios, traditional Transformer-based models (Informer, Autoformer, and Reformer) show significant performance deterioration as the prediction horizon increases. The MP-RWKV model maintains consistent performance even over extended forecasting periods, demonstrating its robustness through parallel processing paths, adaptive fusion mechanisms, and position-aware attention that effectively capture both short-term patterns and long-term dependencies.
The experimental results demonstrate that MP-RWKV consistently outperforms the base RWKV-TS model across all long-term forecasting horizons, with the performance gap increasing at more extended prediction ranges. This indicates that the parallel pathway processing and adaptive integration of temporal and cross-variable information become increasingly beneficial as the forecasting horizon extends. While the base LSTM model shows declining accuracy as the prediction horizon increases due to gradient vanishing issues, the MP-RWKV model maintains stable performance across all tested scenarios.

5. Conclusions

This paper introduces MP-RWKV, an enhanced model that builds upon RWKV-TS to address the limitations of existing approaches in time series forecasting, particularly for power load prediction. Through our comprehensive experimental analysis, we demonstrate that MP-RWKV consistently outperforms other state-of-the-art models across both short-term and long-term forecasting horizons.
The experimental results show that MP-RWKV achieves superior performance in short-term forecasting scenarios, effectively capturing temporal patterns through its context state mechanism. In long-term forecasting, while traditional models like LSTM suffer from gradient vanishing and Transformer-based models show deteriorating performance, MP-RWKV maintains consistent accuracy even at extended prediction horizons, demonstrating its robust handling of temporal dependencies.
The key findings from our experiments include
  • MP-RWKV demonstrates consistently lower MAE values compared to baseline models, including its predecessor RWKV-TS
  • The model maintains stable performance across different prediction horizons, from 24 h to 432 h
  • The parallel processing path approach effectively balances short-term pattern recognition with long-term dependency modeling
  • The adaptive fusion mechanism successfully integrates temporal and cross-variable information, leading to more robust predictions
Future research directions include
  • Extending the model to handle multi-variable forecasting scenarios with even more diverse data sources
  • Incorporating additional parallel processing paths for specialized handling of different data characteristics
  • Exploring applications in other domains requiring time series forecasting
  • Optimizing computational efficiency for real-world deployments
In conclusion, MP-RWKV represents a significant advancement in power load forecasting, building upon and enhancing the capabilities of RWKV-TS. Its consistent performance and versatility make it a promising solution for power system operations and planning, contributing to the development of more efficient and reliable smart grid technologies.

Author Contributions

Conceptualization, A.R. and A.T.; methodology, A.R.; software, A.R. and A.E.; validation, A.R., A.T. and M.A.; formal analysis, A.R.; investigation, A.R.; resources, A.T.; data curation, A.E.; writing—original draft preparation, A.R.; writing—review and editing, A.T. and M.A.; visualization, A.E.; supervision, A.T.; project administration, A.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tang, L.; Wang, X.; Wang, X.; Shao, C.; Liu, S.; Tian, S. Long-term electricity consumption forecasting based on expert prediction and fuzzy Bayesian theory. Energy 2019, 167, 1144–1154. [Google Scholar] [CrossRef]
  2. Kim, T.Y.; Cho, S.B. Predicting the household power consumption using CNN-LSTM hybrid networks. In Intelligent Data Engineering and Automated Learning–IDEAL 2018: 19th International Conference, Madrid, Spain, November 21–23, 2018, Proceedings, Part I 19; Springer: Berlin/Heidelberg, Germany, 2018; pp. 481–490. [Google Scholar]
  3. Fan, D.; Sun, H.; Yao, J.; Zhang, K.; Yan, X.; Sun, Z. Well production forecasting based on ARIMA-LSTM model considering manual operations. Energy 2021, 220, 119708. [Google Scholar] [CrossRef]
  4. Ahmad, T.; Zhang, D.; Huang, C.; Zhang, H.; Dai, N.; Song, Y.; Chen, H. Artificial intelligence in sustainable energy industry: Status Quo, challenges and opportunities. J. Clean. Prod. 2021, 289, 125834. [Google Scholar] [CrossRef]
  5. Yu, C.; Li, Y.; Chen, Q.; Lai, X.; Zhao, L. Matrix-based wavelet transformation embedded in recurrent neural networks for wind speed prediction. Appl. Energy 2022, 324, 119692. [Google Scholar] [CrossRef]
  6. Nowicka-Zagrajek, J.; Weron, R. Modeling electricity loads in California: ARMA models with hyperbolic noise. Signal Process. 2002, 82, 1903–1915. [Google Scholar] [CrossRef]
  7. Chen, X.; Jia, S.; Ding, L.; Xiang, Y. Reasoning over temporal knowledge graph with temporal consistency constraints. J. Intell. Fuzzy Syst. 2021, 40, 11941–11950. [Google Scholar] [CrossRef]
  8. Valipour, M.; Banihabib, M.E.; Behbahani, S.M.R. Comparison of the ARMA, ARIMA, and the autoregressive artificial neural network models in forecasting the monthly inflow of Dez dam reservoir. J. Hydrol. 2013, 476, 433–441. [Google Scholar] [CrossRef]
  9. Divina, F.; Gilson, A.; Goméz-Vela, F.; García Torres, M.; Torres, J.F. Stacking ensemble learning for short-term electricity consumption forecasting. Energies 2018, 11, 949. [Google Scholar] [CrossRef]
  10. Gu, B.; Shen, H.; Lei, X.; Hu, H.; Liu, X. Forecasting and uncertainty analysis of day-ahead photovoltaic power using a novel forecasting method. Appl. Energy 2021, 299, 117291. [Google Scholar] [CrossRef]
  11. Wang, J.; Gao, J.; Wei, D. Electric load prediction based on a novel combined interval forecasting system. Appl. Energy 2022, 322, 119420. [Google Scholar] [CrossRef]
  12. Carpinteiro, O.A.S.; Leme, R.C.; de Souza, A.C.Z.; Pinheiro, C.A.M.; Moreira, E.M. Long-term load forecasting via a hierarchical neural model with time integrators. Electr. Power Syst. Res. 2007, 77, 371–378. [Google Scholar] [CrossRef]
  13. Eldeeb, E. Traffic Classification and Prediction, and Fast Uplink Grant Allocation for Machine Type Communications via Support Vector Machines and Long Short-Term Memory. Master’s Thesis, University of British Columbia, Vancouver, BC, Canada, 2020. [Google Scholar]
  14. Senjyu, T.; Yona, A.; Urasaki, N.; Funabashi, T. Application of recurrent neural network to long-term-ahead generating power forecasting for wind power generator. In Proceedings of the 2006 IEEE PES Power Systems Conference and Exposition, Atlanta, GA, USA, 29 October–1 November 2006; pp. 1260–1265. [Google Scholar]
  15. Hou, H.; Yu, F.R. RWKV-TS: Beyond traditional recurrent neural network for time series tasks. arXiv 2024, arXiv:2401.09093. [Google Scholar]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2017; pp. 5998–6008. [Google Scholar]
  17. Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2021; Volume 34, pp. 22419–22430. [Google Scholar]
  18. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence; AAAI Press: Palo Alto, CA, USA, 2021; Volume 35, pp. 11106–11115. [Google Scholar]
  19. Kitaev, N.; Kaiser, L.; Levskaya, A. Reformer: The efficient transformer. arXiv 2020, arXiv:2001.04451. [Google Scholar]
  20. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
Figure 1. RWKV-TS architecture overview showing the foundational structure upon which our MP-RWKV model builds. The architecture demonstrates the core components including input processing, token embedding, and the recurrent weighted key–value mechanism that enables efficient temporal modeling while maintaining linear computational complexity.
Figure 1. RWKV-TS architecture overview showing the foundational structure upon which our MP-RWKV model builds. The architecture demonstrates the core components including input processing, token embedding, and the recurrent weighted key–value mechanism that enables efficient temporal modeling while maintaining linear computational complexity.
Engproc 97 00026 g001
Figure 2. Power load forecasting architecture based on MP-RWKV-TS showing the parallel processing paths for temporal and cross-variable modeling. The architecture demonstrates how the enhanced model processes multivariate time series data through specialized pathways while maintaining computational efficiency. Key components include the Adaptive Time-Mixing Operator (ATMO) and Cross-Variable Attention (CVA) mechanisms.
Figure 2. Power load forecasting architecture based on MP-RWKV-TS showing the parallel processing paths for temporal and cross-variable modeling. The architecture demonstrates how the enhanced model processes multivariate time series data through specialized pathways while maintaining computational efficiency. Key components include the Adaptive Time-Mixing Operator (ATMO) and Cross-Variable Attention (CVA) mechanisms.
Engproc 97 00026 g002
Figure 3. Statistical analysis of weather factors and electricity consumption patterns across three regions. The figure presents comprehensive data characteristics showing temperature variations (measured in degrees Celsius), humidity levels (percentage), wind speed patterns (m/s), and corresponding power consumption profiles for Regions 1, 2, and 3. Data distributions are presented using scientific notation where appropriate (e.g., 8 × 10 3 instead of 8E3) to maintain clarity and precision in the visualization.
Figure 3. Statistical analysis of weather factors and electricity consumption patterns across three regions. The figure presents comprehensive data characteristics showing temperature variations (measured in degrees Celsius), humidity levels (percentage), wind speed patterns (m/s), and corresponding power consumption profiles for Regions 1, 2, and 3. Data distributions are presented using scientific notation where appropriate (e.g., 8 × 10 3 instead of 8E3) to maintain clarity and precision in the visualization.
Engproc 97 00026 g003
Figure 4. Performance comparison of different models on short-term power load forecasting showing MAE values across prediction horizons from 24 h to 120 h. The results demonstrate MP-RWKV’s consistent superiority over baseline models, with particularly strong performance in the 24–72 h range where immediate temporal correlations are most critical.
Figure 4. Performance comparison of different models on short-term power load forecasting showing MAE values across prediction horizons from 24 h to 120 h. The results demonstrate MP-RWKV’s consistent superiority over baseline models, with particularly strong performance in the 24–72 h range where immediate temporal correlations are most critical.
Engproc 97 00026 g004
Figure 5. Long-term forecasting performance comparison showing MAE values for prediction horizons from 144 h to 432 h. The results highlight MP-RWKV’s ability to maintain consistent accuracy across extended prediction ranges, where traditional models suffer from significant performance degradation due to vanishing gradients and attention mechanism limitations.
Figure 5. Long-term forecasting performance comparison showing MAE values for prediction horizons from 144 h to 432 h. The results highlight MP-RWKV’s ability to maintain consistent accuracy across extended prediction ranges, where traditional models suffer from significant performance degradation due to vanishing gradients and attention mechanism limitations.
Engproc 97 00026 g005
Table 1. Computational and space complexity comparison.
Table 1. Computational and space complexity comparison.
ModelTime ComplexitySpace ComplexityParallelizable
Transformer O ( L 2 D ) O ( L 2 + L D ) Yes
RNN O ( L D 2 ) O ( D ) No
LSTM O ( L D 2 ) O ( D ) No
MP-RWKV-TS O ( L D 2 ) O ( L D + D 2 ) Partial
Table 2. Performance comparison of different models on short-term power load forecasting (48 h input).
Table 2. Performance comparison of different models on short-term power load forecasting (48 h input).
HoursMP-RWKV-TSRWKV-TSInformerAutoformerTransformerReformerLSTM
MSE MAE MSE MAE MSE MAE MSE MAE MSE MAE MSE MAE MSE MAE
240.05520.18450.05850.19020.05970.19290.24640.39970.11530.29430.29830.47380.07870.2245
480.07980.23150.08250.23780.09100.24241.21210.90720.18550.37660.31960.49220.10570.2533
720.08120.23250.08350.23780.09640.25840.27710.41150.18660.37790.44230.58950.12990.2988
960.09980.26450.10350.27050.11650.28790.89020.76590.23540.43960.42280.57310.12330.2879
1200.10250.26850.10650.27550.11590.28480.47390.53710.21060.40430.44120.58580.13720.3157
Note: Bold values indicate the best performance for each metric. MSE: mean-squared error; MAE: Mean Absolute Error. Input sequence length is fixed at 48 h for all experiments.
Table 3. Performance comparison of different models on long-term power load forecasting.
Table 3. Performance comparison of different models on long-term power load forecasting.
HoursMP-RWKV-TSRWKV-TSInformerAutoformerTransformerReformerLSTM
MSE MAE MSE MAE MSE MAE MSE MAE MSE MAE MSE MAE MSE MAE
1440.19550.38250.20150.38950.26900.45610.70610.61810.26920.45170.89490.86720.75090.7675
1920.24850.43150.25680.43980.29790.48131.05630.72840.27360.45660.87620.84490.81280.8054
2400.22550.41050.23350.41820.26910.45420.46640.54510.25100.43530.88680.84470.59700.6731
2880.23350.40850.24150.41620.25440.43051.32930.77130.21110.39200.90920.85780.70850.7285
4320.18950.36250.19650.37050.27860.44400.45190.51480.23570.41730.92530.85920.78080.7629
Note: Bold values indicate the best performance for each metric. MSE: mean-squared error; MAE: Mean Absolute Error. Input sequence length is fixed at 8 h for all experiments.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rizki, A.; Touil, A.; Echchatbi, A.; Ahlaqqach, M. A Parallel Processing Architecture for Long-Term Power Load Forecasting. Eng. Proc. 2025, 97, 26. https://doi.org/10.3390/engproc2025097026

AMA Style

Rizki A, Touil A, Echchatbi A, Ahlaqqach M. A Parallel Processing Architecture for Long-Term Power Load Forecasting. Engineering Proceedings. 2025; 97(1):26. https://doi.org/10.3390/engproc2025097026

Chicago/Turabian Style

Rizki, Adil, Achraf Touil, Abdelwahed Echchatbi, and Mustapha Ahlaqqach. 2025. "A Parallel Processing Architecture for Long-Term Power Load Forecasting" Engineering Proceedings 97, no. 1: 26. https://doi.org/10.3390/engproc2025097026

APA Style

Rizki, A., Touil, A., Echchatbi, A., & Ahlaqqach, M. (2025). A Parallel Processing Architecture for Long-Term Power Load Forecasting. Engineering Proceedings, 97(1), 26. https://doi.org/10.3390/engproc2025097026

Article Metrics

Back to TopTop