Next Article in Journal
A Missing Data Imputation Method for Distribution Network Data Based on TGAN-GP
Previous Article in Journal
Lithium-Ion Battery Open-Circuit Voltage Analysis for Extreme Temperature Applications
Previous Article in Special Issue
Low-Carbon Economic Dispatch Model for Virtual Power Plants Considering Multi-Type Load Demand Response
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ultra-Short-Term Multi-Step Photovoltaic Power Forecasting Based on Similarity-Based Daily Clustering

1
Hubei Key Laboratory of Advanced Technology for Automotive Components, School of Automotive Engineering, Wuhan University of Technology, Wuhan 430070, China
2
Shanxi Transportation Investment and Financing Group Co., Ltd., Taiyuan 030006, China
*
Author to whom correspondence should be addressed.
Energies 2026, 19(1), 29; https://doi.org/10.3390/en19010029 (registering DOI)
Submission received: 31 October 2025 / Revised: 6 December 2025 / Accepted: 11 December 2025 / Published: 20 December 2025

Abstract

Photovoltaic (PV) power generation is inherently intermittent and volatile, complicating power system operation and control. Accurate forecasting is crucial for proactive grid responses and optimal energy resource scheduling. This study proposes a novel hybrid forecasting model that achieves high-precision PV power forecasting by integrating similar-day clustering, generating extreme weather samples, and optimizing the Bidirectional Temporal Convolutional Network (BiTCN) and Bidirectional Gated Recurrent Unit (BiGRU) model via the Animated Oat Optimization (AOO) algorithm. The proposed method outperforms other models in the three evaluation metrics of mean absolute error (MAE), root mean square error (RMSE), and coefficient of determination (R2). The innovations lie in the integration of similar-day clustering with deep learning and the application of AOO for hyperparameter optimization, which significantly enhances forecasting accuracy and robustness.

1. Introduction

Solar energy is increasingly being utilized on a global scale, largely because it stands out as an abundant and environmentally sustainable renewable option. As a key candidate to meet future energy demands, photovoltaic (PV) power generation has attracted significant attention in modern power grids for its efficient use of solar energy. It is projected that by 2025, PV generation will supply one-quarter of the world’s electricity demand [1]. However, the intermittent nature of solar energy necessitates that the generated electricity be either consumed immediately or stored in costly battery systems. To maintain a stable energy supply and enable effective resource planning, achieving high accuracy in PV power forecasting is critical.
The primary approaches for predicting PV power currently encompass physical models, statistical and probabilistic methods, and machine learning models. Physical models do not rely on historical data but rather compute PV power output directly using meteorological information (e.g., weather forecasts, satellite cloud maps) and system parameters such as PV panel installation angles and surface temperatures. In studies based on physical modeling, Zhi et al. [2] constructed a model based on physical principles for PV power forecasting, which incorporates both meteorological parameter prediction and an enhanced MPPT algorithm. Experimental results under various weather conditions demonstrated low prediction error, thereby improving both accuracy and applicability. El Ainaoui et al. [3] introduced new mathematical formulations using implicit single-diode models (SDM) and explicit Das models (DM) to describe variations in model parameters with temperature and irradiance. Field tests at the Moroccan Green Energy Park showed that their method could effectively predict PV performance, achieving an NRMSE below 4.16%, contributing to PV system optimization and enhanced grid stability. Zaimi et al. [4] proposed two new techniques for identifying the physical parameters of single-diode circuit models, which rely on establishing correlations between key photovoltaic indicators and quality factors. Manufacturer data were utilized to determine temperature and radiation coefficients, investigating the influence of environmental factors on model parameters. Testing on KC130GT and SM55 photovoltaic panels validated the effectiveness of the methods and models, providing support for photovoltaic system performance prediction. Tifidat et al. [5] proposed a novel simulation approach for PV module performance based on the single-diode model. By combining numerical and analytical techniques, they reduced the number of unknown parameters and eliminated the need for approximation and iteration, relying only on data under standard test conditions and simplifying the modeling process. Comparative experiments on PV modules using various technologies showed lower error and faster convergence compared to other methods, providing an effective tool for dynamic performance evaluation and design. Although physical modeling does not require historical data and is applicable to any prediction horizon, it faces practical challenges. A major constraint is the need for highly detailed, site-specific parameters tailored to large-scale PV plants, which makes these models less suitable for residential rooftop PV systems [6]. Furthermore, compared to data-driven methods, physical models demand more domain knowledge, as they rely on solving complex differential equations based on mathematical descriptions of photovoltaic conversion. This not only imposes high computational costs but also leads to longer runtimes to achieve accurate results. Despite a solid understanding of the underlying physical phenomena and accurate modeling, even mature and well-studied models can introduce errors [7]. Many fail to account for localized effects such as system faults, cloud movement, partial shading, or snow accumulation—factors that data-driven methods are better equipped to handle [8].
Statistical approaches rely on the analysis of historical data to construct a functional relationship between past and predicted values. Classical statistical techniques such as vector autoregression (VAR) [9] and autoregressive integrated moving average (ARIMA) models [10] were dominant in early studies. Li et al. [11] proposed an autoregressive moving average model with exogenous inputs (ARMAX), incorporating meteorological features to forecast power output. By using easily obtainable parameters such as temperature, precipitation, sunshine duration, and humidity, the model retained the simplicity of ARIMA without relying on solar irradiance forecasts. It also proved to be more general and flexible in practical applications. Based on real-world data from a 2.1 kW grid-connected PV system, the ARMAX model significantly outperformed ARIMA in prediction accuracy. Despite advantages such as structural simplicity and fast computation, statistical models struggle to capture the volatile and dynamic nature of PV data, limiting their predictive performance [12].
In recent years, machine learning has been widely applied in PV power forecasting due to its ease of modeling, high accuracy, and strong generalization capability. Sulaiman et al. [13] successfully applied neural network (NN) models for forecasting the power generation of rooftop PV systems. The authors reported superior accuracy and consistency over alternative approaches. For day-ahead forecasting in a real microgrid, an integrated model (WT-PSO-SVM) was formulated by Eseye et al. [14], which merges wavelet transform, particle swarm optimization, and support vector machine. Wavelet transform was used for preprocessing, and particle swarm optimization for hyperparameter tuning, resulting in improved prediction performance. Wolff et al. [15] evaluated support vector regression (SVR) against physical models by utilizing a fused dataset, which included PV power observations, numerical weather prediction (NWP), and cloud motion vector (CMV) information. SVR showed excellent performance in real microgrid environments. Machine learning models have shown significant promise for PV forecasting; however, their effectiveness is often constrained when it comes to fully modeling the intricate nonlinear dynamics between weather conditions and PV power. It often struggles to capture dynamic phenomena such as sudden cloud cover changes and component degradation, and lacks long-range modeling capabilities, which constrain further accuracy improvements. Overcoming these limitations calls for models with stronger feature extraction and time-series modeling capabilities, paving the way for deep learning applications in PV forecasting.
Owing to its powerful strengths in feature extraction, nonlinear modeling, and generalization, deep learning has emerged as an increasingly prominent subfield of machine learning. Recurrent neural networks (RNNs) [16] are frequently applied to time-series prediction problems. Lee et al. [17] utilized two types of RNNs to forecast PV power without requiring future meteorological information, achieving better results than traditional artificial neural networks (ANNs) and deep neural networks (DNNs). However, RNNs suffer from long-term dependency issues, leading to vanishing or exploding gradients. Hochreiter et al. [18] presented the long short-term memory (LSTM) network to tackle this issue. By leveraging a mechanism of input, forget, and output gates plus a memory cell, the LSTM network effectively overcomes long-term dependency issues, making it a broadly applied tool in PV forecasting. Hu et al. [19] further combined LSTM with self-attention mechanisms, integrating historical and forecast data. Their model, tested on real PV data from a building in Japan, exhibited strong accuracy and adaptability. Ahmed et al. [20] proposed an efficient and practical integrated forecasting method combining LSTM with live data from the Yulara PV plant in Australia. Other commonly used deep learning models include gated recurrent units (GRUs) [21] and convolutional neural networks (CNNs) [22]. Dai et al. [23] used LOWESS smoothing to extract PV features and proposed a CNN-BiGRU-Attention hybrid model optimized via ensemble learning. Chen et al. [24] employed a multi-task learning (MTL) scheme for TPA-LSTM to jointly forecast wind–solar. Their approach utilized the maximal information coefficient for feature selection and wind–PV correlation analysis. The MTL framework’s innovation lay in separating shared and task-specific components, which was further enhanced by a novel loss strategy balancing training speed and loss magnitude. Qu et al. [25] introduced an attention-based long- and short-term memory prediction model (ALSM) under a multi-related and target-variable prediction (MRTPP) scheme. The model combines CNN, LSTM, and attention mechanisms and was validated using historical PV system data from the DKSCA website. Results showed that the MRTPP-based ALSM outperformed traditional input–output schemes and various statistical and AI-based forecasting methods. Furthermore, improved LSTM-based algorithms are widely adopted in PV power forecasting. Bi-LSTM’s bidirectional learning enhances accuracy and convergence speed in time-series tasks [26,27], while GRU, by simplifying LSTM’s gate structure into a unified update gate, offers fewer parameters and easier training, contributing to its popularity [28,29,30,31,32,33].
In summary, while existing PV power forecasting methods have addressed many challenges, several issues remain. First, physics-based methods suffer from low prediction accuracy, poor generalization, and complex parameterization. Traditional statistical methods are inadequate in capturing key PV data characteristics. Moreover, conventional machine learning models are prone to gradient issues under long-term dependencies. Although hybrid models combining various algorithms have become mainstream in PV forecasting, they often focus solely on model structure, neglecting the role of meteorological data analysis. In response to these limitations, a deep learning-based PV power forecasting model that integrates similar-day clustering is proposed in this study, whose structure is illustrated in Figure 1. The main contributions are as follows:
To enhance the model’s adaptability in complex meteorological scenarios, the PV dataset is first processed using a K-medoids clustering algorithm based on Dynamic Time Warping (DTW) distance. This method allows the model to fully account for the impact of varying weather conditions on power generation.
The TimeGAN algorithm is employed to generate synthetic data samples under extreme weather conditions, aiming to mitigate model underfitting and enhance the model’s robustness in such scenarios.
The BiTCN and BiGRU algorithms are integrated to fully leverage their bidirectional structures for capturing latent information within the sequences. Meanwhile, to address the issue of excessive hyperparameters, the AOO algorithm is employed to optimize these parameters, thereby improving prediction accuracy and reducing manual tuning efforts.
The superior performance of the proposed method for photovoltaic power forecasting is validated through a comparative analysis against other models, utilizing Mean Absolute Error (MAE), Root Mean Square Error (RMSE), and R2 as the evaluation indicators.
The remainder of this article is organized as follows: Section 2 covers the theoretical foundations of the employed methods and the proposed model. Section 3 is then dedicated to the case study, data, and preprocessing. Section 4 describes the modeling process and presents a comprehensive analysis of the results. Lastly, Section 5 recaps the principal findings of this study and proposes potential avenues for subsequent research.

2. Methodology

The proposed AOO-BiTCN-BiGRU photovoltaic power forecasting framework centers on four core components: weather classification, sample augmentation, model optimization, and predictive modeling. Each component has a distinct role: (1) K-medoids clustering classifies similar days, enhancing the model’s adaptability to diverse weather patterns; (2) TimeGAN addresses sparse extreme weather samples by generating high-fidelity synthetic data to augment the training set; (3) The AOO algorithm optimizes hyperparameters of the BiTCN-BiGRU model to enhance performance; (4) BiTCN-BiGRU serves as the core prediction network, simultaneously capturing both local temporal features and long-term dependencies in PV power sequences. Through synergistic collaboration, these components ultimately achieve high-accuracy multi-step power forecasting.

2.1. DTW-K-Medoids Clustering Algorithm

This study utilizes a K-medoids clustering algorithm with the DTW distance metric for the efficient extraction of similar days [34]. The K-medoids algorithm addresses the limitations of the K-means algorithm in this research scenario, including its weak robustness to outliers and anomalies, distance distortion, and low interpretability of cluster centers. This approach is intended to fully categorize data samples based on power characteristics and to improve the model’s generalization and structural adaptability for PV power data. The main steps include data construction, DTW distance matrix calculation, and K-medoids clustering, which are outlined as follows:

2.1.1. Data Construction and Representation

The sampling time of the photovoltaic power time series is denoted as T, and the data includes historical records of N days. Each day is treated as a sample, and the sample matrix is constructed as follows:
P i = [ P i , 1 , P i , 2 , , P i , T ] T R T ,           i = 1 , , N
In this context, Pi denotes the photovoltaic power value at each sampling point on the i-th day.
P = [ P 1 T P 2 T P N T ] = [ P 1 , 1 P 1 , 2 P 1 , T P 2 , 1 P 2 , 2 P 2 , T P N , 1 P N , 2 P N , T ] R N × T

2.1.2. Definition of DTW Distance

To accurately measure the similarity between power curves on different days, Dynamic Time Warping (DTW) is introduced as a distance metric between samples. The DTW distance between any two daily power sequences, x1 and y1, is defined as:
x = [ x 1 , x 2 , , x T ] ,             y = [ y 1 , y 2 , , y T ]
D D T W ( x , y ) = min π ( t x , t y ) π ( x t x y t y ) 2                   1 t x T , 1 t y T
where π represents the set of paired time steps that satisfy the boundary constraints, monotonicity, and continuity conditions; t x and t y are the time indices of sequences and aligned on the warping path.
By calculating the DTW distance between all pairs of P i and P j , a distance matrix is constructed:
D = [ D D T W ( P 1 , P 1 ) D D T W ( P 1 , P 2 ) D D T W ( P 1 , P N ) D D T W ( P 2 , P 1 ) D D T W ( P 2 , P 2 ) D D T W ( P 2 , P N ) D D T W ( P N , P 1 ) D D T W ( P N , P 2 ) D D T W ( P N , P N ) ]       D i j = D D T W ( P i , P j )

2.1.3. K-Medoids Clustering

Given the number of clusters k, let S = { 1 , , N } denote the set of sample indices and C = { C 1 , , C k } S the indices of the medoids to be optimized. K-Medoids minimizes the sum of within-cluster DTW distances:
min C r = 1 k i S r D D T W ( P r , P C r )
where S r = { i S | r = a r g min 1 l k D D T W ( P i , P C l ) } represents the cluster for which P C r serves as the medoid.
By iterating through different values of k, all historical daily samples are eventually categorized into 4 similar clusters, which are then used for subsequent modeling and prediction based on the categories.

2.2. TimeGAN

To address the scarcity of extreme weather data samples, this study employs TimeGAN, a framework introduced by Yoon et al. [35] that uses GANs for time-series generation. The model uniquely blends supervised and unsupervised learning, using embedding and recovery mechanisms alongside supervised losses. This approach allows it to learn the dynamic evolution and temporal consistency of data, not just static distributions, enabling the generation of high-quality, structurally similar sequences to augment the dataset.
As illustrated in Figure 2, TimeGAN’s architecture is constructed from four primary modules: an embedding network, a reconstruction network, a generator, and a discriminator. Operating as an autoencoder, the embedding network is responsible for transforming the high-dimensional input data into a low-dimensional latent representation, a process that inherently extracts the key features. The reconstruction network maps latent representations back to the original dimensions, optimized to minimize reconstruction error. Building upon this foundation, the model introduces a GAN architecture within the embedding space to jointly train the generator and discriminator, thereby capturing temporal dependencies within high-dimensional time series. This approach not only reduces the difficulty of adversarial training but also significantly enhances the model’s learning efficiency and generation quality.

2.3. AOO-BiTCN-BiGRU Network

TCNs are an enhanced CNN architecture well-suited for temporal problems, primarily due to their unique structure employing extended causal convolutions and residual blocks. As shown in Figure 3, the extension factor d grows exponentially with depth. This design allows the convolutions to expand their receptive field, sample inputs from preceding layers, and thereby extract features from time series data [36]. Residual connections directly link inputs to outputs, ensuring input information propagates to subsequent layers without excessive transformation, thereby mitigating information loss and vanishing gradient issues. This approach performs forward convolutional computations only on the input sequence, extracting forward data features while neglecting implicit information in the reverse direction. Hence, this study adopts the BiTCN network. The BiTCN network, illustrated in Figure 4, employs a bidirectional structure with a larger receptive field. The introduction of residual blocks mitigates issues like gradient vanishing and slow convergence associated with increased receptive fields. Consequently, the bidirectional temporal convolution architecture of BiTCN more effectively captures hidden features in both forward and backward directions, thereby better capturing long-term dependencies within sequences.
The BiGRU layer is employed to further process the output of the BiTCN, enhancing prediction accuracy by leveraging the precision of contextual information captured from both preceding and subsequent data. Figure 5A illustrates the BiGRU network. BiGRU is a dual-layer GRU network. The forward GRU layer processes features through forward propagation. This mines forward-related data correlations. The backward GRU layer trains the input sequence via backward propagation. This mines backward-related data correlations. This network architecture enables bidirectional feature extraction from inputs, better enhancing feature completeness and globality.
Figure 5B shows the structure of the GRU. The GRU receives two inputs. These are the hidden state h t 1 from the previous time step and the input data X t from the current time step. Its output is the hidden state h t at the current time step. To reduce redundant data information during activation, the GRU performs preliminary parameter estimation before activation, thereby eliminating unnecessary computations. It also utilizes the mechanisms of the reset gate r t and the update gate Z t to update the state [37]. The core function of the reset gate lies in regulating the degree of forgetting historical state information. As the value of r t decreases, the amount of neglected information increases, expressed as follows:
r t = σ ( W r [ h t 1 , X t ] + b r )
The primary function of the update gate is to monitor the extent to which historical state information propagates to the current time point. As the Z t value increases, the amount of information transmitted from the previous moment to the present state correspondingly grows, as expressed by the following formula:
Z t = σ ( W z [ h t 1 , X t ] + b z )
Candidate latent states:
h ˜ t = tan h ( W [ r t h t 1 , X t ] + b )
Implicit State at the Current Moment:
h t = ( 1 Z t ) h t 1 + Z t h ˜ t
In these formulas, W r and b r are the reset gate weight and bias. W z and b z are the update gate weight and bias. W and b are the recurrent connection weight and bias.
The Animated Oat Optimization (AOO) algorithm is a novel intelligent optimization technique. Unlike previous algorithms, it draws inspiration from the three-stage motion mechanism of oat seeds propagating through “rolling-bouncing” in humid environments [38]. This mechanism ingeniously integrates global exploration with local exploitation within a two-stage framework. Figure 6 explains the algorithm’s underlying principles.

2.3.1. Initialization Phase

Before commencing the main strategy, the initial positions of the population must first be established:
X = [ x 1 , 1 x 1 , j x 1 , D i m 1 x 1 , D i m x 2 , 1 x 2 , j x 2 , D i m 1 x 2 , D i m x i , j x N 1 , 1 x N 1 , j x N 1 , D i m 1 x N 1 , D i m x N , 1 x N , j x N , D i m 1 x N , D i m ]
where x i denotes an individual in the i-th subgroup along with its location. D i m indicates the dimensional size of the problem. N represents the total number of individuals in the subpopulation.
x i , j = r × ( U B j L B j ) + L B j , i = 1 , 2 , , N , j = 1 , 2 , , D i m
where r represents a random value chosen from the range [0, 1]. The terms U B j and L B j define the problem’s boundaries for the j-th dimension. U B j is the upper limit. L B j is the lower limit.

2.3.2. Biological Parameter Mapping

The developed mathematical model quantifies the animated oat seed’s dispersal process. Several key biological characteristics are defined. These include the length and mass of the seed’s main awn. The eccentricity coefficient from the rolling process is also incorporated. The equations below are used to calculate these values:
m = 0.5 × r dim L = N × r dim e = 0.5 × r dim c = 1 t T 3
where m , L , and e define the seed’s physical properties (mass, awn length, and eccentricity coefficient). t , T , and c are the iteration control parameters (current iteration, max iterations, and dynamic adjustment factor).

2.3.3. Exploration Phase

To facilitate a wide-ranging exploration of the solution space, this random dispersal method governs position updates according to the following equations:
w = c π × ( 2 × r d i m 1 ) U B
{ X t + 1 ( i ) = 1 N × i = 1 N X t ( i ) + W , i f   m o d   ( i , N / 10 ) = 0 , X t + 1 ( i ) = X b e s t + W , i f   m o d   ( i , N / 10 ) = 1 , X t + 1 ( i ) = X t ( i ) + W , e l s e .
where U B defines the problem’s maximum boundary. X t + 1 ( i ) is the updated location for the i-th individual. X b e s t tracks the best solution found so far in the population.

2.3.4. Development Stage

This stage implements two propagation methods based on obstacle encounters, each with an equal probability. The ‘no obstacle’ scenario, termed ‘hygroscopic rolling,’ is mathematically defined by the following eccentric rotation and torque equations:
A = U B | U B × t × sin ( 2 × π × r ) T |
R = ( m × e + L 2 ) × r d i m ( A , A ) d i m
Levy ( dim ) = 0.01 × μ × σ | ν | 1 / β
σ = ( Γ ( 1 + β ) × sin ( π × β 2 ) Γ ( 1 + β 2 ) × β × 2 ( β 1 2 ) ) 1 β
X t ( i ) = X b e s t + R + c × L e v y ( d i m ) X b e s t
where r d i m ( A , A ) is a random matrix whose dimensions match the target problem. The Levy flight is controlled by several standard parameters. μ is the mean value used to adjust the step size. σ is the scale parameter for the step length distribution. ν is the current velocity vector. β is the stable distribution parameter which shapes the step’s randomness. Finally, Γ is the standard gamma function.
In the event of an obstacle, the seed initiates an energy-driven ejection. This entire process is simulated as simplified projectile motion, with the position update defined as follows:
B = U B | U B × t × cos ( 2 × π × r ) T |
{ k = 0.5 + 0.5 × r x = 3 × r d i m θ = π × r α = 1 π × e r T
J = 2 × k × x 2 × sin ( 2 θ ) m g × r d i m ( B , B ) d i m × ( 1 α )
X t ( i ) = X b e s t + J + c × L e v y ( d i m ) X b e s t
where k , x , θ , and α represent the physical parameters for the projectile model (elasticity, length change, angle, and air resistance). r is a random number between 0 and T .

3. Data Analysis and Parameter Configuration

This section outlines the dataset employed in this study, the comparative methodology, evaluation metrics, parameter configurations, and feature relevance analysis. Furthermore, the experimental hardware configuration comprises a 3.7 GHz Intel® CoreTM i9-10900K CPU, 96.00 GB of memory, and an NVIDIA® GeForce RTX™ 5060 Ti graphics card. The TimeGAN algorithm was implemented in Python 3.10.15based on the PyTorch 2.7.0 framework with CUDA 12.8 support, whilst clustering and prediction algorithms were realized using MATLAB 2023a.

3.1. Data Information

The data for this study were sourced from figshare, a website that publicly shares various datasets [39]. The solar data was sourced from the China State Grid Renewable Energy Generation Forecasting Competition. This dataset originates from a 30 MW nominal capacity solar power station. Table 1 and Table 2 illustrates its key characteristics. The dataset spans the period from 1 January 2019 to 31 December 2019, with a resolution of 15 min. Detailed feature information is presented in Table 1. Energy peaks around midday, with relatively lower levels at sunrise and in the afternoon, tending towards zero during night-time hours. Furthermore, a noticeable lag effect exists between certain meteorological factors and the PV data. As photovoltaic power output is zero at night, night-time data points were removed, retaining only data from 07:30 to 19:30. The dataset was subsequently partitioned into an 80% training set and a 20% validation set.

3.2. Comparative Methodology

To demonstrate the effectiveness of generated weather samples for photovoltaic power forecasting, this paper designed an ablation study comparing accuracy changes before and after data generation. Concurrently, to validate the efficacy of the AOO-BiTCN-BiGRU model for this task, we selected GRU, LSTM, TCN [40], and CNN models for comparative analysis.

3.3. Evaluation Metrics

To assess how well the predicted values matched the actual values, this study utilized three standard statistical metrics: MAE, RMSE, and the coefficient of determination (R2) [41,42,43].
The MAE quantifies the average magnitude of the errors between predicted and actual values:
M A E = 1 n i = 1 n ( Y i Y ^ I )
The RMSE is particularly sensitive to outliers. This is because its calculation squares each error, which gives large errors a disproportionately heavy weight:
R M S E = 1 n i = 1 n ( Y i Y ^ I ) 2
For the error metrics (MAE and RMSE), lower values signify superior predictive accuracy. Conversely, R2 measures the quality of the model’s fit, where a value closer to 1 indicates a more accurate prediction:
R 2 = 1 i = 1 n ( Y i Y ^ I ) 2 i = 1 n ( Y i Y ¯ I ) 2
where Y i , Y ^ I , and   Y ¯ I correspond to the actual, predicted, and mean values, respectively, and n is the total number of samples.

3.4. Parameter Configuration

For the DTW-based K-medoids algorithm, the sequence length was set to 49 points. Based on this data structure, the number of cluster categories was set to 4.
For the TimeGAN algorithm, the detailed parameter settings are presented in Table 3. Within this table, the middle column corresponds to the search range for each parameter, while the rightmost column indicates the final parameter values determined through experimental validation. It is worth noting that the algorithm generates a volume of synthetic samples consistent with the original dataset. To ensure comprehensive exploration of the solution space while balancing efficiency, the maximum number of iterations for the algorithm starts at 3000 and increments by 500 steps each time, ranging up to 7000.
Regarding parameter settings for the AOO-BiTCN-BiGRU model: the historical time step size is set to 49, the prediction step size to 4 (i.e., utilizing one day’s historical data to forecast the subsequent hour’s data); the filter size is set to 5; and the optimizer employs the Adam algorithm.
Dropout, a commonly used regularization technique in neural network training, primarily aims to suppress model overfitting. The selection of the Dropout rate necessitates consideration of the specific application scenario and model architecture, typically requiring cross-validation to determine the optimal value. This paper sets the Dropout rate at 0.1, a value that effectively avoids the reduced learning efficiency or underfitting issues caused by excessively high Dropout rates, while also preventing overfitting resulting from excessively low Dropout rates.
The parameter settings for the AOO optimization algorithm are as follows: population size of 4 and maximum iterations of 10. This algorithm primarily optimizes key parameters within the BiTCN-BiGRU model, specifically including the learning rate, number of neurons in the BiGRU layer, number of filters, and regularization parameters. The upper and lower limits for optimizing each parameter are detailed in Table 4.
After AOO optimization, the final hyperparameter values used in the experiment are summarized in Table 5. These values are derived through iterative search and verification, ensuring they match the model structure and data characteristics to maximize predictive performance.

3.5. Feature Correlation Analysis

In this study, we employed the Pearson correlation coefficient method (PCC) [44] to analyze the degree of association between different variables. As a multifunctional analysis technique, PCC is calculated based on time-series data, focusing on linear correlations between data sequences. In the PCC calculation, we selected weather elements as the comparison series and PV output power as the baseline for comparison. The PCC calculation formula is as follows:
ρ x , y = c o v ( X , Y ) σ x σ y = E ( X μ X ) E ( Y μ Y ) σ X σ Y
The PCC method was applied to calculate the pairwise correlations between seven meteorological factors and power in the original dataset, including total solar radiation (W/m2), direct normal radiation (W/m2), global horizontal radiation (W/m2), air temperature (°C), atmospheric pressure (hPa), relative humidity (%), and PV output power (MW). The PCC heat map is presented in Figure 7. Upon calculating the correlation coefficients, the degree of association between variables can be assessed via the value ranges. The heat map indicates that the factors exerting the greatest influence and exhibiting the highest correlation with PV output power are total solar radiation, direct normal radiation, and global horizontal radiation. Atmospheric pressure exerts the least influence and is therefore excluded. Consequently, the selected input variables for the model comprise total solar radiation, direct normal radiation, global horizontal radiation, air temperature, relative humidity, and PV output power.

4. Results and Discussion

The PV data exhibits strong periodicity. After excluding data points where nighttime power is zero, the remaining dataset contains a total of 49 time points per day (7:30–19:30). The model’s performance was validated on a specific task: predicting the subsequent 4 time steps (1 h) using data from the past 49 time steps (one day). A multi-output strategy was employed for this multi-step forecasting. Finally, comprehensive comparisons were conducted to analyze the model’s performance under varying conditions. This approach better captures dependencies between inputs and outputs, enabling simultaneous prediction of sequences across multiple time steps during model training. The experimental design comprises three parts: (i) K-medoids clustering algorithm evaluation, (ii) ablation experiments and (iii) comparative experiments.

4.1. Evaluation of the K-Medoids Clustering Algorithm

The clustering results, as shown in Figure 8, reveal that periods of stable irradiance (e.g., noon) form tighter clusters, while transitional periods (e.g., sunrise/sunset) exhibit higher variability. Clear-sky days yield the highest photovoltaic power data, whereas overcast days with insufficient light intensity produce relatively lower power, consistent with the physical dynamic characteristics of photovoltaic power generation.
To evaluate the effectiveness of the K-medoids clustering algorithm in capturing intrinsic patterns within photovoltaic power generation data simultaneously, we first visualized the high-dimensional data structure through multidimensional scaling (MDS) projection. MDS reduces feature dimensions while preserving pairwise distances between data points, enabling a qualitative assessment of cluster separability. As shown in Figure 9, the projected data exhibits a clear clustering structure. The boundaries between the four clusters are distinct, with only a few points located on either side of the boundaries. This indicates that the K-medoids algorithm effectively groups photovoltaic power sequences with similar temporal characteristics.

4.2. Ablation Experiments

To validate the efficacy of TimeGAN in augmenting extreme weather samples, we designed an ablation study. Initially, the model was trained using only the original 30 days of rainy-day samples. Following this, performance was assessed on the 17 rainy-day samples within the test set. Subsequently, 14 days of the synthetic data generated by TimeGAN were integrated with the original 30 days of samples, creating an augmented training set of 44 days. The original sequences and generated sequences are visualized in Figure 10. The predictive accuracy on the same 17-day test set was then re-evaluated using this combined dataset. The results (Table 6) confirm that augmenting the data significantly boosts predictive performance: MAE dropped by 39.58%, RMSE decreased by 39.10%, and R2 saw an increase of 0.1386. This clearly demonstrates that sample generation techniques enhance both model accuracy and generalization.

4.3. Comparative Experiments

In contrast to single-step methods that predict a singular future value, this work uses multi-step forecasting. This technique leverages past PV power and meteorological conditions to generate a sequence of predictions over a subsequent time horizon.
The detailed comparison results, summarized in Table 7 and Figure 11, indicate that the AOO-BiTCN-BiGRU model also achieved the highest accuracy in multi-step forecasting across four weather conditions. Results indicate that this model attained the highest average R2 among the five models at 0.9367, while its MAE and RMSE values of 0.9216 and 1.2030, respectively, were the lowest among all models. Compared to other models, this model’s MAE and RMSE were 9.27% and 8.84% lower than the GRU model, 12.52% and 12.9% lower than the LSTM model, 15.09% and 16.46% lower than the TCN model, and 19.36% and 14.16% lower than the CNN model. Additionally, the AOO-BiTCN-BiGRU model achieved higher R2 values than other models by 0.0166, 0.0252, 0.0518, and 0.0178, respectively. While some weather conditions showed slightly higher performance in comparison models, this discrepancy is attributed to the more complex bidirectional structure and larger parameter set of our model, with a large number of parameters. Consequently, its performance in predicting weather conditions with relatively smooth power fluctuations may not significantly outperform other models. However, the average accuracy of this model far exceeds that of the comparison models, indicating that while it demonstrates strong predictive capabilities under weather conditions with smooth power fluctuations, it also possesses robust representational power for weather conditions with intense fluctuations. These results demonstrate that the AOO-BiTCN-BiGRU model possesses strong generalization capabilities and significantly improves prediction accuracy.
The AOO-BiTCN-BiGRU model exhibits high reliability in multi-step photovoltaic power forecasting. This is visually confirmed by Figure 12, which depicts the alignment between the actual and forecasted power values throughout the prediction. The results indicate that this model effectively captures long-term dependencies in PV data, performing particularly well during rainy days with high volatility, where the curves exhibit fewer discrete values. Even when forecast values deviate from actual values, the model converges to the actual values faster and more accurately. The predictions from the AOO-BiTCN-BiGRU model are therefore shown to be balanced, lacking significant overestimation or underestimation biases. Regarding computational efficiency concerns, all models were trained on the unified hardware platform described above. Given that the structural complexity of photovoltaic time-series data is lower than that of multidimensional tensors in image data, the training duration of the proposed AOO-BiTCN-BiGRU model falls within a reasonable range: In single-category weather scenarios, sunny-day training took the longest, approximately 50 min (including hyperparameter optimization and model training), with the pure model training phase taking less than 2 min. The comparison model, featuring a simpler network architecture and requiring no additional hyperparameter tuning, completed sunny-day training within 30 s. Notably, during actual deployment, the model rapidly outputs predictions by loading pre-trained parameters alone, fully meeting the real-time requirements of power grid dispatch.

5. Conclusions

Accurately predicting PV power on an ultra-short timescale is vital for maintaining the stable operation of the power grid and executing peak shaving and valley filling operations. Such forecasting is critical for managing the equilibrium between supply and load, as well as for optimizing energy dispatch. While single-structure models are often challenged by highly volatile data and long-range temporal relationships, composite-structure models generally yield superior accuracy. The latter achieves this by utilizing more intricate neural network architectures capable of learning complex features and data patterns. To address this, our study enhances ultra-short-term PV forecasting accuracy by constructing a hybrid prediction model. This model integrates clustering, optimization, generative adversarial networks, and deep learning methods, thereby achieving high-precision forecasting and robust performance evaluation.
The main conclusions drawn from this work include:
A hybrid PV power forecasting model was developed that fully accounts for the influence of varying weather conditions on PV output. To significantly boost the model’s accuracy and adaptability across complex meteorological conditions, multi-dimensional meteorological data (such as temperature, humidity, wind speed, and solar irradiance) along with historical PV power data were incorporated.
Data generation via the TimeGAN algorithm mitigates model underfitting, thereby enhancing robustness under extreme weather conditions. Exploration of model parameter optimization methods reduces complexity and computational efficiency, ensuring practical feasibility for real-world forecasting.
Validation and evaluation using operational data from actual PV power plants analyzed prediction accuracy and reliability across different seasons and weather conditions. Comparative analysis assessed the hybrid model’s advantages over single-method approaches, providing a more reliable tool for PV power forecasting in photovoltaic-storage integrated charging stations.
In summary, the integrated PV power forecasting algorithm proposed herein achieves a higher level of accuracy compared to the other models. Its consideration of weather factors and generation of extreme weather samples significantly enhances PV power prediction accuracy, holding substantial practical value and real-world significance. Future research may deepen the exploration of the following areas: firstly, expanding data dimensions to incorporate PV module status (real-time temperature, dust coverage, degradation level) and regional microclimate data to reduce prediction errors; secondly, integrating reinforcement learning to optimize model parameter update mechanisms, enabling online adaptive optimization and enhancing long-term operational stability; thirdly, integrating the model with energy management strategies for photovoltaic-storage systems to construct an integrated ‘prediction-scheduling’ decision model, thereby facilitating renewable energy integration.

Author Contributions

Conceptualization, Z.S., D.L. and Q.Y.; Methodology, Y.J. and Q.Y.; Software, Y.J. and Q.Y.; Validation, Y.J., W.G. and F.L.; Formal analysis, Y.J. and Q.Y.; Investigation, Y.J. and D.L.; Resources, Z.S., W.G. and F.L.; Data curation, Y.J.; Writing—original draft, Y.J.; Writing—review & editing, Y.J. and Q.Y.; Visualization, Y.J.; Supervision, Z.S., D.L. and Q.Y.; Project administration, W.G.; Funding acquisition, Z.S., D.L., W.G. and F.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Project of the Shanxi Expressway Group Co., Ltd. (Grant No. GSKJ-2024-04).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

Authors Zhichao Sun, Dongliang Lv, Weicheng Gao and Fengze Liu were employed by Shanxi Transportation Investment and Financing Group Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest. The authors declare that this study received funding from the Shanxi Expressway Group Co., Ltd. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.

References

  1. Zhong, J.Q.; Liu, L.Y.; Sun, Q.; Wang, X.Y. Prediction of Photovoltaic Power Generation Based on General Regression and Back Propagation Neural Network. Energy Procedia 2018, 152, 1224–1229. [Google Scholar] [CrossRef]
  2. Zhi, Y.; Sun, T.; Yang, X.D. A physical model with meteorological forecasting for hourly rooftop photovoltaic power prediction. J. Build. Eng. 2023, 75, 106997. [Google Scholar] [CrossRef]
  3. El Ainaoui, K.; Zaimi, M.; Flouchi, I.; Elhamaoui, S.; El mrabet, Y.; Ibaararen, K.; Bouasria, Y.; Ghennioui, A.; Assaid, E.M. Novel optimized models to enhance performance forecasting of grid-connected PERC PV string operating under semi-arid climate conditions. Sol. Energy 2024, 282, 112976. [Google Scholar] [CrossRef]
  4. Zaimi, M.; El Achouby, H.; Ibral, A.; Assaid, E.M. Determining combined effects of solar radiation and panel junction temperature on all model-parameters to forecast peak power and photovoltaic yield of solar panel under non-standard conditions. Sol. Energy 2019, 191, 341–359. [Google Scholar] [CrossRef]
  5. Tifidat, K.; Maouhoub, N.; Askar, S.S.; Abouhawwash, M. Numerical procedure for accurate simulation of photovoltaic modules performance based on the identification of the single-diode model parameters. Energy Rep. 2023, 9, 5532–5544. [Google Scholar] [CrossRef]
  6. Li, Y.Y.; Song, L.D.; Zhang, S.; Kraus, L.; Adcox, T.; Willardson, R.; Komandur, A.; Lu, N. A TCN-Based Hybrid Forecasting Framework for Hours-Ahead Utility-Scale PV Forecasting. IEEE Trans. Smart Grid 2023, 14, 4073–4085. [Google Scholar] [CrossRef]
  7. Pylorof, D.; Garcia, H.E. Uncertainty-aware photovoltaic generation estimation through fusion of physics with harmonics information using Bayesian neural networks. In Proceedings of the 2023 IEEE Power & Energy Society Innovative Smart Grid Technologies Conference (ISGT), Washington, DC, USA, 16–19 January 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–5. [Google Scholar] [CrossRef]
  8. Koster, D.; Fiorelli, D.; Bruneau, P.; Braun, C. Single-Site Forecasts for 130 Photovoltaic Systems at Distribution System Operator Level, Using a Hybrid-Physical Approach, to Improve Grid-Integration and Enable Future Smart-Grid Operation. Sol. RRL 2023, 7, 2200652. [Google Scholar] [CrossRef]
  9. Kilian, L.; Lütkepohl, H. Structural Vector Autoregressive Analysis; Cambridge University Press: Cambridge, UK, 2017. [Google Scholar]
  10. Box, G.E.P.; Jenkins, G.M. Time Series Analysis: Forecasting and Control; Holden-Day: San Francisco, CA, USA, 1976. [Google Scholar]
  11. Li, Y.T.; Su, Y.; Shu, L.J. An ARMAX model for forecasting the power output of a grid connected photovoltaic system. Renew. Energy 2014, 66, 78–89. [Google Scholar] [CrossRef]
  12. Cao, Y.S.; Liu, G.; Luo, D.H.; Bavirisetti, D.P.; Xiao, G. Multi-timescale photovoltaic power forecasting using an improved Stacking ensemble algorithm based LSTM-Informer model. Energy 2023, 283, 128669. [Google Scholar] [CrossRef]
  13. Sulaiman, M.H.; Jadin, M.S.; Mustaffa, Z.; Daniyal, H.; Azlan, M.N.M. Short-term forecasting of rooftop retrofitted photovoltaic power generation using machine learning. J. Build. Eng. 2024, 94, 109948. [Google Scholar] [CrossRef]
  14. Eseye, A.T.; Zhang, J.H.; Zheng, D.H. Short-term photovoltaic solar power forecasting using a hybrid Wavelet-PSO-SVM model based on SCADA and Meteorological information. Renew. Energy 2018, 118, 357–367. [Google Scholar] [CrossRef]
  15. Wolff, B.; Kühnert, J.; Lorenz, E.; Kramer, O.; Heinemann, D. Comparing support vector regression for PV power forecasting to a physical modeling approach using measurement, numerical weather prediction, and cloud motion data. Sol. Energy 2016, 135, 197–208. [Google Scholar] [CrossRef]
  16. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  17. Lee, D.; Kim, K. PV power prediction in a peak zone using recurrent neural networks in the absence of future meteorological information. Renew. Energy 2021, 173, 1098–1110. [Google Scholar] [CrossRef]
  18. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  19. Hu, Z.H.; Gao, Y.; Ji, S.Y.; Mae, M.; Imaizumi, T. Improved multistep ahead photovoltaic power prediction model based on LSTM and self-attention with weather forecast data. Appl. Energy 2024, 359, 122709. [Google Scholar] [CrossRef]
  20. Ahmed, R.; Sreeram, V.; Togneri, R.; Datta, A.; Arif, M.D. Computationally expedient Photovoltaic power Forecasting: A LSTM ensemble method augmented with adaptive weighting and data segmentation technique. Energy Conv. Manag. 2022, 258, 115563. [Google Scholar] [CrossRef]
  21. Chung, J.; Gülçehre, Ç.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar] [CrossRef]
  22. Kim, Y. Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1746–1751. [Google Scholar] [CrossRef]
  23. Dai, Y.; Yu, W.; Leng, M. A hybrid ensemble optimized BiGRU method for short-term photovoltaic generation forecasting. Energy 2024, 299, 131458. [Google Scholar] [CrossRef]
  24. Chen, Y.J.; Xiao, J.W.; Wang, Y.W.; Li, Y.Z. Regional wind-photovoltaic combined power generation forecasting based on a novel multi-task learning framework and TPA-LSTM. Energy Convers. Manag. 2023, 297, 117715. [Google Scholar] [CrossRef]
  25. Qu, J.Q.; Qian, Z.; Pei, Y. Day-ahead hourly photovoltaic power forecasting using attention-based CNN-LSTM neural network embedded with multiple relevant and target variables prediction pattern. Energy 2021, 232, 120996. [Google Scholar] [CrossRef]
  26. Huang, X.Q.; Li, Q.; Tai, Y.H.; Chen, Z.Q.; Liu, J.; Shi, J.S.; Liu, W.M. Time series forecasting for hourly photovoltaic power using conditional generative adversarial network and Bi-LSTM. Energy 2022, 246, 123403. [Google Scholar] [CrossRef]
  27. Zhen, H.; Niu, D.X.; Wang, K.K.; Shi, Y.C.; Ji, Z.S.; Xu, X.M. Photovoltaic power forecasting based on GA improved Bi-LSTM in microgrid without meteorological information. Energy 2021, 231, 120908. [Google Scholar] [CrossRef]
  28. Yang, K.; Cai, Y.J.; Cheng, J.R. A deep learning model based on multi-attention mechanism and gated recurrent unit network for photovoltaic power forecasting. Comput. Electr. Eng. 2025, 123, 110250. [Google Scholar] [CrossRef]
  29. Li, Q.; Zhang, X.Y.; Ma, T.J.; Liu, D.G.; Wang, H.; Hu, W. A Multi-step ahead photovoltaic power forecasting model based on TimeGAN, Soft DTW-based K-medoids clustering, and a CNN-GRU hybrid neural network. Energy Rep. 2022, 8, 10346–10362. [Google Scholar] [CrossRef]
  30. Dai, Y.M.; Wang, Y.X.; Leng, M.M.; Yang, X.Y.; Zhou, Q. LOWESS smoothing and Random Forest based GRU model: A short-term photovoltaic power generation forecasting method. Energy 2022, 256, 124661. [Google Scholar] [CrossRef]
  31. Sushmi, N.B.; Subbulekshmi, D. Real-time ultra short-term irradiance forecasting using a novel R-GRU model for optimizing PV controller dynamics. Results Eng. 2025, 26, 105046. [Google Scholar] [CrossRef]
  32. Guo, X.F.; Zhan, Y.; Zheng, D.; Liao, L.Y.; Qi, Q. Research on short-term forecasting method of photovoltaic power generation based on clustering SO-GRU method. Energy Rep. 2023, 9, 786–793. [Google Scholar] [CrossRef]
  33. Elmousaid, R.; Drioui, N.; Elgouri, R.; Agueny, H.; Adnani, Y. Ultra-short-term global horizontal irradiance forecasting based on a novel and hybrid GRU-TCN model. Results Eng. 2024, 23, 102817. [Google Scholar] [CrossRef]
  34. Pan, Y.; Wu, M.Z.; Zhang, L.M.; Chen, J.J. Time series clustering-enabled geological condition perception in tunnel boring machine excavation. Autom. Constr. 2023, 153, 104954. [Google Scholar] [CrossRef]
  35. Yoon, J.; Jarrett, D.; van der Schaar, M. Time-series Generative Adversarial Networks. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  36. Limouni, T.; Yaagoubi, R.; Bouziane, K.; Guissi, K.; Baali, E. Accurate one step and multistep forecasting of very short-term PV power using LSTM-TCN model. Renew. Energy 2023, 205, 1010–1024. [Google Scholar] [CrossRef]
  37. Hosseini, S.M.; Ebrahimi, A.; Mosavi, M.R.; Shahhoseini, H.S. A novel hybrid CNN-CBAM-GRU method for intrusion detection in modern networks. Results Eng. 2025, 28, 107103. [Google Scholar] [CrossRef]
  38. Wang, R.B.; Pan, J.S.; Hu, R.B.; Geng, F.D.; Xu, L.; Chu, S.C.; Meng, Z.Y.; Mirjalili, S. The Animated Oat Optimization Algorithm: A nature-inspired metaheuristic for engineering optimization and a case study on Wireless Sensor Networks. Knowl. Based Syst. 2025, 318, 113589. [Google Scholar] [CrossRef]
  39. Chen, Y.; Xu, J. Solar and wind power data from the Chinese State Grid Renewable Energy Generation Forecasting Competition. Sci. Data 2022, 9, 577. [Google Scholar] [CrossRef] [PubMed]
  40. Bai, S.; Kolter, J.Z.; Koltun, V. An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar] [CrossRef]
  41. Mayer, M.J. Benefits of physical and machine learning hybridization for photovoltaic power forecasting. Renew. Sust. Energ. Rev. 2022, 168, 112772. [Google Scholar] [CrossRef]
  42. Mayer, M.J.; Gróf, G. Extensive comparison of physical models for photovoltaic power forecasting. Appl. Energy 2021, 283, 116239. [Google Scholar] [CrossRef]
  43. Liu, X.; Liu, Y.; Kong, X.; Ma, L.; Besheer, A.H.; Lee, K.Y. Deep neural network for forecasting of photovoltaic power based on wavelet packet decomposition with similar day analysis. Energy 2023, 271, 126963. [Google Scholar] [CrossRef]
  44. Huang, L.H.; Hu, B.; Wan, S.T.; Lu, B. Research on pipeline flange leakage detection method based on random forest and Pearson correlation coefficient. Appl. Acoust. 2025, 240, 110918. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed forecast model.
Figure 1. Flowchart of the proposed forecast model.
Energies 19 00029 g001
Figure 2. Schematic Diagram of TimeGAN.
Figure 2. Schematic Diagram of TimeGAN.
Energies 19 00029 g002
Figure 3. Dilated causal convolution.
Figure 3. Dilated causal convolution.
Energies 19 00029 g003
Figure 4. Schematic diagram of BiTCN.
Figure 4. Schematic diagram of BiTCN.
Energies 19 00029 g004
Figure 5. Schematic diagram of BiGRU.
Figure 5. Schematic diagram of BiGRU.
Energies 19 00029 g005
Figure 6. Schematic diagram of the AOO Algorithm.
Figure 6. Schematic diagram of the AOO Algorithm.
Energies 19 00029 g006
Figure 7. PCC heat map.
Figure 7. PCC heat map.
Energies 19 00029 g007
Figure 8. Clustering results obtained for the four weather categories, namely rainy (weather 1), cloudy (weather 2), sunny (weather 3) and intermittent (weather 4).
Figure 8. Clustering results obtained for the four weather categories, namely rainy (weather 1), cloudy (weather 2), sunny (weather 3) and intermittent (weather 4).
Energies 19 00029 g008
Figure 9. MDS projection diagram.
Figure 9. MDS projection diagram.
Energies 19 00029 g009
Figure 10. Visualization of Original and Generated Sequences.
Figure 10. Visualization of Original and Generated Sequences.
Energies 19 00029 g010
Figure 11. Comparison of proposed model (M1) against baseline architectures: GRU (M2), LSTM (M3), TCN (M4) and CNN (M5).
Figure 11. Comparison of proposed model (M1) against baseline architectures: GRU (M2), LSTM (M3), TCN (M4) and CNN (M5).
Energies 19 00029 g011
Figure 12. Predicted results for four weather types.
Figure 12. Predicted results for four weather types.
Energies 19 00029 g012
Table 1. Statistical overview of meteorological and PV power data.
Table 1. Statistical overview of meteorological and PV power data.
NameUnitMinMeanMaxStd
Total Solar IrradianceW/m20163.241214.54245.39
Direct Normal IrradianceW/m20142.021056.65213.49
Global Horizontal IrradianceW/m2021.22157.8931.90
Air Temperature°C−8.0418.0147.638.56
AtmospherehPa881.40956.421037.7830.54
Relative Humidity%11.8371.7110018.50
PowerMW04.2329.416.52
Table 2. Power Generation and PV Panel Information of the Solar Energy Station.
Table 2. Power Generation and PV Panel Information of the Solar Energy Station.
PV Panel ModelNominal Generation Output Capacity (MW)Number of PV Panels Installations
HR-260P-18/Bbd0.933567
HR-260P-18/Bbd1.927234
GCL-M6/60G2800.15541
YL260P-29b4.6217,782
JC260m-24/Bb6.9626,763
CS6K-260P-PG1.565986
CS6K-255P-PG6.4725,383
CS6K-250P-PG0.301211
TSM-260PC05A2.328908
SYP260P4.2416,326
JMPV-HM6VBM2/60-3400.531559
Table 3. TimeGAN Hyperparameter Demonstration.
Table 3. TimeGAN Hyperparameter Demonstration.
HyperparameterSearch ScopeValue
Time Series Length4949
Batch Size[64,128,256]64
Max Steps[3000–7000]6500
Hidden Layer Size[32,48,64]64
Gamma[0.5–1.0]0.7
Table 4. Hyperparameter optimization scope.
Table 4. Hyperparameter optimization scope.
HyperparameterLower LimitUpper Limit
Learning Rate0.00010.01
Number of BiGRU Units10100
Number of Filters20120
L2 Regularization Parameter0.000010.005
Table 5. Optimized Hyperparameters of the AOO-BiTCN-BiGRU Model.
Table 5. Optimized Hyperparameters of the AOO-BiTCN-BiGRU Model.
HyperparameterSunnyCloudyIntermittentRainy
Learning Rate5.3144 × 10−40.00360.00300.0030
Number of BiGRU Units428191100
Number of Filters1102012056
L2 Regularization Parameter0.00420.00430.00450.005
Table 6. Ablation study results: TimeGAN augmentation for predictive performance.
Table 6. Ablation study results: TimeGAN augmentation for predictive performance.
SampleNumber of Test SamplesMAERMSER2
Original rainy day training dataset17 days0.67710.89310.7972
Combined training dataset17 days0.40910.54390.9358
Table 7. Performance comparison (MAPE, RMSE, R2) of forecasting models under different weather conditions.
Table 7. Performance comparison (MAPE, RMSE, R2) of forecasting models under different weather conditions.
Day WeatherErrorPrimary ModelGRULSTMTCNCNN
SunnyMAE1.30331.44351.39421.23121.8494
RMSE1.67501.83931.87231.58812.1863
R20.97110.96200.96050.97590.9465
CloudyMAE1.12241.23911.42971.46851.2614
RMSE1.51211.66041.84892.00031.6004
R20.93560.91890.89630.87710.9337
IntermittentMAE0.81000.93940.88211.02690.8078
RMSE1.04481.18211.10781.33241.0381
R20.90760.87230.89200.83970.9144
RainyMAE0.45070.44100.50780.61480.6529
RMSE0.58020.59660.69590.83970.7806
R20.93230.92720.89710.84680.8808
AverageMAE0.92161.01581.05351.08541.1429
RMSE1.20301.31961.38121.44011.4014
R20.93670.92010.91150.88490.9189
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jin, Y.; Sun, Z.; Lv, D.; Gao, W.; Liu, F.; Yu, Q. Ultra-Short-Term Multi-Step Photovoltaic Power Forecasting Based on Similarity-Based Daily Clustering. Energies 2026, 19, 29. https://doi.org/10.3390/en19010029

AMA Style

Jin Y, Sun Z, Lv D, Gao W, Liu F, Yu Q. Ultra-Short-Term Multi-Step Photovoltaic Power Forecasting Based on Similarity-Based Daily Clustering. Energies. 2026; 19(1):29. https://doi.org/10.3390/en19010029

Chicago/Turabian Style

Jin, Yongcheng, Zhichao Sun, Dongliang Lv, Weicheng Gao, Fengze Liu, and Qinghua Yu. 2026. "Ultra-Short-Term Multi-Step Photovoltaic Power Forecasting Based on Similarity-Based Daily Clustering" Energies 19, no. 1: 29. https://doi.org/10.3390/en19010029

APA Style

Jin, Y., Sun, Z., Lv, D., Gao, W., Liu, F., & Yu, Q. (2026). Ultra-Short-Term Multi-Step Photovoltaic Power Forecasting Based on Similarity-Based Daily Clustering. Energies, 19(1), 29. https://doi.org/10.3390/en19010029

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop