Next Article in Journal
Computational Analysis of the Effects of Power on the Electromagnetic Characteristics of Microwave Systems with Plasma
Previous Article in Journal
Study on Low-Carbon Design Strategy of Block-Scale Science and Technology Industrial Park Based on Solar Energy Utilization Potential and Heat Island Effects
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Short-Term Photovoltaic Power Prediction Through Multi-Method Data Processing and SFOA-Optimized CNN-BiLSTM

1
Hubei Technology Innovation Center for Smart Hydropower, Wuhan 430010, China
2
School of Automation, Wuhan University of Technology, Wuhan 430070, China
*
Author to whom correspondence should be addressed.
Energies 2025, 18(19), 5124; https://doi.org/10.3390/en18195124
Submission received: 22 August 2025 / Revised: 16 September 2025 / Accepted: 24 September 2025 / Published: 26 September 2025

Abstract

The increasing global demand for renewable energy poses significant challenges to grid stability due to the fluctuation and unpredictability of photovoltaic (PV) power generation. To enhance the accuracy of short-term PV power prediction, this study proposes an innovative integrated model that combines Convolutional Neural Networks (CNN) and Bidirectional Long Short-Term Memory (BiLSTM), optimized using the Starfish Optimization Algorithm (SFOA) and integrated with a multi-method data processing framework. To reduce input feature redundancy and improve prediction accuracy under different conditions, the K-means clustering algorithm is employed to classify past data into three typical weather scenarios. Empirical Mode Decomposition is utilized for multi-scale feature extraction, while Kernel Principal Component Analysis is applied to reduce data redundancy by extracting nonlinear principal components. A hybrid CNN-BiLSTM neural network is then constructed, with its hyperparameters optimized using SFOA to enhance feature extraction and sequence modeling capabilities. The experiments were carried out with historical data from a Chinese PV power station, and the results were compared with other existing prediction models. The results demonstrate that the Root Mean Square Error of PV power generation prediction for three scenarios are 9.8212, 12.4448, and 6.2017, respectively, outperforming all other comparative models.

1. Introduction

As the world places greater importance on environmental protection and sustainable development, the share of renewable energy in the energy mix has been growing steadily. Solar energy, acclaimed as a clean, emission-free, and sustainable resource, is an optimal choice for green energy development [1]. Photovoltaic (PV) power generation, the principal means of harnessing solar energy, is experiencing rapid growth. Nevertheless, PV output is affected by multiple factors, making it fluctuating and unpredictable. This variability challenges grid stability and dispatch [2]. Consequently, improving PV power forecasting and mitigating its grid impacts have become key to ensuring dependable and efficient system operation [3].
Traditional PV power prediction methods include physical models and statistical approaches, which have been widely studied [4]. Physical models rely on detailed meteorological and environmental factors, but they often suffer from high computational complexity and sensitivity to input accuracy [5]. Statistical methods like exponential smoothing [6], auto-regressive moving average [7], and autoregressive integrated moving average [8] capture historical patterns but struggle with nonlinear and complex relationships, degrading accuracy.
Deep learning techniques have garnered considerable attention for PV power prediction in recent years due to their powerful feature extraction and nonlinear modeling capabilities [9]. Many scholars have applied Convolutional Neural Networks (CNN) [10], gated recurrent units [11], and Long Short-Term Memory (LSTM) [12] to improve forecasting accuracy. Building on this, many studies adopt hybrid architectures. Zang et al. [13] proposed a spatiotemporal hybrid CNN-LSTM model for short-term solar radiation prediction and evaluated its performance in different seasons and weather conditions, showing its superior predictive ability. Limouni et al. [14] proposed an LSTM-TCN-based forecasting model to improve prediction accuracy, and the results demonstrated that the proposed model outperforms baseline and conventional models in forecasting accuracy. Li et al. [15] introduced a Stacking ensemble of optimized Neural Prophet, LSTM, and Informer models, which achieved good performance by delivering lower errors, improved stability across different forecasting horizons, and better adaptability to seasonal changes in solar power generation. However, deep learning models often face hyperparameter selection challenges, significantly affecting their performance.
To address this problem, scholars have explored various optimization algorithms to meet the challenge of optimizing deep learning parameters. Hou et al. [16] used the whale optimization algorithm (WOA) to optimize the training iterations, learning rate, and number of hidden neurons of the LSTM to enhance the prediction accuracy of the model. Li et al. [17] proposed a short-term PV power generation prediction method by combining the sparrow search algorithm (SSA) and BiLSTM, and verified its effectiveness by comparing it under different seasons and weather. Zhou et al. [18] utilized the RIME optimization algorithm (RIME) to modify the hyperparameters of neural networks and proved the capability of RIME in short-term photovoltaic power prediction. Zhang et al. [19] introduced a short-term PV forecasting model based on a secondary decomposition method, and optimized the BiLSTM hyperparameters using the Crested Porcupine Optimization algorithm (CPO). The results showed that CPO effectively tuned the model parameters, leading to improved forecasting performance. However, these algorithms still have some drawbacks, such as long optimization times and the risk of falling into local optima, resulting in suboptimal performance.
Further, data quality is another key factor that directly affects the accuracy of power prediction. The power generation data series is a non-smooth, non-linear signal with various time scales and contains numerous hidden patterns. Integrating signal decomposition and dimensionality reduction methods with deep learning can improve the accuracy of power predictions. Wang et al. [20] proposed a PCA-SSA-VMD-based hybrid model to enhance offshore wind power forecasting for economic scheduling and demonstrated that the proposed model effectively enhances prediction precision and validates its effectiveness. Zhang et al. [21] proposed a combined model that combines Empirical Mode Decomposition (EMD) and Kernel Principal Component Analysis (KPCA) for data preprocessing with a deep learning framework for wind power forecasting, achieving superior prediction accuracy compared to benchmark models. Peng et al. [22] developed a hybrid model that combines symplectic geometry model decomposition, KPCA, and an optimized BiLSTM, achieving highly accurate forecasts of wind and photovoltaic power. However, this method is rarely used for PV power forecasting.
Based on the analysis above, this research proposes a hybrid prediction framework that combines data classification, selection, and optimization algorithms with neural networks. The main contributions and innovations are summarized as follows:
(1)
Classify the data into multiple categories by using clustering and construct differentiated prediction models for various weather conditions to improve the model’s adaptability to complex weather.
(2)
The feature data are decomposed by using EMD and downscaled by using KPCA to extract key features, reduce redundancy, and improve computational efficiency. The processed model is then compared with the unprocessed counterpart to verify its advantages in improving the stability of PV power forecasting and its adaptability to complex meteorological conditions.
(3)
The SFOA is first introduced into the field of photovoltaic forecasting to automatically optimize the hyperparameters of models. It is compared with other optimization algorithms, demonstrating its advantage in improving the generalization ability of photovoltaic forecasting models.
The rest of the paper is arranged as follows: Section 2 describes the principle behind each method employed. Section 3 introduces the prediction model and details the experimental steps. Section 4 shows the experimental data along with the analysis methodology. Section 5 shows the numerical results and compares them with alternative methods. Section 6 provides the conclusions and points out issues for further research.

2. Principle

2.1. K-Means Clustering Algorithm

K-means is a widely used clustering algorithm that divides a dataset into K clusters by minimizing within-cluster variance. The specific procedure is as follows:
Step 1: Given data D = { x 1 , x 2 , , x N } , randomly initialize K centroids { c 1 , c 2 , , c K } .
Step 2: For each data point x i D , compute its Euclidean distance from each cluster centroids c   j by using the formula d ( x i , c   j ) = x i c   j   . Each data point is then allocated to the cluster corresponding to the nearest centroid. This results in K clusters, where each data point x i is assigned to the cluster whose centroid is closest.
Step 3: After assigning all data points to their respective clusters, the centroids of the clusters are recalculated. The new centroid c   j for each cluster C   j is computed as the mean of the data points in that cluster:
c   j = 1 | C j | x i C j x i
where c   j is the set of data points in cluster j , and C   j is the number of points in the cluster.
Step 4: Steps 2 and 3 are repeated until the cluster centroids stabilize, or the maximum number of iterations is reached, indicating the algorithm’s convergence.

2.2. Empirical Mode Decomposition

EMD is a decomposition method for nonlinear, non-smooth signals, breaking the signal into intrinsic modal functions (IMFs) that capture oscillatory components at different time scales. The specific decomposition process is as follows:
Step 1: Given a signal x ( t ) , the first step is to identify the local extrema. These extrema points are used to construct the lower and upper envelopes of the signal, denoted as u ( t ) and l ( t ) , respectively.
Step 2: The local mean m ( t ) is calculated as the average of these envelopes. The first component is h 1 ( t ) = x ( t ) m ( t ) . If h 1 ( t ) satisfies the stopping criteria, which requires equal numbers of extrema and zero-crossings, it is considered the first IMF.
Step 3: The residual r 1 ( t ) = x ( t ) h 1 ( t ) becomes the new signal, and the procedure repeats to find further IMFs, continuing until the residual r n ( t ) meets the termination condition.
Step 4: The decomposition stops when the residual signal r n ( t ) becomes monotonic or reaches a predefined threshold δ ( t ) . The final decomposition is:
x ( t ) = h 1 ( t ) + h 2 ( t ) + + h n ( t ) + r n ( t )
where h 1 ( t ) , h 2 ( t ) ,…, h n ( t ) are the IMFs, and r n ( t ) is the residual.

2.3. Kernel Principal Component Analysis

After signal decomposition, the increased input dimension can complicate the model and reduce its generalization ability. KPCA employs a kernel function to transform data into a high dimensional feature space for nonlinear dimensionality reduction. In this space, principal component analysis is applied to extract the key features of the data. Unlike traditional PCA, KPCA enables nonlinear feature extraction by leveraging the kernel function, which avoids high-dimensional mapping. The specific procedure is as follows:
Step 1: Select a kernel function and map the original data into a high-dimensional space, obtaining the corresponding data matrix. The polynomial kernel function is defined as:
K ( x i , x j ) = φ ( x i ) d φ ( x j )
where x i , x j are original data samples, and d is the degree of the highest-order term.
Step 2: Center the kernel matrix and obtain the centered kernel matrix K c . The equation is presented below:
K c = K I N K K I N + I N K I N
where I N represents an N × N matrix whose elements are all 1 / N .
Step 3: Calculate the eigenvalues and eigenvectors of the centered kernel matrix.
Step 4: The cumulative contribution of the eigenvalues r 1 , r 2 , , r n is calculated to quantify the contribution of each eigenvalue to the total variability of the data. When the cumulative contribution r t reaches a predefined threshold p , the top t principal components a 1 , a 2 , , a t are selected as the reduced-dimensional representation of the data.

2.4. SFOA-CNN-BiLSTM Model

2.4.1. CNN-BiLSTM Neural Network Model

The CNN-BiLSTM model is a hybrid architecture combining Convolutional Neural Networks and Bidirectional Long Short-Term Memory Networks. This architecture leverages the strengths of both models, making it particularly effective for tasks like sequential data analysis.
(1)
CNN Layer
In the CNN layer, multiple convolutional layers are used to automatically capture spatial hierarchies and local patterns from the input data. These layers apply convolutional filters, generating feature maps that capture local dependencies and patterns. The CNN layer for a single filter w applied to an input sequence x can be expressed as:
y = f ( x w + b )
where x represents the input sequence, w represents the convolutional filter, b represents the bias term, and f represents an activation function.
(2)
BiLSTM Layer
The BiLSTM layer is employed to capture the sequential dependencies in the data. Unlike traditional LSTMs that capture only past information, BiLSTM networks capture sequence information by processing data bidirectionally. This enables them to capture information from both the past and the future, making them particularly effective for tasks that require understanding both preceding and succeeding elements. A single LSTM unit computes the following relations:
i t = σ ( W i [ h t 1 , x t ] + b i )
f t = σ ( W f [ h t 1 , x t ] + b f )
o t = σ ( W o [ h t 1 , x t ] + b o )
c t = f t c t 1 + i t tanh ( W c [ h t 1 , x t ] + b c )
h t = o t tanh ( c t )
where i t , f t , o t are the input, forget, and output gates, respectively, h t is the hidden state, c t is the cell state, W are the weights and b are the biases, x t is the current input, and h t 1 is the previous hidden state.
The BiLSTM can process the input sequence in two directions. The outputs of both directions are then concatenated to provide a more comprehensive representation of the sequence. For each time step t , the hidden state h t of the BiLSTM output is a splice of the forward and reverse LSTM outputs:
h t = [ h t , h t ]
where h t and h t indicate the forward and reverse LSTM outputs at time step t , respectively.
(3)
Overall CNN-BiLSTM Structure
In the CNN-BiLSTM model, the CNN layers initially extract local features from the input sequence. These features are then passed to the BiLSTM layers, which model the temporal dependencies of the sequence in both directions. The specific model parameters are presented in Table 1, and the structure of the CNN-BiLSTM model is depicted in Figure 1.

2.4.2. SFOA Principle

SFOA is a novel bio-inspired metaheuristic inspired by the behaviors of starfish, including exploration, preying, and regeneration [23]. It employs a hybrid search strategy that integrates both five-dimensional and unidimensional search patterns. This integration can not only optimize resource utilization but also significantly enhance computational efficiency, enabling SFOA to swiftly and accurately navigate through complex solution spaces and identify optimal solutions more effortlessly. The main ideas of SFOA are expressed as follows:
(1)
Initialization
X i j = l j + r ( u j l j ) , i = 1 , 2 , , N , j = 1 , 2 , , D
where X i j denotes the j th dimensional position of the i th starfish, r denotes a random number between (0, 1), u j and l j are bounds of the j th dimensional design variable, respectively.
(2)
Exploration phase
If the optimization problem has a dimension greater than 5, SFOA uses the following mathematical model:
Y i , p T = X i , p T + a 1 ( X i , p T X i , p T ) cos θ , r 0.5
Y i , p T = X i , p T a 1 ( X best , p T X i , p T ) sin θ , r > 0.5
a 1 = ( 2 r 1 ) π
θ = π 2 T T max
where Y i , p T and X i , p T denote the obtained position and the current position of a starfish, respectively; X best , p T denotes the p -dimension of the best current position, p is a randomly selected 5-dimension in the D -dimension, r denotes the random number between (0, 1); T is the current number of iterations, and T max is the maximal number of iterations, θ is in the range of [0, π / 2 ].
If the optimization problem has no dimension greater than 5, the exploration phase uses a one-dimensional search pattern to update the position.
Y i , q T = E t X i , p T + A 1 ( X k 1 , p T X i , p T ) + A 2 ( X k 2 , p T X i , p T )
E t = T max T T max cos θ
where X k 1 , p T and X k 2 , p T are the positions of two randomly chosen starfish in the p-dimension, A 1 and A 2 are two random numbers between (−1, 1), and p is a randomly chosen number in the D dimension.
(3)
Exploitation phase
SFOA employs a parallel bidirectional search strategy, which leverages information from other starfish and the current best position within the population. The distances can be calculated as:
d m = ( X best T X m p T ) , m = 1 , , 5
where d m represents the five distances between the global best and other starfish, and m p refers to five randomly selected starfish. The first update rule for each starfish is modeled as follows:
Y i T = X i T + r 1 d m 1 + r 2 d m 2
where r 1 and r 2 are random numbers between (0, 1), d m 1 and d m 2 are randomly selected in d m . In addition, if i = N , another update rule is modeled as follows:
Y i T = exp T × N T max X i T
where T is the current iteration, T max is the maximum iterative number, and N is the population size.

2.4.3. The Steps of the SFOA-CNN-BiLSTM Model

The network performance and convergence speed are determined by the values of its hyperparameters, which can be optimized by SFOA. The main steps in the SFOA optimization of networks are as follows:
Step 1: Determine the optimized hyperparameter range and initialize the SFOA population.
Step 2: Select RMSE as the fitness function.
Step 3: Entering the optimization phase of the SFOA, the process undergoes iterative optimization through two distinct stages.
Step 4: Check the stopping condition. If the maximum number of iterations is reached, the optimization process stops.
Step 5: Obtain the best set of hyperparameters.
Figure 2 shows the steps of optimization of CNN-BiLSTM using SFOA.

2.5. Evaluation Metrics

To assess the model’s performance, six performance metrics are used in this study: RMSE, Mean Squared Error (MSE), Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), Pearson Correlation Coefficient (R), and Coefficient of Determination (R2). These metrics are commonly used to assess the accuracy and reliability of forecasting models. The detailed equations for each metric are expressed as follows:
RMSE = 1 n i = 1 n ( y i y ^ i ) 2
MSE = 1 n i = 1 n ( y i y ^ i ) 2
MAE = 1 n i = 1 n | y i y ^ i |
MAPE = 1 n i = 1 n y i y ^ i y i × 100 %
R = i = 1 n ( y i y ¯ ) ( y ^ i y ^ ¯ ) i = 1 n ( y i y ¯ ) 2 i = 1 n ( y ^ i y ^ ¯ ) 2
R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2
where n is the number of samples, y i is the real power value, y ^ i is the predicted power value, and y ¯ is the average power value.

3. PV Power Prediction Based on the Developed Fusion Method

In this paper, a fusion method for PV power prediction based on similar day clustering with EMD-KPCA-SFOA-CNN-BiLSTM is developed. The main steps are illustrated in Figure 3 and they are described as follows:
Step 1: Data acquisition and processing. The raw power generation dataset is first preprocessed and the most representative feature parameters are screened by correlation coefficient analysis. Then, a similar day clustering method is used to group the data.
Step 2: Feature data decomposition and dimensionality reduction. The feature data is first decomposed by using EMD. Subsequently, KPCA is applied to reduce dimensionality to obtain the principal components affecting the PV output power.
Step 3: Hyperparameter optimization and model training. In training the CNN-BiLSTM model, the SFOA is employed to optimize key hyperparameters to enhance overall model performance.
Step 4: PV power forecasting. During the forecasting stage, the preprocessed data is input into the neural network with optimized hyperparameters to generate short-term PV power predictions.
Step 5: Results analysis. Different evaluation metrics are used to judge the prediction results and to validate the advantages of the developed fusion method in comparison with other prediction methods.

4. Data Analysis and Prediction Process

4.1. Data Description and Preprocessing

The experiments were conducted on a system running Windows 11, equipped with an AMD Ryzen 7 5800H processor and an NVIDIA GeForce RTX 3050 GPU. All experiments were executed using MATLAB 2021b, with GPU acceleration enabled for enhanced computational performance.
The photovoltaic power generation data used in the study were collected from a power station in China, with a total of 92 days of records gathered from June to August 2017. The data was collected every 15 min. However, since the system does not generate power at night, only the period from 6:00 AM to 7:00 PM is included and this gives 52 data points per day. The collected data were as follows: total radiation, direct radiation, scattered radiation, component temperature, temperature, pressure, relative humidity, and PV power generation.
During the process of collecting or transmitting the raw photovoltaic power generation data, data loss and abnormal fluctuations may occur. In this study, missing values and outliers are filled or replaced with the average values from the corresponding time points before and after the problematic data points. The descriptions of the data items in the database are presented in Table 2.

4.2. Characteristic Analysis

To analyze the relationship between the extracted features and their predictive ability, we used the Pearson correlation coefficient (PCC) to examine the correlation between each feature and the PV power. By calculating the PCC for each feature with the PV power, we can identify which features have higher predictive value in short-term PV power forecasting models. The calculation method for the PCC is shown as follow.
PCC = i = 1 n ( X i X ¯ ) ( Y i Y ¯ ) i = 1 n ( X i X ¯ ) 2 i = 1 n ( Y i Y ¯ ) 2
where X i is the different values of feature X , Y i is the different values of feature Y , X ¯ is the mean value of feature X , and Y ¯ is the mean value of feature Y .
The result of the PCC between the environmental factors and PV power is displayed in Table 3. A PCC value closer to 1 indicates a stronger correlation between the features and the PV power.
As can be seen from Table 3, the correlation between total radiation, direct radiation, scattered radiation, and component temperature with PV power is strong. Therefore, these features were selected as the input features for the subsequent sessions (Feature I: Total Radiation, Feature II: Direct Radiation, Feature III: Scattered Radiation, Feature IV: Component Temperature).

4.3. Similar Day Division

In practice, weather conditions impact PV power generation in various ways. To account for this, we applied the k-means clustering algorithm to group the data based on different weather scenarios. This classification allows the subsequent models to make more accurate predictions tailored to specific weather conditions. Through clustering, the 92 days of data are divided into three typical PV output day scenarios: Scenario I includes 41 typical days, representing strong sunlight and sunny weather. Scenario II includes 32 typical days, representing moderate sunlight and cloudy weather. Scenario III includes 19 typical days, representing weak sunlight and rainy weather. Figure 4 shows the results of the k-means clustering algorithm.

4.4. Analysis of the EMD Results

In this section, building upon the previous analyses, we further applied EMD to decompose the four features and obtain the IMF components and residual components for each feature under different scenarios. Taking Scenario III as an example, Figure 5 illustrates the decomposition results of EMD. For all three scenarios, the results of the EMD for the four features are presented in Table 4.

4.5. KPCA Dimension Reduction

The IMF components obtained from EMD may exhibit high-dimensional redundancy and nonlinear coupling, which can lead to dimensionality issues and increase the risk of overfitting when directly used in prediction models. To better capture the nonlinear relationships between the features and reduce model complexity, we applied kernel principal component analysis to the feature sequence data. We selected components with a cumulative contribution greater than 90%, ensuring that the selected principal components more effectively portray the original data and enhance information synthesis. The processed data were subsequently employed as input samples for predicting future PV output power. Figure 6 illustrates the proportion of information captured by using KPCA across the three scenarios and Table 5 shows the number of components of each scenario after the KPCA dimensionality reduction.

4.6. Model Hyperparameter Settings

In this section, based on the CNN-BiLSTM neural network, the optimal hyperparameters of the network model are obtained by using SFOA computation to enhance the model performance and prediction accuracy. The hyperparameters optimized by the SFOA include the learning rate, dropout rate, the filters’ number in each CNN layer, the hidden units’ number in each BiLSTM layer, and the regularization parameters. In addition, we chose MSE loss as the loss function and the parameters are solved and continuously updated by using the Adam solver. The optimization range of hyperparameters is shown in Table 6.
The dimensionality-reduced feature data and historical photovoltaic power were input into the CNN-BiLSTM model optimized by SFOA. We divided the dataset into training and testing sets at a ratio of 8:2.

5. Results and Discussions

5.1. Compared with Other Power Prediction Models

To validate the predictive effectiveness and high accuracy of the combined model, we employed five different models to make predictions on the test sets of three distinct scenarios. The prediction models used in this section are listed below: ① Model I: BiLSTM; ② Model II: CNN-BiLSTM; ③ Model III: EMD-KPCA-CNN-BiLSTM; ④ Model IV:SFOA-CNN-BiLSTM; and ⑤ Model V:EMD-KPCA-SFOA-CNN-BiLSTM. In the three scenarios, we chose August 18, August 19, and August 27, and Figure 7 shows the single-day prediction results of each model in the three different scenarios. At the same time, Table 7 shows the evaluation metrics for the prediction results across the entire test set (the results for Scenario III with Model I are excluded because of its poor performance).
By analyzing the results from the table, it can be obtained that:
(1)
The CNN-BiLSTM model improves its performance by adding a Convolutional Neural Network on the top of BiLSTM, and it helps the model extract more advanced information from local features. Compared with the single BiLSTM model, there is a significant improvement in all indicators, which proves the effectiveness of the CNN layer in enhancing the feature extraction ability of the model.
(2)
The EMD-KPCA-CNN-BiLSTM model adds EMD and KPCA techniques to the CNN-BiLSTM model. The RMSEs in the three scenarios decreased by 4.67%, 2.91%, and 2.67%, while the MAEs decreased by 8.19%, 6.65%, and 1.23%, respectively. This demonstrates that the combination of EMD and KPCA effectively reduces dimensionality while preserving the key information in the data, thereby lowering computational complexity and enhancing prediction performance.
(3)
The SFOA-CNN-BiLSTM model optimizes hyperparameters to further enhance performance. SFOA significantly reduces errors, leading to substantial improvements in RMSE, MSE, MAE, and MAPE.
(4)
The EMD-KPCA-SFOA-CNN-BiLSTM model combines four methods to maximize the performance. The model can handle complex time series data efficiently and performs the best among all the models. All the metrics are optimal in different scenarios, which proves its superiority.

5.2. Compared with Other Optimization Algorithms

To verify the superiority of SFOA in dealing with short-term PV power forecasting problems, we used three other typical optimization algorithm methods to replace SFOA in Model V above, all of which have demonstrated effective optimization results in PV forecasting. The comparison optimization methods used in the section are listed below: ① WOA; ② SSA; and ③ RIME. Figure 8 shows the same single-day prediction results as in the previous section for each method for three different scenarios and Table 8 shows the evaluation metrics for the prediction results across the entire test set. To provide a clearer comparison of model performances, we have included a global comparative figure (see Figure 9) that summarizes the performance of the different models across the three scenarios. This consolidated figure offers a comprehensive view of how each model performs under varying conditions, allowing for a more direct comparison of their strengths and improvements.
By analyzing the results from the table, it can be seen that:
(1)
In Scenario I, the prediction results of the methods are generally close to each other and SFOA performs the best in most of the indicators, which means that its prediction results have the smallest deviation from the actual values. Although WOA is slightly better in MAPE, the advantage of SFOA is more obvious in the overall error level. In addition, the higher correlation coefficient R (0.9546) and coefficient of determination R2 (0.9092) also indicate that SFOA is not only smaller in absolute error, but also has stronger data fitting and interpretation capabilities, and is therefore considered the best choice in this scenario.
(2)
In Scenarios II and III, a comparison across methods shows that SFOA consistently outperforms the others in all error metrics. This superior performance indicates that SFOA is highly effective at capturing data variability and achieving an optimal model fit. Even under more challenging prediction tasks, it maintains impressive prediction accuracy and strong explanatory power.
To validate the generalizability of the model, we extended the experiment by using an additional dataset sourced from a 1.8 MW distributed solar PV plant located in Uluru (Ayers Rock), Central Australia. Specifically, we selected Yulara Solar Site-2, which is equipped with a 226.8 kW polycrystalline silicon system, and analyzed data collected from 1 December 2017, to 30 November 2018 [24]. The data was divided into four seasons: Spring, Summer, Autumn, and Winter. The experimental procedure and related parameters remained consistent with the previous setup. First, the PCC was calculated, followed by EMD for denoising, and KPCA for feature dimensionality reduction. Subsequently, the four optimization algorithms were applied for training and prediction. Figure 10 presents the results of the experiment.
The results show that SFOA demonstrates excellent predictive capability on the new dataset as well. In the comparison across the four seasons, SFOA remains the overall leader. Although in Winter, SFOA’s MAE (3.589) is slightly higher than WOA (3.532) and SSA (3.530), and R (0.991) is slightly smaller than SSA (0.992), other metrics still outperform those of the other algorithms. In summary, SFOA consistently maintains low errors and high accuracy across different climatic conditions in all seasons, demonstrating its robust performance.
Furthermore, to assess the computational cost of each optimization algorithm, we measured the time taken for each algorithm to complete one optimization iteration. In all experiments, the population size was set to 10, and the number of iterations was set to 10, with each optimization algorithm successfully converging. According to the data in Table 9, SFOA outperforms WOA and SSA in terms of optimization iteration time, while its computational time remains comparable to that of RIME. These findings demonstrate that, while improving prediction accuracy, SFOA does not significantly increase computational resource overhead, indicating a good balance between accuracy enhancement and computational efficiency.

6. Conclusions

To accurately predict PV power generation, this research introduces a short-term PV power prediction approach according to multi-method data processing and a CNN-BiLSTM model optimized by the SFOA. Initially, typical days are clustered using the K-means algorithm. Subsequently, key components are retained and redundant information is eliminated through the EMD and KPCA. Finally, a hybrid SFOA-CNN-BiLSTM model is developed to achieve accurate PV power prediction. The advantage of the proposed method is verified and analyzed through relevant case studies, leading to the following conclusions:
(1)
Combining EMD and KPCA to preprocess the input data enables the model to effectively filter out noise and irrelevant information. This process enhances the model’s ability to capture the underlying patterns and trends in the data, leading to more accurate and reliable predictions. Under three different scenarios, the RMSEs of the models processed with EMD-KPCA were reduced by 4.67%, 2.91%, and 2.67%, respectively, compared to those without it. The results illustrate the effectiveness of EMD-KPCA in improving the accuracy of PV data processing.
(2)
The SFOA-CNN-BiLSTM power prediction model can help to solve the problem of difficult parameterization of traditional deep learning networks and improve prediction accuracy. Compared to the model without SFOA, all metrics have improved dramatically. At the same time, compared with other optimization algorithms, SFOA has improved in most of the metrics and has the most comprehensive performance.
(3)
The proposed hybrid model outperforms other comparative models in most metrics, while also maintaining a low computational cost, thus demonstrating strong robustness and generalizability.
The hybrid model proposed in this paper can achieve good results in predicting PV power. However, there are still some shortcomings:
(1)
In this study, Pearson correlation was used to calculate and select relevant features. However, relying solely on Pearson correlation may overlook nonlinear and lagged relationships. In future research, we will pay more attention to selecting these types of features to improve the model’s performance and capture more complex dependencies.
(2)
Optimizing the hyperparameters of the network by using SFOA can significantly improve prediction accuracy. However, it has some weaknesses that may lead to suboptimal results, particularly in scenarios involving small real values, where relative errors have a greater impact. As a result, this can lead to a slightly higher MAPE compared to other algorithms, despite maintaining strong overall prediction accuracy. To address this issue, we will improve the SFOA in the next phase, or adopt more robust evaluation metrics and explore incorporating weighted error indicators or other techniques during the training process to mitigate bias in low-power regions, further enhancing the model’s accuracy and robustness.
(3)
The model currently relies solely on data-driven methods, without accounting for real-time physical mechanisms. Incorporating factors like irradiance, module temperature, and cloud dynamics could enhance the model’s credibility and transferability in real-world settings. These physical variables directly impact system performance and would improve the model’s robustness, making it more applicable in dynamic environments. Future work will explore integrating these physical elements to strengthen the model’s real-world accuracy.

Author Contributions

Conceptualization, X.H.; methodology, X.H.; software, Z.Z.; validation, Z.Z.; formal analysis, T.Y. and Z.S.; investigation, T.Y. and Z.S.; writing—original draft preparation, Z.Z. and Y.S. (Yixin Su); writing—review and editing, Z.Z.; visualization, Y.S. (Yun Shao); supervision, X.H. and Y.S. (Yixin Su); project administration, X.H. and Y.S. (Yixin Su); funding acquisition, Y.S. (Yixin Su). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Open Research Fund of Hubei Technology Innovation Center for Smart Hydropower, (Grant No. 152356001924), and the National Natural Science Foundation of Hubei Province, (Grant No. 2024AFD364).

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

Authors Xiaojun Hua, Tao Ye, Zida Song and Yun Shao were employed by the company Hubei Technology Innovation Center for Smart Hydropower. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Kakran, S.; Rathore, J.S.; Sidhu, A.; Kumar, A. Solar Energy Advances and CO2 Emissions: A Comparative Review of Leading Nations’ Path to Sustainable Future. J. Clean. Prod. 2024, 475, 143598. [Google Scholar] [CrossRef]
  2. Maghami, M.R.; Hizam, H.; Gomes, C.; Radzi, M.A.; Rezadad, M.I.; Hajighorbani, S. Power Loss Due to Soiling on Solar Panel: A Review. Renew. Sustain. Energy Rev. 2016, 59, 1307–1316. [Google Scholar] [CrossRef]
  3. Jiang, M.; Ding, K.; Chen, X.; Cui, L.; Zhang, J.; Yang, Z.; Cang, Y.; Cao, S. Research on Time-Series Based and Similarity Search Based Methods for PV Power Prediction. Energy Convers. Manag. 2024, 308, 118391. [Google Scholar] [CrossRef]
  4. Massidda, L.; Bettio, F.; Marrocu, M. Probabilistic Day-Ahead Prediction of PV Generation. A Comparative Analysis of Forecasting Methodologies and of the Factors Influencing Accuracy. Sol. Energy 2024, 271, 112422. [Google Scholar] [CrossRef]
  5. de Oliveira Santos, L.; AlSkaif, T.; Barroso, G.C.; de Carvalho, P.C.M. Photovoltaic Power Estimation and Forecast Models Integrating Physics and Machine Learning: A Review on Hybrid Techniques. Sol. Energy 2024, 284, 113044. [Google Scholar] [CrossRef]
  6. Ferbar Tratar, L.; Strmčnik, E. The Comparison of Holt–Winters Method and Multiple Regression Method: A Case Study. Energy 2016, 109, 266–276. [Google Scholar] [CrossRef]
  7. Rigby, A.; Baker, U.; Lindley, B.; Wagner, M. Generation and Validation of Comprehensive Synthetic Weather Histories Using Auto-Regressive Moving-Average Models. Renew. Energy 2024, 224, 120157. [Google Scholar] [CrossRef]
  8. Shadab, A.; Ahmad, S.; Said, S. Spatial Forecasting of Solar Radiation Using ARIMA Model. Remote Sens. Appl. Soc. Environ. 2020, 20, 100427. [Google Scholar] [CrossRef]
  9. Khan, Z.A.; Hussain, T.; Baik, S.W. Dual Stream Network with Attention Mechanism for Photovoltaic Power Forecasting. Appl. Energy 2023, 338, 120916. [Google Scholar] [CrossRef]
  10. Wang, H.; Yi, H.; Peng, J.; Wang, G.; Liu, Y.; Jiang, H.; Liu, W. Deterministic and Probabilistic Forecasting of Photovoltaic Power Based on Deep Convolutional Neural Network. Energy Convers. Manag. 2017, 153, 409–422. [Google Scholar] [CrossRef]
  11. Nhat, N.N.V.; Huu, D.N.; Nguyen, T.T.H. Short-Term Multi-Step Forecasting of Rooftop Solar Power Generation Using a Combined Data Decomposition and Deep Learning Model of EEMD-GRU. J. Renew. Sustain. Energy 2024, 16, 016101. [Google Scholar] [CrossRef]
  12. Luo, X.; Zhang, D.; Zhu, X. Deep Learning Based Forecasting of Photovoltaic Power Generation by Incorporating Domain Knowledge. Energy 2021, 225, 120240. [Google Scholar] [CrossRef]
  13. Zang, H.; Liu, L.; Sun, L.; Cheng, L.; Wei, Z.; Sun, G. Short-Term Global Horizontal Irradiance Forecasting Based on a Hybrid CNN-LSTM Model with Spatiotemporal Correlations. Renew. Energy 2020, 160, 26–41. [Google Scholar] [CrossRef]
  14. Limouni, T.; Yaagoubi, R.; Bouziane, K.; Guissi, K.; Baali, E.H. Accurate One Step and Multistep Forecasting of Very Short-Term PV Power Using LSTM-TCN Model. Renew. Energy 2023, 205, 1010–1024. [Google Scholar] [CrossRef]
  15. Li, Y.; Liu, G.; Cao, Y.; Chen, J.; Gang, X.; Tang, J. WNPS-LSTM-Informer: A Hybrid Stacking Model for Medium-Term Photovoltaic Power Forecasting with Ranked Feature Selection. Renew. Energy 2025, 244, 122687. [Google Scholar] [CrossRef]
  16. Hou, Z.; Zhang, Y.; Liu, Q.; Ye, X. A Hybrid Machine Learning Forecasting Model for Photovoltaic Power. Energy Rep. 2024, 11, 5125–5138. [Google Scholar] [CrossRef]
  17. Li, Y.; Huang, W.; Lou, K.; Zhang, X.; Wan, Q. Short-Term PV Power Prediction Based on Meteorological Similarity Days and SSA-BiLSTM. Syst. Soft Comput. 2024, 6, 200084. [Google Scholar] [CrossRef]
  18. Zhou, D.; Liu, Y.; Wang, X.; Wang, F.; Jia, Y. Combined Ultra-Short-Term Photovoltaic Power Prediction Based on CEEMDAN Decomposition and RIME Optimized AM-TCN-BiLSTM. Energy 2025, 318, 134847. [Google Scholar] [CrossRef]
  19. Zhang, J.; Sun, T.; Guo, X.; Lu, M. Short-Term Photovoltaic Power Prediction with CPO-BILSTM Based on Quadratic Decomposition. Electr. Power Syst. Res. 2025, 243, 111511. [Google Scholar] [CrossRef]
  20. Wang, Z.; Ying, Y.; Kou, L.; Ke, W.; Wan, J.; Yu, Z.; Liu, H.; Zhang, F. Ultra-Short-Term Offshore Wind Power Prediction Based on PCA-SSA-VMD and BiLSTM. Sensors 2024, 24, 444. [Google Scholar] [CrossRef]
  21. Zhang, Z.; Deng, A.; Wang, Z.; Li, J.; Zhao, H.; Yang, X. Wind Power Prediction Based on EMD-KPCA-BiLSTM-ATT Model. Energies 2024, 17, 2568. [Google Scholar] [CrossRef]
  22. Peng, S.; Zhu, J.; Wu, T.; Yuan, C.; Cang, J.; Zhang, K.; Pecht, M. Prediction of Wind and PV Power by Fusing the Multi-Stage Feature Extraction and a PSO-BiLSTM Model. Energy 2024, 298, 131345. [Google Scholar] [CrossRef]
  23. Zhong, C.; Li, G.; Meng, Z.; Li, H.; Yildiz, A.R.; Mirjalili, S. Starfish Optimization Algorithm (SFOA): A Bio-Inspired Metaheuristic Algorithm for Global Optimization Compared with 100 Optimizers. Neural Comput. Applic. 2025, 37, 3641–3683. [Google Scholar] [CrossRef]
  24. DKA Solar Centre. Available online: https://dkasolarcentre.com.au/source/yulara/yulara-2-fixed-service-station (accessed on 11 May 2025).
Figure 1. The structure of the CNN-BiLSTM neural network model.
Figure 1. The structure of the CNN-BiLSTM neural network model.
Energies 18 05124 g001
Figure 2. Schematic diagram of the SFOA-CNN-BiLSTM process.
Figure 2. Schematic diagram of the SFOA-CNN-BiLSTM process.
Energies 18 05124 g002
Figure 3. General framework of PV power prediction method.
Figure 3. General framework of PV power prediction method.
Energies 18 05124 g003
Figure 4. Clustering data graph of similar days: (a) strong sunlight and sunny weather; (b) moderate sunlight and cloudy weather; and (c) weak sunlight and rainy weather.
Figure 4. Clustering data graph of similar days: (a) strong sunlight and sunny weather; (b) moderate sunlight and cloudy weather; and (c) weak sunlight and rainy weather.
Energies 18 05124 g004
Figure 5. IMFs of each feature after EMD: (a) Total Radiation; (b) Direct Radiation; (c) Scattered Radiation; and (d) Component Temperature.
Figure 5. IMFs of each feature after EMD: (a) Total Radiation; (b) Direct Radiation; (c) Scattered Radiation; and (d) Component Temperature.
Energies 18 05124 g005
Figure 6. Chart of contribution rate distribution: (a) Scenario I; (b) Scenario II; and (c) Scenario III.
Figure 6. Chart of contribution rate distribution: (a) Scenario I; (b) Scenario II; and (c) Scenario III.
Energies 18 05124 g006
Figure 7. Forecast results for a typical day by using different models: (a) Scenario I; (b) Scenario II; and (c) Scenario III.
Figure 7. Forecast results for a typical day by using different models: (a) Scenario I; (b) Scenario II; and (c) Scenario III.
Energies 18 05124 g007
Figure 8. Forecast results for a typical day by using different methods: (a) Scenario I; (b) Scenario II; and (c) Scenario III.
Figure 8. Forecast results for a typical day by using different methods: (a) Scenario I; (b) Scenario II; and (c) Scenario III.
Energies 18 05124 g008
Figure 9. Global comparison of model performance across three scenarios.
Figure 9. Global comparison of model performance across three scenarios.
Energies 18 05124 g009
Figure 10. Key performance metrics comparison for the extended experiment.
Figure 10. Key performance metrics comparison for the extended experiment.
Energies 18 05124 g010
Table 1. Parameter configuration of the overall model.
Table 1. Parameter configuration of the overall model.
ParameterValue
Input layer time steps8
Output layer variables1
Maximum number of iterations200
Maximum number of epochs200
CNN layers2
BiLSTM layers3
OptimizerAdam
Activation functionReLU
Initial learning rate0.01
L2 Regularization0.002
Table 2. Variable definitions in the Database.
Table 2. Variable definitions in the Database.
Variables UnitMin. ValueMax. Value
Total RadiationW/m20927.67
Direct RadiationW/m20853.4564
Scattered RadiationW/m2089.0563
Component Temperature°C1653.8
Temperature°C040.7
PressurehPa997.81020
Relative Humidity%0116.5395
Table 3. The results of PCC.
Table 3. The results of PCC.
FeaturesPCC
Total Radiation0.538
Direct Radiation0.538
Scattered Radiation0.538
Component Temperature0.292
Temperature0.074
Pressure0.055
Relative Humidity−0.011
Table 4. The results of the EMD.
Table 4. The results of the EMD.
ScenariosFeaturesIMFsResidual Components
Scenario IFeature I91
Feature II91
Feature III91
Feature IV91
Scenario IIFeature I81
Feature II81
Feature III81
Feature IV81
Scenario IIIFeature I81
Feature II81
Feature III81
Feature IV61
Table 5. The results of the KPCA dimensionality reduction.
Table 5. The results of the KPCA dimensionality reduction.
ScenariosNumber of Components After KPCA
Scenario I10
Scenario II8
Scenario III9
Table 6. Hyperparameters and the corresponding tuning ranges.
Table 6. Hyperparameters and the corresponding tuning ranges.
HyperparameterRange
Learning rate[0.001, 0.01]
CNN 1 number of filters[5, 20]
CNN 2 number of filters[5, 20]
BiLSTM 1 number of hidden units[5, 100]
BiLSTM 2 number of hidden units[5, 100]
BiLSTM 3 number of hidden units[5, 100]
Regularization parameter[0.001, 0.01]
Dropout rate[0.1, 0.5]
Note: Superscripts 1–3 denote the layer index in the network (e.g., BiLSTM 1, BiLSTM 2, BiLSTM 3 correspond to the 1st, 2nd, and 3rd BiLSTM layers).
Table 7. Prediction evaluation of five models in three scenarios.
Table 7. Prediction evaluation of five models in three scenarios.
ScenariosModelsRMSEMSEMAEMAPERR2
Scenario IModel I13.0366169.951710.59941.09150.93100.8401
Model II10.9585120.08868.16770.77380.94760.8870
Model III10.4462109.12267.49890.60530.95090.8973
Model IV10.0626101.25587.27790.48460.95210.9047
Model V9.821296.45616.74600.38210.95460.9092
Scenario IIModel I16.1908262.143612.69172.74390.86490.7076
Model II14.1580200.448210.79982.72180.89960.7764
Model III13.7453188.933610.08142.03790.90490.7893
Model IV12.8802165.89909.09242.12770.91220.8150
Model V12.4448154.87268.72851.38950.91670.8273
Scenario IIIModel II6.937348.12605.47901.94080.80150.5982
Model III6.752445.59485.41171.59850.81070.6193
Model IV6.365740.52184.88731.30510.82220.6617
Model V6.201738.46164.48831.08290.83170.6789
Table 8. Comparison results with other algorithm methods.
Table 8. Comparison results with other algorithm methods.
ScenariosMethodsRMSEMSEMAEMAPERR2
Scenario IWOA9.983099.66026.88760.34440.95220.9062
SSA9.908798.18226.79630.43860.95270.9076
RIME10.2335104.72547.48050.60140.95010.9014
SFOA9.821296.45616.74600.38210.95460.9092
Scenario IIWOA13.3452178.0959.46491.84450.90510.8014
SSA13.0029169.07429.24861.93980.90570.8114
RIME13.1256172.28049.31651.58820.90750.8079
SFOA12.4448154.87268.72851.38950.91620.8272
Scenario IIIWOA6.632143.98514.96161.28050.81170.6328
SSA6.473741.90935.06791.47530.82240.6501
RIME6.625943.90244.98151.31110.80040.6335
SFOA6.201738.46164.48831.08290.83170.6789
Table 9. Time per iteration of optimization algorithms in different seasons.
Table 9. Time per iteration of optimization algorithms in different seasons.
Time Per Iteration (s)SpringSummerAutumnWinter
WOA387.39267.39389.56380.54
SSA394.18326.58462.82429.7.5
RIME333.10228.77342.31368.83
SFOA332.82266.72341.37377.20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hua, X.; Zhang, Z.; Ye, T.; Song, Z.; Shao, Y.; Su, Y. Enhanced Short-Term Photovoltaic Power Prediction Through Multi-Method Data Processing and SFOA-Optimized CNN-BiLSTM. Energies 2025, 18, 5124. https://doi.org/10.3390/en18195124

AMA Style

Hua X, Zhang Z, Ye T, Song Z, Shao Y, Su Y. Enhanced Short-Term Photovoltaic Power Prediction Through Multi-Method Data Processing and SFOA-Optimized CNN-BiLSTM. Energies. 2025; 18(19):5124. https://doi.org/10.3390/en18195124

Chicago/Turabian Style

Hua, Xiaojun, Zhiming Zhang, Tao Ye, Zida Song, Yun Shao, and Yixin Su. 2025. "Enhanced Short-Term Photovoltaic Power Prediction Through Multi-Method Data Processing and SFOA-Optimized CNN-BiLSTM" Energies 18, no. 19: 5124. https://doi.org/10.3390/en18195124

APA Style

Hua, X., Zhang, Z., Ye, T., Song, Z., Shao, Y., & Su, Y. (2025). Enhanced Short-Term Photovoltaic Power Prediction Through Multi-Method Data Processing and SFOA-Optimized CNN-BiLSTM. Energies, 18(19), 5124. https://doi.org/10.3390/en18195124

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop