Next Article in Journal
Foundation Models for Cybersecurity: A Comprehensive Multi-Modal Evaluation of TabPFN and TabICL for Tabular Intrusion Detection
Next Article in Special Issue
Enhancing Distribution Network Resilience Using Genetic Algorithms
Previous Article in Journal
Channel Capacity Analysis of Partial-CSI SWIPT Opportunistic Amplify-and-Forward (OAF) Relaying over Rayleigh Fading
Previous Article in Special Issue
A Deep Reinforcement Learning Approach for Energy Management in Low Earth Orbit Satellite Electrical Power Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning-Based Ensemble Method for Parameter Estimation of Solar Cells Using a Three-Diode Model

1
Department of Engineering Science, National Cheng Kung University, Tainan 701401, Taiwan
2
Department of Electrical Engineering, Kun Shan University, Tainan 710303, Taiwan
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(19), 3790; https://doi.org/10.3390/electronics14193790
Submission received: 20 August 2025 / Revised: 16 September 2025 / Accepted: 22 September 2025 / Published: 24 September 2025

Abstract

Accurate parameter estimation of solar cells is critical for early-stage fault diagnosis in photovoltaic (PV) power systems. A physical model based on three-diode configuration has been recently introduced to improve model accuracy. However, nonlinear and recursive relationships between internal parameters and PV output, along with parameter drift and PV degradation due to long-term operation, pose significant challenges. To address these issues, this study proposes a deep learning-based ensemble framework that integrates outputs from multiple optimization algorithms to improve estimation precision and robustness. The proposed method consists of three stages. First, the collected data were preprocessed using some data processing techniques. Second, a PV power generation system is modeled using the three-diode structure. Third, several optimization algorithms with distinct search behaviors are employed to produce diverse estimations. Finally, a hybrid deep learning model combining convolutional neural networks (CNNs) and long short-term memory (LSTM) networks is used to learn from these results. Experimental validation on a 733 kW PV power generation system demonstrates that the proposed method outperforms individual optimization approaches in terms of prediction accuracy and stability.

1. Introduction

Accurate estimation of solar cell parameters is essential for predicting power output and facilitating early-stage fault detection in photovoltaic (PV) power generation systems [1]. Manufacturers typically provide key specifications, such as short-circuit current, open-circuit voltage, and the maximum power point (MPP) parameters of voltage and current under standard test conditions (STCs). These specifications serve as the basis for developing physical models that estimate real-world power generation under various weather conditions.
Several physical models have been proposed for PV modules, including the single-diode, two-diode, and three-diode models. The single-diode model incorporates parameters such as photocurrent, dark saturation current, ideality factor, series resistance, and shunt resistance [2,3,4,5], offering a balance between simplicity and accuracy. The two-diode model [6,7,8,9] extends this by adding a second diode, which improves performance under low irradiance conditions.
In the three-diode model [10,11,12,13,14,15,16,17,18,19], the dark saturation current is represented by three distinct diodes, accounting for (i) diffusion and recombination currents associated with the emitter and the P–N junction, (ii) recombination current in the depletion region, and (iii) currents arising from leakage and grain boundary effects. This model provides the most comprehensive representation, improving accuracy by capturing a wider range of internal physical mechanisms. Nevertheless, the parameters count increases to nine, thereby introducing greater complexity to the model.
Despite improvements in modeling, estimating parameters remains challenging due to the exponential nonlinearities and recursive dependencies between internal parameters and output power. Conventional iterative methods converge rapidly but are susceptible to local optima if initialized poorly [20,21,22,23]. In contrast, multi-agent optimization approaches, such as evolutionary computing (EC), swarm intelligence (SI), and physics-based (PB) methods, leverage multi-point search and parallelism to achieve more reliable solutions.
EC algorithms simulate natural evolution through mutation and crossover, with examples including the flower pollination algorithm (FPA) [24], differential evolution (DE) [25], and genetic algorithm (GA) [26]. SI methods are inspired by collective behaviors, such as particle swarm optimization (PSO) [27], whale optimization algorithm (WOA) [28], bonobo optimizer (BO) [29], and salp swarm algorithm (SSA) [30]. PB algorithms rely on physical phenomena, including charged system search (CSS) [31] and lightning search algorithm (LSA) [32]. References [33,34] provide a detailed introduction and review of the parameter estimation methods for single-diode and two-diode power generation models. Table 1 summarizes the three-diode models reported in the literature. Most studies employed SI-based optimization algorithms to obtain accurate parameter estimates. As presented in [1,12], the three-diode model achieves improved estimation accuracy compared with the single-diode model, yielding an MRE reduction of approximately 0.017–0.040%. Relative to the two-diode model, the accuracy improvement ranges from 0.001% to 0.008%, depending on the optimization technique applied. Furthermore, References [1,12] expanded the parameter set from nine to sixteen, enabling more refined estimates. A similar transformation is adopted in this study.
As shown in Table 1 and References [24,25,26,27,28,29,30,31,32], most prior studies employ either single or hybrid optimization methods for parameter estimation. Although multi-agent optimization algorithms have shown promising results, various optimization algorithms with distinct search behaviors often yield diverse estimates under varying weather conditions, parameter drift, and PV modules degradation after long-term operation. To address these limitations, this study proposes a deep learning-based ensemble framework that combines convolutional neural networks (CNNs) and long short-term memory (LSTM) networks [35] to integrate the estimated outputs from distinct optimization algorithms. The key contributions are as follows:
  • Estimation results using the three-diode model are compared with those from other diode models to verify its accuracy.
  • Multiple optimization algorithms, including EC, SI, and PB methods, are employed to generate complementary estimates, thereby enhancing ensemble effectiveness and mitigating overfitting.
  • A hybrid CNN–LSTM architecture is developed to improve predictive accuracy and stability in three-diode parameter estimation.
  • As the output power of PV systems decreases due to degradation and parameter drift, the proposed ensemble method integrates multiple optimization algorithms to enhance prediction accuracy, thereby improving the reliability of practical deployments.
The remainder of this paper is organized as follows. Section 2 introduces the three-diode model. Section 3 presents the proposed parameter estimation methodology. Section 4 reports the simulation results for a 733 kW PV power generation system. Section 5 provides discussions, and Section 6 concludes the paper.

2. The Three-Diode Model

Figure 1 illustrates the configuration of a three-diode model. The output current ( I p v ) can be expressed as:
I p v = I p h I s a t 1 exp V p v + I p v R s n i d l 1 V T 1 I s a t 2 exp V p v + I p v R s n i d l 2 V T 1 I s a t 3 exp V p v + I p v R s n i d l 3 V T 1 V p v + I p v R s R h ,
where I p h is the photocurrent, R s is the series resistance, R s h is the parallel resistance, I s a t 1 , I s a t 2 a n d I s a t 3 are the dark saturation currents,   n i d l 1 ,     n i d l 2 ,   n i d l 3 are the ideality factors, V p v is the output voltage, and V T is the thermal voltage which is varied with the absolute temperature.
Since the open-circuit voltage of a solar cell is approximately 0.6 V, multiple solar cells are connected in series to form a solar module. These modules are further arranged in series and parallel to construct a PV array, thereby achieving the desired voltage, current, and power output, as shown in Figure 2. The output current for a PV array can be rewritten from (1) as:
I p v = N p I p h N p I s a t 1 exp V p v + I p v R s N s / N p N s n i d l 1 V T 1 N p I s a t 2 exp V p v + I p v R s N s / N p N s n i d l 2 V T 1 N p I s a t 3 exp V p v + I p v R s N s / N p N s n i d l 3 V T 1 V p v + I p v R s N s / N p R h ,
where N s and N p denote the number of modules connected in series and in parallel, respectively.
The aforementioned nine parameters ( I p h , I s a t 1 , I s a t 2 ,   I s a t 3 ,   n i d l 1 ,   n i d l 2 ,   n i d l 3   R s , and R s h ) directly influence the output characteristics of the PV power generation system. However, these key parameters ( I p h , I s a t 1 ,   n i d l 1 , R s , and R s h ) are in turn affected by the short-circuit current ( I s c ), open-circuit voltage ( V o c ), and the voltage ( V m ) and current ( I m ) at the MPP as follows [2,3,4,5]:
I p h = 1 + R s R s h I s c ,
I s a t 1 = I s c V o c R s h exp V o c n i d l 1 V T ,
n i d l 1 = I s a t 1 V m + I m R s I m e x p V m + I m R s n i d l 1 V T / V T ,
R s = I s a t 1 V m + I m R s I m e x p V m + I m R s n i d l 1 V T l n I s c I m I s a t 1 V m I m ,
R s h = I s c I p h I s c R s .
To simplify the parameter extraction process, Ishaque et al. [9] assumed I s a t 1   =   I s a t 2   = I s a t 3 , and based on the Shockley diode diffusion theory [36], set n i d l 1 1.2 , and selected n i d l i 2 > 1.2 to achieve the best estimation results.
Moreover, the manufacturer-provided specifications ( I s c , V o c , V m , and I m ) vary with solar irradiance and module temperature as shown below:
I s c = G G r I s c r + β I s c T T r ,
V o c = V o c r + β V o c T T r ,
V m = V m r + β V o c T T r ,
I m = G G r I m r + β I s c T T r ,
where I s c r , V o c r , V m r , and I m r represent the values under STC, with irradiance of 1000 W/m2, temperature of 298 K, and air mass of 1.5. β I s c and β V o c   denote temperature coefficients for short-circuit current and open-circuit voltage, respectively.
To improve the estimation accuracy of the three-diode model under varying irradiance and temperature conditions, the original nine parameters are extended to 14, as conducted from Equation (3) to Equation (11). The parameter control framework is illustrated in Figure 3, where these 14 parameters act as control variables for determining the nine key parameters. Optimization algorithms are then employed to identify the optimal set of control variables. Once obtained, the output current and power are computed using Equation (2).

3. The Proposed Method

Figure 4 illustrates the overall framework of the proposed ensemble learning approach, which comprises four stages: (i) data preprocessing (Stage 1), (ii) model establishment (Stage 2), (iii) preliminary estimation using individual optimization algorithms (Stage 3), and (iv) final estimation with the proposed ensemble method (Stage 4). The following subsections provide detailed descriptions of these stages.

3.1. Data Preprocessing

3.1.1. Missing Data Compensation

Missing data are compensated using interpolation or regression techniques. If more than 50% of the data for a given day are missing, the data from that day are excluded from further analysis.

3.1.2. Outlier Removal

Outlier detection and removal are essential for improving data quality. Several approaches exist, including the interquartile range (IQR), median absolute deviation (MAD), classification models, clustering-based methods, and the Z-score. In this study, the Z-score method is employed, as it provides a normalized measure of deviation from the mean, defined as:
Z = x μ σ     ,
where x is a data point, μ is the mean, and σ is the standard deviation. A higher Z value indicates a greater deviation from the mean.

3.1.3. Data Smoothing

Data smoothing reduces noise and emphasizes long-term trends in time series analysis. The moving average method is applied, defined as:
M A t ( n ) = 1 n w i = t n + 1 t x i ,
where n w is the moving window size and M A t ( n ) is the moving average at time t.

3.1.4. Feature Selection

To select the most relevant input variables for PV power generation, the Pearson correlation coefficient (PCC) is calculated as:
r = x y x y n d x 2 x 2 n d y 2 y 2 n d ,
where x is the weather variable, y is the PV power output, n d is the number of samples, and r is the correlation coefficient.

3.1.5. Data Classification

K-means clustering [37] is employed to partition n data point (x1, x2, …, xn) into k clusters by minimizing the sum of squared Euclidean distances between each point and its assigned cluster center as:
a r g   m i n = i = 1 k j = 1 n w j i X j R i 2 ,
where X j is the jth observation, R i is the ith cluster center, w j i is the weight, and   is the Euclidean distance. R i and w j i can be expressed individually as follows:
R i = j = 1 n w j i X j j = 1 n w j i ,
w j i = 1 , i f   X j R i   X j R m ,   m i 0 , o t h e r w i s e ,
After data compensation, abnormal data removal, data smoothing, feature extraction, and data classification, a three-diode power generation model can be established based on the manufacturer’s parameters and historical data, including PV power, solar irradiance, and module temperature. Since the manufacturer typically provides only four parameters ( I s c , V o c , V m , and I m ), the remaining unknown parameters must be estimated using Newton–Raphson method. Once all nine parameters are obtained, an optimization algorithm can be applied to perform a more comprehensive search for parameter solutions. The following section introduces five optimization algorithms with distinct characteristics.

3.2. Optimization Algorithms

As shown in Figure 4, five optimization algorithms are used to produce individual outputs, including the Newton–Raphson method, particle swarm optimization (PSO), multiverse optimizer (MVO), evolutionary strategy (ES), and atom search optimization (ASO). These algorithms were selected for their diverse characteristics, which provide complementary estimates, thereby enhancing ensemble effectiveness and reducing the risk of overfitting. Specifically, the Newton–Raphson method was employed as a single-agent approach for parameter searching, while PSO is categorized under swarm intelligence (SI), MVO and ASO under population-based (PB) methods, and ES under evolutionary computation (EC). This diversity ensures that the proposed framework can be generalized to a wide range of scenarios by leveraging the strengths of different algorithmic classes. A detailed description of the five optimization algorithms is provided in the following subsection.

3.2.1. Newton–Raphson Method

The Newton–Raphson method [20] employs a single agent to search for parameter solutions. Let the parameter vector be defined as
ω = I s c r ,     V o c r ,   I m r ,     V m r ,   I s a t 1 ,   n i d l 1 ,   n i d l 2 ,   n i d l 3 , G r ,   T r ,   β I S C ,   β V O C ,   R s ,   R s h ,
The solution procedure is outlined as follows:
1.
Parameter initialization: Define the admissible range for each parameter.
2.
Fitness definition: Let x 1 = ω , and formulate the objective function f V p v , I p v , ω     as:
f V p v , I p v , ω   = I p v I p h { I s a t 1 exp V p v + I p v R s n i d l 1 V T 1 N p I s a t 2 exp V p v + I p v R s n i d l 2 V T 1 N p I s a t 3 exp V p v + I p v R s n i d l 3 V T 1 V p v + I p v R s R h } = 0 ,
where f V p v , I p v , ω   = 0   is a nonlinear recursive function, and its fitness value is expressed as:
f M A E = 1 N i = 1 N f V p v , I p v , ω   ,
where N is the number of data points, MAE is the mean absolute error, which is less sensitive to outliers.
3.
Initialization: Randomly generate an initial value x 1 ( 0 ) , and evaluate both f ( x 1 ( 0 ) ) and its derivative f ( x 1 ( 0 ) ) .
4.
Parameter update: Update the parameter values according to Equation (21):
x 1 ( 1 ) = x 1 ( 0 ) f ( x 1 ( 0 ) ) f ( x 1 ( 0 ) )   ,
5.
Convergence check: Evaluate f ( x 1 ( 1 ) ) . If the value falls below a predefined threshold, the algorithm terminates and the converged x 1 is recorded. Otherwise, repeat Step 4.
6.
Output calculation: Once convergence is achieved, compute the output current, voltage, and power using Equation (2).

3.2.2. Particle Swarm Optimization (PSO)

PSO is a population-based stochastic optimization algorithm introduced by James Kennedy and Russ Eberhart in 1995 [38]. It is inspired by the social behavior of birds flocking or fish schooling. PSO employs a swarm of particles (potential solutions) “fly” through the search space, adjusting their positions based on their own best position (personal experience) and the best positions found by the swarm (social learning) as follows:
v i t + 1 = w v i t + c 1 · r 1 × p i x i t + c 2 · r 2 × g t x i t ,  
x i t + 1 = x i t + v i t + 1 ,  
where x i t is the position vector of the ith particle, v i t   is the velocity vector of the ith particle, g ( t ) is the global best position, w is an inertia weight, c 1 and c 2 are the acceleration factors that control particles towards the optimal position. r 1 and r 2   [ 0 ,   1 ] are random numbers used to maintain group diversity.

3.2.3. Multiverse Optimizer (MVO)

MVO is a nature-inspired optimization algorithm introduced by Mirjalili et al. in 2016 [39]. It is inspired by the multiverse theory in physics, which includes three core phenomena: white holes, black holes, and wormholes. The solution steps using MVO are described as follows:
1.
Initialization: Randomly generate the initial population of universes (solutions) as follows:
ω i = x i 1 ,   x i 2 ,   . . . ,   x i d   f o r   i = 1 ,   2 ,   . . .   N p ,
where N p is the population size and d is the variable dimension. In this paper, d = 14 , representing the number of parameters to be estimated.
2.
Fitness evaluation: Compute the inflation rate (IR) of each universe, which is the inverse of the fitness value of Equation (20). IR is used for sorting and probability exchange.
3.
White- and black-hole exchange: Exchange variables between universes based on fitness ranking. If the inflation rate of the universe ω j   is high, transmit its information to other universes ω i through white holes. Information exchange is performed using the roulette wheel selection method, as follows:
x i k n e w = x j k , i f   r a n d < I R j x i k , o t h e r w i s e ,
where r a n d [ 0 ,   1 ] is a random number, I R j   is the selected inflation rate, and x i k is the parameter value of the kth dimension of the ith universe.
4.
Wormhole tunneling: Adjust each universe’s position toward the best-known universe with a certain probability. All universes have the opportunity to approach the optimal solution ω * through the wormhole. Equation (25) is updated as follows:
x i k n e w = x j k + T D R · ( u b k l b k ) · r a n d , i f   r a n d < 0.5 x i k , T D R · ( u b k l b k ) · r a n d , o t h e r w i s e ,
where u b k and l b k are the upper and lower bounds of the k dimension, respectively, T D R   is the traveling distance rate, controlling the magnitude of perturbation.
5.
Termination: Repeat step 2 until a maximum number of iterations or convergence criteria are met.

3.2.4. Evolution Strategies (ES)

ES is a class of black-box optimization algorithm inspired by the concept of natural evolution [40]. It is widely used in non-convex, non-differentiable, or noisy optimization problems, especially in high-dimensional continuous domains. The evolving steps using ES are described as follows:
1.
Initialization: Start with a randomly initialized population of individuals (solutions) as:
P ( 0 ) = x 1 ( 0 ) ,   x 2 ( 0 ) ,   . . . ,   x N ( 0 ) R d ,
where individual x 1 ( 0 ) R d is a d -dimensional vector.
2.
Mutation: Each individual is perturbed by adding a normally distributed noise to generate offspring:
x j = x i + ρ · ϵ j ,   ϵ j ~ N ( 0 , σ ) ,  
where ρ is the step size of mutation, and N ( 0 , σ ) is a standard normal distribution.
3.
Fitness evaluation: Calculate the fitness value of each offspring using Equation (20), as follows:
f ( x j ) ,   f o r   a l l   j 1 ,   2 ,   . . . , N    
4.
Selection: Sort all populations and select the top-performing individuals (based on fitness) to form the next generation, as follows:
P ( t + 1 ) = S e l e c t T o p μ x j j = 1 N .  
After the best performing individuals are selected, half of them ( μ ) remain in the population.
5.
Termination: Repeat step 2 until convergence or a maximum number of iterations are met.

3.2.5. Atom Search Optimization (ASO)

ASO is a metaheuristic optimization algorithm introduced by M. Ghaffari Hadigheh et al. in 2020 [41]. It is inspired by the motion and interaction of atoms based on principles from molecular dynamics. Described below are the steps using ASO method:
1.
Initialization: Randomly initialize the positions x i R D and velocity v i R D of all atoms as:
x i d = l b d + r a n d · u b d l b d , i = 1 ,   2 ,   . . . ,   N
2.
Fitness evaluation: Evaluate the fitness of each atom ( f i ) using Equation (20).
3.
Mass calculation: Convert fitness to mass as follows:
m i = f w o r s t f i f w o r s t f b e s t ϵ ,
where ϵ is the well depth of potential used to control attraction strength.
4.
Force calculation: Use the Lennard-Jones (LJ) potential to calculate forces between atoms i and j:
F i j = 24 · ϵ 2 δ r i j 13 δ r i j 7 · x i x j r i j ,
F i = j K b e s t ,   j i F i j ,
where r i j = x i x j is the interatomic distances and δ is a distance constant where the potential is zero if there is a balance between attraction and repulsion.
5.
Acceleration calculation: Apply Newton–Raphson’s second law to evaluate acceleration as follows:
a i = F i M i + ϵ ,
where M i   is the normalization of mass.
6.
Velocity and position update: Update velocity and position as follows:
v i ( t + 1 ) = r a n d · v i ( t + 1 ) + a i ,
x i ( t + 1 ) = x i ( t ) + v i ( t + 1 ) ,
7.
Boundary handling: Ensure that all position values remain within the defined search space:
x i d = m i n ( m a x ( x i d , l b d ) , u b d ) ,
8.
Termination: Repeat step 2 until convergence or a maximum number of iterations are met.
Table 2 shows the structured comparison between the Newton–Raphson method, PSO, MVO, ES, and ASO. After generating individual initial estimates, the proposed approach employs a hybrid CNN–LSTM model to integrate the outputs of the five aforementioned methods. Since these optimization algorithms exhibit different characteristics, CNNs are used to extract their features, while the LSTM network is employed to capture temporal dependencies, thereby providing more accurate and stable estimates.

3.3. Deep Learning-Based Ensemble Method

The proposed hybrid CNN–LSTM operates as an ensemble framework that aggregates the outputs of various optimization algorithms to generate the final estimates. Specifically, it serves a dual role: functioning as a feature extractor and as a weight allocator, thereby assigning appropriate significance to each optimization algorithm. CNNs are utilized to extract informative temporal features, while LSTM is incorporated to mitigate the gradient vanishing issue in sequence learning and to capture long-term dependencies. Introduced below are the CNNs and LSTM networks.

3.3.1. CNNs

The concept of CNNs was first introduced by Yann LeCun et al. in 1998 [42] for handwritten digit recognition. A CNN typically consists of convolutional layers, pooling layers, and fully connected layers, as illustrated in Figure 5. A brief description of each layer is provided below:
(i)
Convolutional layers:
  • Serve as the fundamental building blocks of CNNs.
  • Utilize learnable filters (kernels) to convolve across the input, capturing local spatial or temporal dependencies.
  • Early layers primarily extract low-level features, while deeper layers capture more abstract and high-level representations.
(ii)
Pooling layers:
  • Reduce the dimensionality of spatial or temporal features while preserving essential information.
  • Enhance translation invariance and decrease computational complexity.
(iii)
Fully connected layers:
  • Transform the extracted features from convolutional and pooling layers into a one-dimensional vector.
  • Enable high-level reasoning at the final stage, supporting tasks such as classification, regression, or other predictions.
CNNs exhibit strong capabilities in automatic feature extraction and translation invariance, enabling the recognition of features irrespective of their position [43]. However, CNNs alone are limited in modeling long-term temporal dependencies. To address this limitation, LSTM is integrated in this work to complement CNNs and enhance sequence learning.

3.3.2. LSTM

LSTM was first introduced by Sepp Hochreiter and Jürgen Schmidhuber in 1997 [44]. It was designed as a solution to improve the vanishing gradient problem in recurrent neural networks (RNNs). LSTM has been extensively applied to time-series forecasting, speech recognition, and anomaly detection. The overall structure of an LSTM network is illustrated in Figure 6, where the inputs are derived from the outputs of CNNs.
An LSTM unit comprises four key components: input gate ( i t ), memory cell ( c ~ t ), forget gate ( f t ), and output gate ( O t ). These elements collectively regulate the flow of information, which can be mathematically formulated as follows:
i t = σ ( W i x t + U i h t 1 + b i ) ,
c ~ t = t a n h ( W c x t + U c h t 1 + b c ) ,
f t = σ ( W f x t + U f h t 1 + b f ) ,
o t = σ ( W o x t + U o h t 1 + b o ) ,
At each time step, these gates determine how much of the past information should be retained, how much new information should be incorporated, and how much should be output. The sigmoid ( σ ) and tanh ( φ ) activation functions are used to modulate these flows. In this study, LSTM is utilized to enhance temporal modeling capabilities of CNNs.

4. Numerical Results

4.1. Data Preprocessing Results

The proposed method was evaluated using a 733 kW PV power generation system. Data were collected from July 2021 to July 2022 at an hourly resolution, including historical PV power output, solar irradiance and module temperature. Outliers were removed using the Z-score method, and the remaining dataset was divided into 75% for training and 25% for testing. The choice of a 75%/25% split was based on the availability of data for each weather condition. Although an 80%/20% split is commonly adopted, the testing data corresponding to rainy days were relatively limited after classification in this study. Therefore, a 75%/25% split was employed to ensure sufficient representation for both training and testing.
During training, solar irradiance and module temperature were used as input variables, while PV power generation served as the output. In the testing stage, predicted global horizontal irradiance (GHI) and air temperature were employed as inputs, as they exhibit stronger correlation with PV output based on the PCC method. The forecasted weather data were obtained from the SOLCAST platform [45]. All experiments were implemented in Python 3.9 on a Windows 11 platform equipped with an Intel Core i7-10700 CPU, 32 GB RAM. Estimation accuracy was evaluated using the mean relative error (MRE), mean absolute Error (MAE), normalized mean absolute error (NMAE), and symmetric mean absolute percentage error (sMAPE), defined as:
M R E = 1 N m i = 1 N m P i P ^ i P c a p × 100 % ,
M A E = 1 N m i = 1 N m P i P ^ i ,
N M A E = M A E P ¯ ,
s M A P E = 1 N m i = 1 N m P i P ^ i P i + P ^ i 2 × 100 % ,
where P c a p is the rated PV capacity, P i is the actual power, P ^ i is the estimated power, P ¯ is the mean of actual power, and N m denotes the number of testing samples. As indicated in (43)–(46), MRE reduces the influence of low-power generation levels on the error percentage. MAE quantifies the average magnitude of the errors regardless of sign. NMAE scales MAE by the mean value, making the error relative to the data scale. The sMAPE metric is a percentage-based symmetric error that mitigates the issue of inflated error values when actual power approaches zero.
The Z-score method was adopted to eliminate outliers. After normalization, the mean value ( μ ) was set to 0 and the standard deviation ( σ ) to 1. Data points with Z-scores greater than 1.5 were considered outliers and subsequently removed.
The elbow curve shown in Figure 7 was derived using k-means clustering, with the sum of squared errors (SSE) expressed as:
S S E = i = 1 N T P i P ^ i 2 ,
where N T is the number of training samples. When outliers are removed, the optimal number of clusters was determined using SSE. As illustrated in Figure 7, SSE decreases rapidly until K = 5, after which the reduction becomes marginal. Therefore, the data were classified into five clusters (Figure 8), corresponding to sunny, partly cloudy, overcast, heavily overcast, and rainy conditions. Note that when the data were divided into five classes, the continuity of the monthly time series was broken, and only daily time series could be preserved. The advantages of this classification are threefold: (i) it avoids dominance by the majority class (e.g., sunny days) while maintaining performance across rare but critical conditions (e.g., rainy days); (ii) it enables operators to identify which weather conditions cause the larger prediction errors; and (iii) it facilitates decision-making in energy management and fault diagnosis by clarifying the class membership of each data segment.

4.2. Estimation Results of Different Diode Models

Table 3 summarizes the parameter ranges used in the optimization process. Figure 9 presents the optimization performance of various diode models using the PSO method, with the maximum number of iterations and population size set to 100 and 50, respectively. Across 100 iterations, the three-diode model achieves a lower mean absolute error (MAE) than the single- and two-diode models. Figure 10 further demonstrates the convergence behavior of PSO over 10 independent runs, showing stable and consistent convergence. Table 4 compares the estimation performance of the three diode models using PSO, where the average estimation errors for the single-, two-, and three-diode models are 4.049%, 4.005%, and 3.976%, respectively. Although the accuracy improvement of the three-diode model over the two-diode model is relatively small (~0.03%), even such a minor reduction can result in meaningful energy prediction gains for large-scale PV systems. For instance, in a 733 kW PV plant, a 0.03% improvement corresponds to approximately 550 kWh annually (based on 2500 h per year). Such an improvement is non-negligible for long-term PV deployment and energy yield estimation. It should also be noted that, while the three-diode model provides slightly higher estimation accuracy, it incurs longer computation time after 100 iterations compared with the single- and two-diode models.

4.3. Preliminary Estimation Results

Building upon the comparative analysis of diode models in Section 4.2, preliminary estimation results are obtained using five optimization algorithms. The parameter settings of these algorithms, which are primarily determined based on prior studies [38,39,40,41] and refined through manual step-by-step adjustment to achieve stable convergence and reliable estimation results under the given dataset, are summarized in Table 5. These choices were guided by commonly adopted ranges in the literature to ensure fairness and reproducibility. The internal parameter values obtained for each algorithm are reported in Table 6. The estimation curves under various weather conditions are illustrated in Figure 11. Across all methods, estimation errors are generally higher on rainy days due to increased fluctuations in weather conditions. Overall, the Newton–Raphson method exhibits consistently lower accuracy than the metaheuristic algorithms. A comparative summary of the five optimization approaches is provided in Table 7, highlighting their respective strengths and weaknesses under different weather conditions. Specifically, the MVO achieves the best estimation performance on rainy and heavily overcast days while avoiding the worst performance under any weather condition. This robustness can be attributed to the inherent mechanism of MVO, which preserves population diversity through multi-universe exploration. Such a mechanism mitigates the risk of premature convergence and enhances resilience against random noise in the input data.
ES delivers the most accurate results on overcast days but performs poorly on partly cloudy and sunny days. ASO achieves the best accuracy on sunny days but produces the least reliable results on heavily overcast days. By contrast, the Newton–Raphson method demonstrates the most unstable performance overall, producing the best results on partly cloudy days but ranking worst under three of the four weather conditions. Newton–Raphson relies heavily on precise derivative information. In noisy or degraded PV conditions, inaccurate gradient estimation amplifies errors during iterative updates, leading to unstable convergence. Hence, its sensitivity is fundamentally associated with noise amplification rather than PV degradation itself.

4.4. Final Estimation Results

For the final estimation stage, an ensemble architecture integrating CNNs and LSTM networks is employed, aiming to enhance both feature extraction and temporal dependency modeling, thereby improving overall prediction accuracy. The parameter settings for CNNs and LSTM are provided in Table 8. The CNN–LSTM consists of three CNN layers (Conv2D, MaxPooling2D, and Flatten), one LSTM layer with 64 units, and two fully connected Dense layers (32 and 1 units, respectively). Relu was selected for CNN and intermediate Dense layers to mitigate the vanishing-gradient problem, while the output layer employed a linear activation. The CNNs function as a pre-projection layer before the LSTM, receiving five inputs from individual optimization algorithms and linearly projecting 16 features to the LSTM for efficient training. While LSTM models typically adopt the tanh activation function, the Relu activation function was used in this study to achieve better estimation performance.
To evaluate the sensitivity of hyperparameters in the CNN–LSTM, an additional experiment was conducted by varying batch size, optimizer, and learning rate within a reasonable range while keeping the other parameters fixed. The results, illustrated in Figure 12, indicate that the optimal settings for batch size, optimizer, and learning rate are 25, Adam, and 0.001, respectively. Furthermore, the number of timesteps was set to 5. To mitigate overfitting, the dropout rate was set to 0.3 for the CNN layer and 0.5 for the LSTM layer. The patience parameter for early stopping was fixed at 5 epochs, and the L2 regularization coefficient was set to 0.0001.
Table 9 compares the proposed ensemble method against several optimization algorithms under various weather conditions. The LSTM alone achieves the best estimation on sunny days but performs worst on partly cloudy days. By contrast, the proposed ensemble method delivers the most accurate predictions on rainy and heavily overcast days, although its performance improvement is relatively small under other weather scenarios. Overall, in terms of average MRE, the proposed approach achieves a lower estimation error compared with all other optimization algorithms. The source code of the proposed CNN–LSTM model and the baseline methods is available in [46,47].
Table 10, Table 11, Table 12 and Table 13 present the comparative results of various methods under the proposed ensemble framework in terms of MRE, MAE, NMAE, and sMAPE. Baseline models include single approaches such as extreme gradient boosting (XGBoost), light gradient boosting machine (LightGBM), gated recurrent unit (GRU), CNN, and LSTM, as well as hybrid models including attention LSTM (A–LSTM) and CNN-A-LSTM. As shown in Table 10, the proposed CNN–LSTM method achieves the lowest average error (3.968%), marginally outperforming the other approaches and demonstrating robustness across diverse weather conditions. Nevertheless, the hybrid model requires substantially higher training time compared with tree-based models (XGBoost and LightGBM) and other deep learning models (GRU and CNN), indicating a trade-off between computational efficiency and accuracy. Furthermore, Table 11 further shows that CNN–LSTM attains the best performance with the lowest average MAE (29.13 kW), slightly surpassing the benchmark methods. The results of NMAE in Table 12 are consistent with those of MAE. Moreover, Table 13 indicates that large estimation errors occur on rainy days due to greater weather variability; nevertheless, the proposed CNN–LSTM still outperforms the other baselines.

5. Discussions

The experimental findings lead to the following key observations:
1.
The three-diode model provides superior estimation accuracy compared with the single- and two-diode models, albeit with a longer computation time (~12 min). Nevertheless, this additional computational burden does not hinder its applicability for fault detection in PV systems.
2.
As presented in Table 4, the optimization results indicate a decrease in parallel resistance and an increase in series resistance compared with the Newton–Raphson method. The Newton–Raphson method is employed as a benchmark because the manufacturer typically provides only four parameters ( I s c , V o c , V m , and I m ). The remaining unknown parameters are estimated using the Newton–Raphson method. Therefore, the parameters obtained via the Newton–Raphson method are regarded as the baseline solutions derived from the manufacturer’s data. This outcome suggests possible PV module oxidation and degradation [48,49], thereby requiring maintenance to restore performance.
3.
The Newton–Raphson method exhibits instability due to its gradient-based nature and sensitivity to initial values (Table 5). In contrast, multi-agent optimization approaches leverage multi-point search and parallelism to achieve more reliable solutions.
4.
The proposed CNN–LSTM ensemble method consistently achieves stable performance and yields lower MRE compared with standalone LSTM and other individual optimization algorithms (Table 7). Although the difference in mean MRE between CNN–LSTM (3.968%) and the best individual method PSO (3.976%) is relatively small (0.008 p.p.), this improvement translates into meaningful gains in energy prediction accuracy when applied to a 733 kW PV system. In practical deployment, it corresponds to approximately 146.6 kWh annually (based on 2500 h per year).

6. Conclusions

This study proposed a deep learning-based ensemble framework for accurate parameter estimation of solar cells using a three-diode model. By integrating the outputs of diverse optimization algorithms including Newton–Raphson, PSO, MVO, ES, and ASO through a hybrid CNN–LSTM network, the method effectively leverages complementary search behaviors to improve robustness and precision. The main findings are summarized as follows:
1.
Enhanced model accuracy: Compared with single- and two-diode models, the three-diode structure demonstrated superior capability in capturing nonlinear and physical effects, thereby improving estimation reliability.
2.
Robust ensemble performance: The proposed ensemble method consistently outperformed tree-based models (XGBoost and LightGBM) and other deep learning models (GRU, CNN, and LSTM) across all evaluation metrics, including MRE, MAE, NMAE, and sMAPE.
3.
Deep learning integration: The CNN–LSTM architecture successfully combined feature extraction and temporal sequence learning, mitigating overfitting while delivering more stable parameter estimates.
4.
Practical validation: Experimental results on a 733 kW PV power system indicated a decrease in parallel resistance and an increase in series resistance compared with the Newton–Raphson method. The Newton–Raphson method is employed as a benchmark since it relies on manufacturer-provided parameters to generate the estimated results. This outcome suggests possible PV module oxidation and degradation, thereby requiring maintenance to restore performance.
5.
Deployment aspect: Although the difference in mean MRE between the CNN–LSTM and the best-performing individual method (PSO) is relatively small (0.008%), this improvement translates into meaningful practical benefits. In particular, it yields an additional energy prediction gain of approximately 146.6 kWh per year in a 733 kW PV system, underscoring the practical significance of the proposed framework, especially when applied to larger PV systems.
6.
Scalability aspect: The proposed method has been validated on a 733 kW PV system and can be readily extended to larger-scale PV power generation systems. With appropriate parameter tuning, such as adjustments to the number of parallel and series modules, the proposed framework can be adapted to different system configurations, thereby ensuring its applicability to a wide range of deployment scenarios.
In conclusion, the proposed ensemble approach provides a promising solution for solar cell parameter estimation by integrating the strengths of physical modeling, optimization algorithms, and deep learning. Future work will focus on extending the framework to larger PV arrays, incorporating statistical validation (e.g., confidence intervals or Diebold–Mariano tests), analyzing the sensitivity of optimization parameters, and performing a quantitative analysis of deployment costs in practical PV operation and maintenance. Moreover, the adaptive parameter threshold requires further investigation, particularly under long-term PV operation with environmental fluctuations.

Author Contributions

This paper is a collaborative work by all the authors. Conceptualization, C.-M.H. and S.-P.Y.; methodology, C.-M.H. and S.-P.Y.; software, F.-R.S. and C.-H.C.; validation, C.-M.H. and S.-J.C.; writing—original draft preparation, C.-M.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science and Technology Council, Taiwan, under grant no. NSTC 114-2221-E-168-001.

Data Availability Statement

There is no statement of the availability of data and materials.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
A–LSTMAttention LSTM
ASOAtom Search Optimization
BOBonobo Optimizer
CNNsConvolutional Neural Networks
CSSCharged System Search
DEDifferential Evolution
ECEvolutionary Computing
ESEvolutionary Strategy
FPAFlower Pollination Algorithm
GAGenetic Algorithm
GHIGlobal Horizontal Irradiance
GRUGated Recurrent Unit
IQRInterquartile Range
LightGBMLight gradient boosting machine
LSTMLong Short-Term Memory
LSALightning Search Algorithm
MADMedian Absolute Deviation
MAEMean Absolute Error
MPPMaximum Power Point
MREMean Relative Error
MVOMultiverse Optimizer
PBPhysics-Based method
PCCPearson Correlation Coefficient
PSOParticle Swarm Optimization
PVPhotovoltaic
RNNsRecurrent Neural Networks
SISwarm Intelligence
SSASalp Swarm Algorithm
SSESum of Squared Errors
STCStandard Test Condition
WOAWhale Optimization Algorithm
XGBoostExtreme gradient boosting

References

  1. Huang, C.M.; Chen, S.J.; Yang, S.P.; Huang, Y.C.; Huang, P.Y. Parameter Estimation and Preliminary Fault Diagnosis for Photovoltaic Modules Using a Three-Diode Model. Energies 2024, 17, 3214. [Google Scholar] [CrossRef]
  2. González-Cagigal, M.A.; Rosendo-Macías, J.A.; Gómez-Expósito, A. Estimation of Equivalent Model of Photovoltaic Array Using Unscented Kalman Filters. J. Mod. Power Syst. Clean Energy 2024, 12, 819–827. [Google Scholar] [CrossRef]
  3. Kumar, M.; Panda, K.P.; Naayagi, R.T.; Thakur, R.; Panda, G. Enhanced Optimization Techniques for Parameter Estimation of Single-Diode PV Modules. Electronics 2024, 13, 2934. [Google Scholar] [CrossRef]
  4. Singla, M.K.; Gupta, J.; Nijhawan, P.; Singh, P.; Giri, N.C.; Hendawi, E.; Abu El-Sebah, M.I. Parameter Estimation Techniques for Photovoltaic System Modeling. Energies 2023, 16, 6280. [Google Scholar] [CrossRef]
  5. Cárdenas-Bravo, C.; Barraza, R.; Sánchez-Squella, A.; Valdivia-Lefort, P.; Castillo-Burns, F. Estimation of Single-diode Pho-tovoltaic Model Using the Differential Evolution Algorithm with Adaptive Boundaries. Energies 2021, 14, 3925. [Google Scholar] [CrossRef]
  6. Hakmi, S.H.; Alnami, H.; Moustafa, G.; Ginidi, A.R.; Shaheen, A.M. Modified Rime-Ice Growth Optimizer with Polynomial Differential Learning Operator for Single- and Double-Diode PV Parameter Estimation Problem. Electronics 2024, 13, 1611. [Google Scholar] [CrossRef]
  7. Elshara, R.; Gullari, A.H.; Rahebi, J.; Lopez-Guede, J.M. PV Cells and Modules Parameter Estimation Using Coati Optimization Algorithm. Energies 2024, 17, 1716. [Google Scholar] [CrossRef]
  8. Aalloul, R.; Elaissaoui, A.; Benlattar, M.; Adhiri, R. Emerging Parameters Extraction Method of PVModules Based on the Survival Strategies of Flying Foxes Optimization (FFO). Energies 2023, 16, 3531. [Google Scholar] [CrossRef]
  9. Ishaque, K.; Salam, Z.; Taheri, H. Simple, Fast and Accurate Two-Diode Model for Photovoltaic Modules. Sol. Energy Mater. Sol. Cells 2011, 95, 586–594. [Google Scholar] [CrossRef]
  10. Abdelminaam, D.S.; Alluhaidan, A.S.; Ismail, F.H.; El-Rahman, S. Parameters Extraction of the Three-Diode Photovoltaic Model Using Crayfish Optimization Algorithm. IEEE Access 2024, 12, 109342. [Google Scholar] [CrossRef]
  11. Nunes, H.; Pombo, J.; Mariano, S.; Calado, M.R. Photovoltaic Power Estimation of a Polycrystalline Silicon Module Under Partial Shading. In Proceedings of the IEEE International Conference on Environment and Electrical Engineering, Rome, Italy, 17–20 June 2024; pp. 1–6. [Google Scholar]
  12. Huang, C.M.; Huang, Y.C.; Chen, S.J.; Yang, S.P.; Huang, P.Y.; Chiu, C.H. Parameters Estimation of PV Modules for a Three-Diode Model Using an Enhanced Salp Swarm Algorithm. In Proceedings of the IEEE International Conference on Industrial Technology (ICIT), Bristol, UK, 25–27 March 2024; pp. 106–111. [Google Scholar]
  13. Dal, S.; Sezgin, N. Optimal Parameter Extraction of Triple-Diode Photovoltaic Model Using Frilled Lizard Optimization. In Proceedings of the IEEE Global Energy Conference (GEC), Batman, Turkiye, 4–6 December 2024; pp. 310–314. [Google Scholar]
  14. Gupta, J.; Hussain, A.; Singla, M.K.; Nijhawan, P.; Haider, W.; Kotb, H.; Kareem, M.; AboRas, M. Parameter Estimation of Different Photovoltaic Models Using Hybrid Particle Swarm Optimization and Gravitational Search Algorithm. Appl. Sci. 2023, 13, 249. [Google Scholar] [CrossRef]
  15. Gatla, V.R.; Injeti, S.K.; Kotte, S.; Polamarasetty, P.K.; Nuvvula, R.S.S.; Vardh, A.S.S.; Singh, M.; Khan, B. An Effective Approach for Extracting the Parameters of Solar PV Models Using the Chaotic War Strategy Optimization Algorithm with Modified Newton Raphson Method. IEEE J. Electron Devices Soc. 2023, 12, 849–858. [Google Scholar] [CrossRef]
  16. Qais, M.H.; Hasanien, H.M.; Alghuwainem, S.; Loo, K.H.; Elgendy, M.A.; Turky, R.A. Accurate Three-Diode Model Estimation of Photovoltaic Modules Using a Novel Circle Search Algorithm. Ain Shams Eng. J. 2022, 13, 101824. [Google Scholar] [CrossRef]
  17. Yousri, D.; Fathy, A.; Rezk, H.; Babu, T.S.; Berber, M.R. A Reliable Approach for Modeling the Photovoltaic System Under Partial Shading Conditions Using Three Diode Model and Hybrid Marine Predators-Slime Mould Algorithm. Energy Convers. Manag. 2021, 243, 114269. [Google Scholar] [CrossRef]
  18. Yousri, D.; Thanikanti, S.B.; Allam, D.; Ramachandaramurthy, V.K.; Eteiba, M.B. Fractional Chaotic Ensemble Particle Swarm Optimizer for Identifying the Single, Double, and Three Diode Photovoltaic Models’ Parameters. Energy 2020, 195, 116979. [Google Scholar] [CrossRef]
  19. Soliman, M.A.; Hasanien, H.M.; Alkuhayli, A. Marine Predators Algorithm for Parameters Identification of Triple-diode Photovoltaic Models. IEEE Access 2020, 8, 155832–155842. [Google Scholar] [CrossRef]
  20. Ayodele, T.R.; Ogunjuyigbe, A.S.O.; Ekoh, E.E. Evaluation of Numerical Algorithms Used in Extracting the Parameters of a Single-diode Photovoltaic Model. Sustain. Energy Technol. Assess. 2016, 13, 51–59. [Google Scholar] [CrossRef]
  21. De Blas, M.A.; Torres, J.L.; Prieto, E.; Garcia, A. Selecting a Suitable Model for Characterizing Photovoltaic Devices. Renew. Energy 2002, 25, 371–380. [Google Scholar] [CrossRef]
  22. Chegaar, M.; Ouennoughi, Z.; Foffmann, A. A New Method for Evaluating Illuminated Solar Cell Parameters. Solid-State Electron. 2001, 45, 293–296. [Google Scholar]
  23. Villalva, M.G.; Gazoli, J.R.; Filho, E.R. Comprehensive Approach to Modeling and Simulation of Photovoltaic Arrays. IEEE Trans. Power Electron 2009, 24, 1198–1208. [Google Scholar] [CrossRef]
  24. Yang, X.S. Flower Pollination Algorithm for Global Optimization. In International Conference on Unconventional Computing and Natural Computation; Springer: Berlin/Heidelberg, Germany, 2012; pp. 240–249. [Google Scholar]
  25. Ishaque, K.; Salam, Z. An Improved Modeling Method to Determine the Model Parameters of Photovoltaic (PV) Modules Using Differential Evolution (DE). Sol. Energy 2011, 85, 2349–2359. [Google Scholar] [CrossRef]
  26. Jervase, J.A.; Bourdoucen, H.; Al-Lawati, A. Solar Cell Parameters Extracting Using Genetic Algorithm. Meas. Sci. Technol. 2001, 12, 19–22. [Google Scholar] [CrossRef]
  27. Ye, M.; Wang, X.; Xu, Y. Parameter Extraction of Solar Cells Using Particle Swarm Optimization. J. Appl. Phys. 2009, 105, 094502. [Google Scholar] [CrossRef]
  28. Omnia, S.; Elazab, H.M.; Hasanien, M.A.E.; Abdeen, A.M. Parameters Estimation of Single- and Multiple-Diode Photovoltaic Model Using Whale Optimisation Algorithm. IET Renew. Power Gener. 2018, 12, 1755–1761. [Google Scholar]
  29. Al-Shamma’a, A.A.; Omotoso, H.O.; Alturki, F.A.; Farh, H.M.H.; Alkuhayli, A.; Alsharabi, K.; Noman, A.M. Parameter Estimation of Photovoltaic Cell Modules Using Bonobo Optimizer. Energies 2022, 15, 140. [Google Scholar] [CrossRef]
  30. Jhon, M.; Andres, F.T.M.; Andres, A.A.R.; Fabio, A.; Oscar, D.G.R.; Jose, M.P. Salp Swarm Optimization Algorithm for Estimating the Parameters of Photovoltaic Panels Based on the Three-Diode Model. Electronics 2021, 10, 3123. [Google Scholar] [CrossRef]
  31. Huang, Y.C.; Huang, C.M.; Chen, S.J.; Yang, S.P. Optimization of Module Parameters for PV Power Estimation Using a Hybrid Algorithm. IEEE Trans. Sustain. Energy 2020, 11, 2210–2219. [Google Scholar] [CrossRef]
  32. Reza, S.; Hussain, S. Parameter Extraction of Solar Cell Model Using the Lightning Search Algorithm with Different Weather. J. Sol. Energy Eng. 2016, 138, 041007–0410017. [Google Scholar] [CrossRef]
  33. Jordehi, A.R. Parameter estimation of solar photovoltaic (PV) cells: A review. Renew. Sustain. Energy Rev. 2016, 61, 354–371. [Google Scholar] [CrossRef]
  34. Chin, J.; Salam, Z.; Ishaque, K. Cell modelling and model parameters estimation techniques for photovoltaic simulator application: A review. Appl. Energy 2015, 154, 500–519. [Google Scholar] [CrossRef]
  35. Ullah, K.; Ahsan, M.; Hasanat, S.M.; Haris, M.; Yousaf, H.; Raza, S.F.; Tandon, R.; Abid, S.; Ullah, Z. Short-term load forecasting: A comprehensive review and simulation study with CNN-LSTM Hybrids Approach. IEEE Access 2024, 12, 111858–111881. [Google Scholar] [CrossRef]
  36. San, C.T.; Noyce, R.N.; Shockley, W. Carrier Generation and Recombination in P–N Junctions and P–N Junction Characteristics. Proc. IRE 1957, 45, 1228–1243. [Google Scholar]
  37. Lloyd, S.P. Least squares quantization in PCM. IEEE Trans. Inform. Theory 1982, 28, 129–137. [Google Scholar] [CrossRef]
  38. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  39. Mirjalili, S.; Mirjalili, S.M.; Hatamlou, A. Multi-Verse Optimizer: A Nature-Inspired Algorithm for Global Optimization. Neural Comput. Appl. 2016, 27, 495–513. [Google Scholar] [CrossRef]
  40. Salimans, T.; Ho, J.; Chen, X.; Sutskever, I. Evolution Strategies as a Scalable Alternative to Reinforcement Learning. arXiv 2017, arXiv:1703.03864. [Google Scholar] [CrossRef]
  41. Ghaffari Hadigheh, M.; Mirjalili, S.; Faris, H. Atom Search Optimization (ASO): A Physics-Inspired Metaheuristic Algorithm. Future Gener. Comput. Syst. 2020, 112, 1180–1198. [Google Scholar]
  42. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-Based Learning Applied to Document Recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  43. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inform. Process. Syst. 2012, 25, 1–9. [Google Scholar] [CrossRef]
  44. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  45. SOLCAST. Available online: https://solcast.com/ (accessed on 1 June 2025).
  46. The Source Code of the Proposed CNN–LSTM Method Is Available on GitHub. Available online: https://github.com/sfr0604/CNN-LSTM_TDM_Model (accessed on 8 September 2025).
  47. The Baseline Implementations Are Accessible Through the Open-Source Libraries in Python. Available online: https://mealpy.readthedocs.io/en/latest/ (accessed on 8 September 2025).
  48. Van Dyk, E.E.; Meyer, E.L. Analysis of the Effect of Parasitic Resistances on the Performance of Photovoltaic Modules. Renew. Energy 2004, 29, 333–344. [Google Scholar] [CrossRef]
  49. Sze, S.M. Physics of Semiconductor Devices; John Wiley and Sons: Hoboken, NJ, USA, 1981; pp. 812–816. [Google Scholar]
Figure 1. Three-diode equivalent circuit model for a PV module.
Figure 1. Three-diode equivalent circuit model for a PV module.
Electronics 14 03790 g001
Figure 2. Physical configuration of a PV array.
Figure 2. Physical configuration of a PV array.
Electronics 14 03790 g002
Figure 3. Parameter control framework for PV array modeling.
Figure 3. Parameter control framework for PV array modeling.
Electronics 14 03790 g003
Figure 4. Schematic diagram of the proposed deep learning-based ensemble method.
Figure 4. Schematic diagram of the proposed deep learning-based ensemble method.
Electronics 14 03790 g004
Figure 5. Architecture of CNNs.
Figure 5. Architecture of CNNs.
Electronics 14 03790 g005
Figure 6. Structure of an LSTM network.
Figure 6. Structure of an LSTM network.
Electronics 14 03790 g006
Figure 7. The elbow curve for training data.
Figure 7. The elbow curve for training data.
Electronics 14 03790 g007
Figure 8. PV power data classification using the k-means method. (A cycle represents the clustering center).
Figure 8. PV power data classification using the k-means method. (A cycle represents the clustering center).
Electronics 14 03790 g008
Figure 9. Optimization process for different diode models using the PSO method.
Figure 9. Optimization process for different diode models using the PSO method.
Electronics 14 03790 g009
Figure 10. Convergence performance of PSO using 10 different runs.
Figure 10. Convergence performance of PSO using 10 different runs.
Electronics 14 03790 g010
Figure 11. Comparison of estimation curves for different weather conditions: (a) Rainy days on 4 August, (b) Heavily overcast days on 12 November, (c) Overcast days on 16 January, (d) Partly cloudy days on 12 February; (e) Sunny days on 30 April.
Figure 11. Comparison of estimation curves for different weather conditions: (a) Rainy days on 4 August, (b) Heavily overcast days on 12 November, (c) Overcast days on 16 January, (d) Partly cloudy days on 12 February; (e) Sunny days on 30 April.
Electronics 14 03790 g011aElectronics 14 03790 g011b
Figure 12. Sensitivity analysis of hyperparameters in the CNN-LSTM model: (a) Batch size, (b) Optimizer, (c) Learning rate.
Figure 12. Sensitivity analysis of hyperparameters in the CNN-LSTM model: (a) Batch size, (b) Optimizer, (c) Learning rate.
Electronics 14 03790 g012
Table 1. Comparison of the three-diode models reported in the literature.
Table 1. Comparison of the three-diode models reported in the literature.
ReferenceNo. of
Parameter
Optimization Algorithm
ECSIPBNewton
[1] *16 √ **
[10]9
[11]9
[12] *16
[13]9
[14]9
[15]9
[16]9
[17]9
[18]9
[19]9
*: Expand nine parameters to 16 parameters for more precise estimates. In this study, the nine parameters are transformed into 14 parameters. **: A check mark indicates that the reference belongs to that category of optimization method.
Table 2. Comparison between the Newton–Raphson method, MVO, ES, PSO, and ASO.
Table 2. Comparison between the Newton–Raphson method, MVO, ES, PSO, and ASO.
MethodCharacteristicsAdvantagesDisadvantages
Newton–Raphson Method [20]
  • Gradient-based optimization.
  • Quadratic convergence near optimum.
  • Very fast convergence if initial guess is close.
  • Accurate for smooth, differentiable problems.
  • Sensitive to initial guess.
  • Not suitable for non-differentiable or highly non-convex problems.
PSO [38]
  • Population-based approach inspired by swarm intelligence.
  • No gradient required.
  • Fast convergence in early iterations.
  • Simple to implement, few parameters.
  • May trap in local optima.
  • Performance degrades in very high-dimensional search spaces.
MVO [39]
  • Nature-inspired metaheuristic approach.
  • Balances exploration and exploitation via inflation rate.
  • Good exploration ability in complex search spaces.
  • Few control parameters.
  • Convergence may be slow for high-dimensional problems.
  • Lacks strong theoretical convergence guarantee.
ES [40]
  • Population-based approach inspired by biological evolution.
  • No gradient needed.
  • Can escape local optima.
  • Works well for continuous optimization.
  • High computational cost.
  • Parameter tuning affects performance.
ASO [41]
  • Metaheuristic approach inspired by the motion and interaction of atoms.
  • Balance between global and local search.
  • Strong global search ability.
  • Effective on high-dimensional problems.
  • Parameter tuning required for best performance.
  • May require more computation than simpler heuristics.
Table 3. The range of parameter settings used for optimization process.
Table 3. The range of parameter settings used for optimization process.
ParameterRangeParameterRange
Short-circuit current, Iscr (A)3.60~5.40Current temperature coefficient, βIsc (A/K)0.000670~0.000690
Open-circuit voltage, Vocr (V)61~69Voltage temperature coefficient, βVoc (V/K)−0.166~−0.05
Current at MPP, Imr (A)50~55Ideality factor (1st diode), nidl11.0~1.2
Voltage at MPP, Vmr (V)3.30~3.65Ideality factor (2nd diode), nidl21.2~2.0
Irradiance under STC, Gr (W/m2)1000~1050Ideality factor (3rd diode), nidl31.2~3.0
Temperature under STC, Tr (K)290~310Saturation current (1st diode), Isat1 (A)1.16 × 10 15 ~1.16 × 10 7
Parallel resistance under STC, Rshr (Ω)60~180Saturation current (2nd diode), Isat2 (A)1.16   × 10 15 ~1.16 × 10 7
Series resistance under STC, Rsr (Ω)2.0~6.0Saturation current (3rd diode), Isat3 (A)1.16   × 10 15 ~1.16   × 10 7
Table 4. Comparison of the three different diode models using the PSO method.
Table 4. Comparison of the three different diode models using the PSO method.
ModelsSingle-DiodeTwo-DiodeThree-Diode
MRE4.049%4.005%3.976%
Calculation time (s)460.16683.77743.76
Table 5. Parameter settings of different optimization algorithms.
Table 5. Parameter settings of different optimization algorithms.
AlgorithmParameterValue
PSO Inertia   weight   ( w )0.4
Acceleration   factor   ( c 1 )0.5
Acceleration   factor   ( c 2 )0.55
MVO Traveling   distance   rate   ( T D R )[1, 0]
ES Step   size   of   mutation   ( ρ )0.15
ASO Well   depth   of   potential   ( ϵ ) 1.0
Distance   at   zero   potential   ( δ )1.0
Note: Maximum number of iterations and population size are set at 100 and 50, respectively.
Table 6. The internal parameter values obtained by each algorithm.
Table 6. The internal parameter values obtained by each algorithm.
ParameterNewtonPSOMVOESASO
Short-circuit current, Iscr (A)3.664.504.083.963.84
Open-circuit voltage, Vocr (V)66.463.566.364.561.0
Current at MPP, Imr (A)3.513.483.313.543.53
Voltage at MPP, Vmr (V)5250.250.052.451.2
Irradiance under STC, Gr (W/m2)10001046.39001040.01011.9
Temperature under STC, Tr (K)298290.03290305.89300.52
Parallel resistance under STC, Rshr (Ω)15061.5560.0121.07123.58
Series resistance under STC, Rsr (Ω)2.45.996.04.165.52
Current temperature coefficient, βIsc (A/K)0.0006810.0006720.0006700.0006850.000684
Voltage temperature coefficient, βVoc (V/K)−0.1660−0.0379−0.1000−0.0639−0.0942
Ideality factor (1st diode), nidl11.21.01.011.171.16
Ideality factor (2nd diode), nidl21.86091.21.981.791.38
Ideality factor (3rd diode), nidl31.86091.23.01.871.64
Saturation current (1st diode), Isat1 (A)1.16 × 10 15 9.75 × 10 13 1.00 × 10 15 5.03   × 10 8 4.29 × 10 10
Saturation current (2nd diode), Isat2 (A)1.16   × 10 15 9.75 × 10 13 1.00 × 10 15 5.03 × 10 8 4.29   × 10 10
Saturation current (3rd diode), Isat3 (A)1.16   × 10 15 9.75   × 10 13 1.00 × 10 15 5.03   × 10 8 4.29 × 10 10
Table 7. Comparison of the five different methods (MRE%).
Table 7. Comparison of the five different methods (MRE%).
Weather ConditionsNewtonPSOMVOESASO
Rainy3.53 w2.642.63 b2.702.68
Heavily overcast5.49 w4.624.60b4.654.67 w
Overcast4.60 w4.524.544.49 b4.50
Partly cloudy4.30 b4.574.594.62 w4.57
Sunny3.563.533.533.58 w3.52 b
Average4.2953.9763.9784.0083.988
b: Best estimation; W: Worst estimation.
Table 8. Parameter settings of CNNs and LSTM.
Table 8. Parameter settings of CNNs and LSTM.
MethodLayerParameterValueInput ShapeOutput Shape
CNNsConv2Dfilter16(N 1, 5, 1, 1, 1)(N, 5, 1, 1, 16)
kernel size(1, 1)
activationRelu 2
MaxPooling2Dpool size(1, 1)(N, 5, 1, 1, 16)(N, 5, 1, 1, 1)
Flatten (N, 5, 1, 1, 16)(N, 5, 16)
LSTM units64(N, 5, 16)(N, 64)
activationRelu
Denseunits32(N, 64)(N, 32)
activationRelu
Denseunits1(N, 32)(N, 1)
activationlinear
1: N: the number of samples; 2: Relu: rectified linear unit.
Table 9. Comparison between the proposed and the other methods (MRE%).
Table 9. Comparison between the proposed and the other methods (MRE%).
Weather ConditionsSingle Optimization MethodEnsemble Method
NewtonPSOMOVESASOLSTMCNN-LSTM
Rainy3.53 w2.642.63 2.702.682.632.58 b
Heavily overcast5.49 w4.624.604.654.67 4.874.47 b
Overcast4.60 w4.524.544.49 b4.504.644.59
Partly cloudy4.30 b4.574.594.624.574.65 w4.63
Sunny3.563.533.533.58 w3.523.49 b3.57
Average4.2953.9763.9784.0083.9884.0563.968
b: Best estimation; W: Worst estimation.
Table 10. Comparison of various methods under the proposed ensemble framework (MRE%).
Table 10. Comparison of various methods under the proposed ensemble framework (MRE%).
Weather ConditionsSingle ModelHybrid Model
XGBoostLightGBMGRUCNNLSTMA-LSTMCNN-A-LSTMCNN-LSTM
Rainy2.6772.6372.6192.6412.5872.6052.5802.577
Heavily overcast4.5964.5624.5784.8214.6684.6944.5584.471
Overcast4.5374.5414.6604.6944.5954.5694.5854.590
Partly cloudy4.8314.8804.6594.5794.5724.5534.5984.631
Sunny3.8453.4933.5223.5283.5023.5023.5363.571
Average4.0974.0234.0084.0533.9853.9843.9713.968
Training time (s)0.320.214.5928.8266.6290.7114.4105.8
Table 11. Comparison of various methods under the proposed ensemble framework (MAE (kW)).
Table 11. Comparison of various methods under the proposed ensemble framework (MAE (kW)).
Weather ConditionsXGBoostLightGBMGRUCNNLSTMCNN-LSTM
Rainy19.62 19.33 19.20 20.72 18.9719.22
Heavily overcast33.69 33.44 33.56 37.63 34.1633.51
Overcast33.25 33.29 34.16 36.02 33.7633.61
Partly cloudy35.41 35.77 34.15 35.83 33.5933.48
Sunny28.18 25.60 25.82 27.28 25.6725.83
Average30.03 29.49 29.38 31.50 29.23 29.13
Table 12. Comparison of various methods under the proposed ensemble framework (NMAE).
Table 12. Comparison of various methods under the proposed ensemble framework (NMAE).
Weather ConditionsXGBoostLightGBMGRUCNNLSTMCNN-LSTM
Rainy0.3850.3790.3760.4060.3720.377
Heavily overcast0.2410.2390.2400.2690.2440.240
Overcast0.1690.1690.1730.1830.1710.171
Partly cloudy0.1400.1420.1350.1420.1330.133
Sunny0.0990.0900.0910.0960.0900.091
Average0.2070.2040.2030.2190.2020.202
Table 13. Comparison of various methods under the proposed ensemble framework (sMAPE%).
Table 13. Comparison of various methods under the proposed ensemble framework (sMAPE%).
Weather ConditionsXGBoostLightGBMGRUCNNLSTMCNN-LSTM
Rainy66.1565.6966.0785.8865.9365.70
Heavily overcast49.4048.8748.2563.5849.3547.87
Overcast29.1030.0431.3433.9030.0630.80
Partly cloudy27.5027.8127.2230.8126.4526.63
Sunny13.8112.5712.3913.5812.2612.51
Average37.1937.0037.0545.5536.8136.70
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, S.-P.; Shih, F.-R.; Huang, C.-M.; Chen, S.-J.; Chiua, C.-H. A Deep Learning-Based Ensemble Method for Parameter Estimation of Solar Cells Using a Three-Diode Model. Electronics 2025, 14, 3790. https://doi.org/10.3390/electronics14193790

AMA Style

Yang S-P, Shih F-R, Huang C-M, Chen S-J, Chiua C-H. A Deep Learning-Based Ensemble Method for Parameter Estimation of Solar Cells Using a Three-Diode Model. Electronics. 2025; 14(19):3790. https://doi.org/10.3390/electronics14193790

Chicago/Turabian Style

Yang, Sung-Pei, Fong-Ruei Shih, Chao-Ming Huang, Shin-Ju Chen, and Cheng-Hsuan Chiua. 2025. "A Deep Learning-Based Ensemble Method for Parameter Estimation of Solar Cells Using a Three-Diode Model" Electronics 14, no. 19: 3790. https://doi.org/10.3390/electronics14193790

APA Style

Yang, S.-P., Shih, F.-R., Huang, C.-M., Chen, S.-J., & Chiua, C.-H. (2025). A Deep Learning-Based Ensemble Method for Parameter Estimation of Solar Cells Using a Three-Diode Model. Electronics, 14(19), 3790. https://doi.org/10.3390/electronics14193790

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop