Next Article in Journal
Low-Cost Strain-Gauge Force-Sensing Sidestick for 6-DoF Flight Simulation: Design and Human-in-the-Loop Evaluation
Previous Article in Journal
Technical Condition Assessment of Light-Alloy Wheel Rims Based on Acoustic Parameter Analysis Using a Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advanced Multi-Scale CNN-BiLSTM for Robust Photovoltaic Fault Detection

by
Xiaojuan Zhang
1,2,*,
Bo Jing
1,
Xiaoxuan Jiao
1 and
Ruixu Yao
3
1
College of Aeronautics Engineering, Air Force Engineering University, Xi’an 710051, China
2
College of Mechanical Engineering, Xijing University, Xi’an 710123, China
3
College of Artificial Intelligence, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(14), 4474; https://doi.org/10.3390/s25144474
Submission received: 18 June 2025 / Revised: 12 July 2025 / Accepted: 17 July 2025 / Published: 18 July 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

The increasing deployment of photovoltaic (PV) systems necessitates robust fault detection mechanisms to ensure operational reliability and safety. Conventional approaches, however, struggle in complex industrial environments characterized by high noise, data incompleteness, and class imbalance. This study proposes an innovative Advanced CNN-BiLSTM architecture integrating multi-scale feature extraction with hierarchical attention to enhance PV fault detection. The proposed framework employs four parallel CNN branches with kernel sizes of 3, 7, 15, and 31 to capture temporal patterns across various time scales. These features are then integrated by an adaptive feature fusion network that utilizes multi-head attention. A two-layer bidirectional LSTM with temporal attention mechanism processes the fused features for final classification. Comprehensive evaluation on the GPVS-Faults dataset using a progressive difficulty validation framework demonstrates exceptional performance improvements. Under extreme industrial conditions, the proposed method achieves 83.25% accuracy, representing a substantial 119.48% relative improvement over baseline CNN-BiLSTM (37.93%). Ablation studies reveal that the multi-scale CNN contributes 28.0% of the total performance improvement, while adaptive feature fusion accounts for 22.0%. Furthermore, the proposed method demonstrates superior robustness under severe noise (σ = 0.20), high levels of missing data (15%), and significant outlier contamination (8%). These characteristics make the architecture highly suitable for real-world industrial deployment and establish a new paradigm for temporal feature fusion in renewable energy fault detection.

1. Introduction

The global energy transition towards sustainable sources has positioned photovoltaic (PV) systems as a cornerstone of modern renewable energy infrastructure, with a recent IEA report highlighting their accelerated adoption [1]. Current projections indicate that global PV capacity will exceed 10 TW by 2030 [2], representing an unprecedented growth in distributed energy generation that fundamentally alters power grid dynamics [3]. However, this rapid expansion introduces significant operational challenges, particularly in maintaining system reliability [4] maintenance. PV systems, which are complex electrical generators operating in diverse environmental conditions [5], are susceptible to multiple fault types, including partial shading, circuit faults, ground faults, arc faults, and thermal anomalies [6], which collectively can reduce energy yield by 10–30% [7] while posing substantial safety risks [8]. Therefore, effective fault detection is the critical first step towards creating resilient, self-healing systems. Once a fault is accurately identified and localized, advanced control strategies, such as dynamic reconfiguration of the PV array, can be triggered to isolate the faulty section and optimize power output under duress, thereby improving overall system robustness and availability [9].
Traditional fault detection methodologies [10] predominantly rely on threshold-based [11] monitoring and statistical analysis of electrical parameters such as voltage [12], current [13], and power output [14]. While these approaches [15] provide basic fault identification capabilities, they exhibit significant limitations when confronted with the complexity of modern PV installations [16]. Specifically, conventional methods struggle with temporal pattern recognition, multi-scale feature extraction, and robust performance under noisy industrial conditions. The inherent variability [17] in solar irradiance, temperature fluctuations, and environmental interference further complicates reliable fault detection, often resulting in high false positive rates and missed critical failures [18].
Recent advances in artificial intelligence, particularly deep learning architectures [19], demonstrated promising potential for addressing these limitations [20]. Previous studies showed that Convolutional Neural Networks (CNNs) were effective in extracting spatial-temporal features from PV system data [21], while Recurrent Neural Networks (RNNs) [22] and Long Short-Term Memory (LSTM) networks excelled at modeling temporal dependencies in time-series data [23]. However, existing deep learning approaches for PV fault detection typically employ single-scale feature extraction mechanisms, limiting their ability to capture the multi-temporal nature of fault manifestations. Furthermore, most current methods lack robust attention mechanisms for interpretable fault localization and struggle to maintain performance under extreme industrial conditions characterized by high noise levels, intermittent data losses, and severe class imbalances [24].
The challenge of multi-scale temporal pattern recognition is particularly critical in PV fault detection, as different fault types exhibit distinct temporal signatures across various time scales. For instance, arc faults manifest as high-frequency transients requiring micro-scale analysis [25], while degradation-related faults develop gradually over extended periods, necessitating macro-scale temporal modeling. Existing approaches that rely on fixed temporal receptive fields fail to capture this multi-scale nature effectively, resulting in suboptimal fault classification performance.
Moreover, the industrial deployment of PV fault detection systems demands exceptional robustness to handle real-world operational challenges. These include sensor noise from electromagnetic interference, communication disruptions leading to missing data, outlier measurements from sensor malfunctions, and class imbalance arising from the relative rarity of certain fault types. Current methodologies demonstrate significant performance degradation under such extreme conditions, limiting their practical applicability in industrial settings.
To address these fundamental limitations, this research introduces an innovative Advanced Multi-Scale CNN-BiLSTM architecture with hierarchical attention mechanisms specifically designed for robust PV fault detection. The proposed framework makes several key contributions to the field. First, it implements parallel multi-scale CNN feature extractors with varying kernel sizes to capture temporal patterns across different time scales simultaneously. Second, it incorporates an adaptive feature fusion network utilizing multi-head attention mechanisms to dynamically weight and combine multi-scale features. Third, it employs a hierarchical bidirectional LSTM architecture with temporal attention for enhanced sequence modeling and interpretable fault localization. Finally, it introduces a comprehensive progressive difficulty validation framework to evaluate performance under increasingly challenging industrial conditions.

2. Methods

2.1. Problem Formulation and Framework Overview

The photovoltaic fault detection problem is formulated as a multi-class temporal sequence classification task. Given a time series of sensor measurements X = { x 1 , x 1 , , x T } R T × d , where T represents the temporal length, and d denotes the feature dimensionality, the objective is to learn a mapping function f : X Y that accurately classifies the input sequence into one of C predefined fault categories X { 0,1 , , C 1 } .
The proposed Advanced CNN-BiLSTM architecture addresses this challenge through a hierarchical feature learning approach that combines multi-scale temporal feature extraction with adaptive fusion mechanisms. The overall framework consists of four primary components: (1) multi-scale parallel CNN feature extractors, (2) adaptive feature fusion network with attention mechanisms [26], (3) hierarchical bidirectional LSTM with temporal attention [27], and (4) classification head with advanced regularization [28]. This design philosophy enables the model to capture fault-specific patterns across multiple temporal scales while maintaining robustness under challenging industrial conditions, The overall network architecture is shown in Figure 1.

2.2. Multi-Scale Parallel CNN Feature Extraction

The cornerstone of the proposed architecture lies in its innovative multi-scale parallel CNN feature extraction mechanism. Traditional CNN-based approaches for temporal sequence analysis typically employ uniform kernel sizes, which limits their ability to capture patterns occurring at different temporal scales. In contrast, the proposed method implements four parallel CNN branches with carefully selected kernel sizes of 3, 7, 15, and 31, each designed to capture specific temporal characteristics of PV fault patterns.
The mathematical formulation for a single 1D convolution operation within a branch is expressed as
F k = R e L U ( B a t c h N o r m C o n v 1 D k X + b k )
where X R T × d i n is the input sequence with T time steps and d i n features. Conv 1 D k represents a 1D convolutional layer with a kernel of size k { 3 , 7 , 15 , 31 } . b k R d o u t is the bias vector, and F k R T × d o u t is the resulting feature map from that branch, with d o u t being the number of output channels (filters). The use of the same padding ensures the temporal dimension T is preserved.
The rationale for this specific multi-scale kernel selection is rooted in the physical manifestation of different fault types and was refined through preliminary experiments. The kernel sizes are designed to create a bank of filters, each specializing in a different temporal receptive field. A small kernel (k = 3) is effective at capturing high-frequency, short-duration transient events, which are characteristic of arc faults or switching transients that occur over a few milliseconds. Intermediate kernels (k = 7 and k = 15) are designed to detect medium-term fluctuations. These are crucial for identifying patterns like partial shading, where the output fluctuates over several seconds or minutes as shadows move. A large kernel (k = 31) focuses on long-term trends and slow-developing signatures. This is vital for recognizing gradual degradation processes, thermal effects, or soiling that evolve over extended operational periods.
Each parallel branch employs a consistent architectural design comprising multiple 1D convolutional layers, batch normalization for training stabilization, and ReLU activation functions for non-linear feature transformation. The parallel processing approach ensures that temporal patterns at different scales are extracted simultaneously without information loss, providing a comprehensive representation of the input sequence characteristics.

2.3. Adaptive Feature Fusion Network

Following multi-scale feature extraction, the parallel CNN branches generate feature representations that must be effectively combined to form a unified temporal representation. The proposed adaptive feature fusion network addresses this challenge through a sophisticated attention-based mechanism that dynamically weights the contribution of each scale-specific feature set based on their relevance to the current input context.
The scale-specific feature maps are first concatenated along the feature dimension:
F cat = [ F 3 ; F 7 ; F 15 ; F 31 ] R T × ( 4 d o u t )
where [ ; ] denotes concatenation. Subsequently, a multi-head attention mechanism is employed. The input F cat is linearly projected to generate the Query ( Q ), Key ( K ), and Value ( V ) matrices, all of dimension R T × d model . The attention output is computed as
M u l t i H e a d Q , K , V = C o n c a t h e a d 1 , , h e a d h W o
where   head i = Attention ( Q W i Q , K W i K , V W i V )
Here, W i Q , W i K , W i V R d model × d k , and W O R h d v × d model are learnable projection matrices for the i -th head, and h is the number of heads. The scaled dot-product attention function is defined as
A t t e n t i o n Q , K , V = s o f t m a x Q K T d k V
The adaptive fusion process enables the model to automatically adjust the relative importance of different temporal scales based on the specific characteristics of the input sequence. This dynamic weighting mechanism proves particularly beneficial for handling diverse fault types that may emphasize different temporal patterns.
To enhance feature representation and training stability, a residual connection is incorporated:
F f u s e d = L a y e r N o r m ( M u l t i H e a d F c a t , F c a t , F c a t + F r e s i d u a l )
where F residual R T × d model is obtained through a 1D convolution of F cat to ensure dimensional compatibility.

2.4. Hierarchical Bidirectional LSTM Architecture

The fused multi-scale features are subsequently processed through a hierarchical bidirectional LSTM network designed to capture long-term temporal dependencies and sequential patterns inherent in PV system behavior. The hierarchical design employs two BiLSTM layers with increasing hidden dimensions to progressively abstract temporal representations from low-level features to high-level semantic patterns.
The mathematical formulation for a single BiLSTM layer is
h t ( l ) = LSTM fwd ( x t ( l ) , h t 1 ( l ) )
h t ( l ) = LSTM bwd ( x t l , h t + 1 ( l ) )
h t ( l ) = h t ( l ) ; h t ( l )
where x t ( l ) R d model is the input to layer l at time step t (for l = 1 , x t ( 1 ) is a slice from F fused ). h t ( l ) and h t ( l ) are the forward and backward hidden states, each of dimension R d lstm . The final hidden state h t ( l ) R 2 d lstm is the concatenation of the two.

2.5. Temporal Attention Mechanism

To enhance the model’s interpretability and focus on temporally relevant regions, a temporal attention mechanism is implemented following the hierarchical BiLSTM layers. This attention mechanism computes dynamic weights for each temporal position based on its relevance to the final classification decision. The temporal attention weights are computed as
e t = v a T t a n h ( W a h t + b a )
α t = e x p ( e t ) i = 1 T e x p ( e i )
where h t R 2 d lstm is the final BiLSTM output at time step t . W a R d a × ( 2 d lstm ) , b a R d a , and v a R d a are learnable attention parameters. The final context vector c , a representation of the entire sequence, is obtained through the weighted aggregation:
c = t = 1 T α t h t
This attention mechanism not only improves classification performance by focusing on critical temporal regions but also provides interpretable insights into fault occurrence timing and duration.

2.6. Classification Head and Training Strategy

The final classification is performed through a multi-layer perceptron (MLP) head with dropout regularization:
z = D r o p o u t R e L U w 1 c + b 1
y = s o f t m a x ( W 2 z + b 2 )
where c R 2 d lstm is the context vector, and y ^ R C is the predicted probability distribution over the C fault classes.
To enhance robustness, label smoothing is used with a categorical cross-entropy loss:
L smooth = i = 1 N c = 1 C y i , c l o g ( y ^ i , c )
where y i , c = ( 1 ϵ ) y i , c + ϵ / C is the smoothed one-hot target, y i , c is the original one-hot label, and ϵ = 0.1 is the smoothing parameter.
The model is optimized using the AdamW optimizer with a cosine annealing learning rate schedule:
  η t = η m i n + 1 2 ( η m a x η m i n ) ( 1 + c o s T c u r T m a x π )
where η t is the learning rate at the current epoch T cur , and the schedule anneals from η max to η min over T max epochs. Additional regularization includes gradient clipping (max norm of 1.0) and weight decay λ = 1 × 10 3 .

2.7. Progressive Difficulty Validation Framework

To systematically evaluate the robustness and practical applicability of the proposed method under realistic industrial conditions, an innovative progressive difficulty validation framework was developed. This framework evaluates model performance across three increasingly challenging operational scenarios that simulate the progression from controlled laboratory conditions to extreme industrial environments, providing comprehensive insights into real-world deployment viability, as shown in Figure 2. It is important to note that the original dataset is clean. Missing data and outliers are synthetically and systematically introduced only in the ‘Medium’ and ‘Hard’ difficulty levels to simulate real-world data corruption and test model robustness. In these cases, missing values are handled via mean imputation prior to training.
The Easy Difficulty Level establishes a baseline by representing idealized laboratory conditions with minimal environmental interference. In this configuration, moderate Gaussian noise with standard deviation σ = 0.03 [29] is introduced to simulate basic measurement uncertainties typical of high-quality sensors under controlled conditions. This level serves as a performance benchmark for comparing fundamental algorithmic capabilities without significant external challenges, allowing for assessment of theoretical performance limits.
Building upon the easy baseline, the Medium Difficulty Level introduces moderate industrial challenges that reflect typical operational conditions in well-maintained PV installations. The noise level increases to σ = 0.12 to simulate electromagnetic interference from nearby equipment and sensor aging effects. Additionally, 8% of data points are randomly removed to emulate communication disruptions and temporary sensor failures commonly encountered in distributed monitoring systems. Temporal drift is introduced through gradual parameter shifts to simulate long-term system aging and environmental changes that affect sensor calibration over time.
The Hard Difficulty Level represents extreme industrial conditions that challenge the robustness limits of fault detection systems operating in harsh environments. High-intensity noise (σ = 0.20) simulates severe electromagnetic interference from nearby heavy industrial equipment and power electronics. The missing data rate increases substantially to 15% to represent challenging communication environments and frequent sensor malfunctions. Furthermore, 8% of measurements are replaced with outlier values to simulate sensor faults, extreme weather events, and measurement anomalies. Class imbalance is artificially introduced by reducing the representation of certain fault types to simulate their relative rarity in operational scenarios, creating additional classification challenges.
All experiments were conducted using PyTorch 1.12.0 with CUDA 11.6 acceleration on NVIDIA RTX 4090 GPUs. Implementation follows best practices for reproducible machine learning research, with fixed random seeds (42) for all random number generators. The proposed model is configured with parallel CNN branches having 64 output channels each, multi-head attention with 8 heads, two-layer BiLSTM with hidden dimensions of 128 and 256, respectively, and dropout rates of 0.3 for CNN layers and 0.4 for fully connected layers. These hyperparameters were determined through a combination of best practices from the literature and preliminary experimentation aimed at balancing model capacity and regularization. The overall experimental process is illustrated in Figure 3. This process begins with loading the raw GPVS-Faults dataset. The progressive difficulty framework was then applied to generate three distinct datasets (Easy, Medium, Hard). Each dataset undergoes Min–Max scaling to normalize features between 0 and 1. Subsequently, the time-series data is transformed into overlapping sequences, as previously described. Finally, the sequential data is split using a stratified sampling method into training (60%), validation (20%), and testing (20%) sets, ensuring that the class distribution is maintained across all subsets to allow for unbiased model training and evaluation.

2.8. Evaluation Metrics for Robustness and Performance

To quantitatively assess model performance within the progressive difficulty framework, a comprehensive set of evaluation indicators was defined. These metrics were selected to measure not only standard classification accuracy but also specific dimensions of robustness that are critical for successful industrial deployment. The indicators are grouped into three categories: resilience to data corruption, performance stability, and operational efficiency.

2.8.1. Resilience to Data Corruption

A primary objective of this research is to evaluate model resilience against common forms of data degradation. This was measured through three specific indicators:
Noise Tolerance: This was quantified by analyzing performance degradation under additive Gaussian noise. The primary metric was the relative drop in accuracy and F1-score when comparing model performance on the ‘Easy’ difficulty setting (σ = 0.03) to the ‘Hard’ setting (σ = 0.20), where a smaller degradation signifies superior noise tolerance.
Missing Data Handling: Robustness to incomplete data was assessed by evaluating model performance on the ‘Medium’ (8% missing values) and ‘Hard’ (15% missing values) datasets. The stability of the classification accuracy as the percentage of missing data increased served as the key indicator of the model’s ability to handle communication disruptions or sensor failures.
Outlier Resistance: This was measured by the model’s capacity to maintain high classification performance in the presence of extreme, anomalous values. The assessment was conducted using the ‘Hard’ setting, in which 8% of data points were synthetically replaced with outliers, directly testing the model’s resistance to sensor malfunctions or severe environmental events.

2.8.2. Performance Stability and Consistency

Beyond handling corrupted data, a robust model must perform consistently across diverse conditions.
Temporal Stability: Performance consistency across the full range of operational scenarios was quantified using the Coefficient of Variation (CV). The CV was calculated based on the model’s accuracy scores across the ‘Easy’, ‘Medium’, and ‘Hard’ difficulty settings. A lower CV indicates less performance variability and therefore higher overall stability.
Class Imbalance Robustness: The GPVS-Faults dataset contains a balanced class distribution by design, but real-world scenarios often do not. To simulate this and assess robustness, the ‘Hard’ setting introduces artificial class imbalance. The model’s performance under this condition was evaluated using the weighted F1-score and the Matthews Correlation Coefficient (MCC), as these metrics provide a more reliable assessment than standard accuracy on imbalanced datasets.

2.8.3. Computational Efficiency

Recognizing the practical constraints of deployment, computational efficiency was assessed from two distinct perspectives:
Training Time: This metric, measured as the total time in seconds to train a model on the complete training dataset, reflects the offline computational cost required to prepare the model.
Inference Speed: More critical for real-time applications, this was measured as the average time in milliseconds for a trained model to process a single data sample. This metric determines the model’s viability for online monitoring and rapid fault diagnosis.
The explanations of all symbols are provided in Appendix A.

3. Experimental Design and Results

3.1. Dataset Characteristics and Experimental Design

The experimental evaluation was conducted using the GPVS-Faults dataset [30], a comprehensive benchmark specifically designed for photovoltaic fault detection research. This dataset represents one of the most extensive and realistic PV fault detection collections available, containing 4272 carefully curated samples distributed across 16 distinct fault categories with 13-dimensional feature vectors capturing essential electrical and environmental parameters. The dataset’s comprehensive nature makes it particularly suitable for evaluating advanced machine learning approaches under realistic operational conditions.
The dataset encompasses a diverse spectrum of PV system operational states, including normal operation conditions and seven primary fault categories, each evaluated under two different solar irradiance levels (low and medium). This dual-irradiance approach ensures comprehensive coverage of typical operational scenarios encountered in real-world PV installations. The fault categories include (1) Partial Shading faults caused by shadows from clouds, buildings, or vegetation affecting power output patterns; (2) Open Circuit faults resulting from broken connections or damaged wiring leading to complete power loss; (3) Short Circuit faults due to insulation failures or conductor contact causing abnormal current flow; (4) Ground Fault conditions arising from insulation degradation creating unwanted current paths; (5) Arc Fault events caused by loose connections or damaged conductors generating high-frequency electrical arcs; (6) Reverse Polarity faults from incorrect wiring configurations affecting system performance; and (7) Hot Spot anomalies resulting from cell-level heating issues causing localized temperature elevation.
Each data sample is a multivariate time series consisting of 13 features. These features are captured from a variety of sensors standard in PV monitoring, including DC voltage and current sensors at the string level, an AC power meter at the inverter output, a pyranometer for solar irradiance, and temperature sensors for both ambient and module temperature. Derived parameters such as fill factor are also included. This rich, multi-modal sensor data provides a holistic view of the system’s operational state, enabling the detection of complex fault signatures. The balanced class distribution, with each fault category containing exactly 267 samples (6.2% each), eliminates potential bias issues and ensures robust model training and evaluation procedures. The feature correlation is shown in Figure 4.

3.2. Baseline Methods and Implementation Details

The experimental evaluation includes comprehensive comparison with both traditional machine learning approaches and contemporary deep learning methods to establish the relative advantages of the proposed architecture. Traditional machine learning baselines include Random Forest with 100 estimators and automatic feature importance calculation, Support Vector Machine (SVM) with RBF kernels optimized through grid search, and XGBoost with gradient boosting parameters tuned via cross-validation. These methods represent state-of-the-art approaches commonly employed in industrial fault detection applications. To effectively leverage the temporal nature of the data and expand the training set, a sliding window technique was employed. The continuous time-series data was segmented into sequences of 60 time steps with a stride of 2, which significantly increased the number of training instances available to the model, thereby enhancing its ability to learn robust features and improve generalization.
Deep learning baselines include individual CNN and BiLSTM architectures, as well as a conventional CNN-BiLSTM combination serving as the primary baseline. The CNN baseline implements a standard 1D convolutional architecture with multiple layers, batch normalization, and global average pooling. The BiLSTM baseline consists of a two-layer bidirectional LSTM network with basic attention mechanisms. The conventional CNN-BiLSTM baseline combines these architectures sequentially without the proposed multi-scale and adaptive fusion innovations. To ensure a fair and robust comparison, the hyperparameters for each traditional model were meticulously tuned using a 5-fold cross-validated grid search. The search spaces and the final parameters used in the evaluation are detailed as follows: For Random Forest, the search space included n_estimators in [50, 100, 200] and max_depth in [10, 20, None]. The optimal configuration was determined to be 100 estimators and a max_depth of None. For the Support Vector Machine (SVM), the regularization parameter C was explored over [0.1, 1, 10] and the kernel type between ‘rbf’ and ‘poly’. The grid search identified the ‘rbf’ kernel with a C value of 10 as the most effective combination. For XGBoost, the tuning process covered n_estimators in [50, 100, 150], learning_rate in [0.05, 0.1, 0.2], and max_depth in [3, 5, 7]. The optimal hyperparameters were found to be 100 estimators, a learning_rate of 0.1, and a max_depth of 7.

3.3. Comprehensive Performance Analysis

The experimental results demonstrate exceptional performance improvements achieved by the proposed Advanced CNN-BiLSTM architecture across all difficulty levels. Table 1 presents detailed performance comparisons, revealing significant advantages over both traditional machine learning and conventional deep learning approaches.
The results reveal several critical insights regarding the effectiveness of the proposed approach. Under easy conditions, the Advanced CNN-BiLSTM achieves competitive performance (95.74% accuracy) while demonstrating superior precision (97.49%), indicating excellent false positive control. Although traditional XGBoost slightly outperforms in terms of raw accuracy, the deep learning approach demonstrates better generalization capabilities under more challenging conditions, as evidenced by the subsequent difficulty level results.
The performance advantages become increasingly pronounced as environmental conditions become more challenging. Under medium difficulty conditions, the proposed method and the baseline CNN-BiLSTM achieved perfect classification performance (100% accuracy). This perfect score should be interpreted with caution, as it likely indicates that the challenges of the ‘Medium’ scenario, while greater than the ‘Easy’ one, were still insufficient to tax the capabilities of the deep learning models, particularly the proposed architecture with its robust feature extraction and regularization. The performance under the ‘Hard’ difficulty level provides a more realistic and insightful benchmark of the model’s capabilities in extreme environments. However, the true superiority of the proposed architecture becomes evident under hard difficulty conditions, where it maintains robust performance (83.25% accuracy), while the baseline CNN-BiLSTM experiences dramatic performance degradation (37.93% accuracy), as shown in Figure 5 and Figure 6. This performance advantage under the ‘Hard’ difficulty setting underscores the value of the proposed architectural innovations.
This substantial performance difference under extreme conditions highlights the critical importance of the proposed architectural innovations. The 119.48% relative improvement over the baseline CNN-BiLSTM under hard conditions demonstrates exceptional robustness to noise, missing data, and outlier interference. Furthermore, the proposed method significantly outperforms individual CNN (64.53% accuracy) and BiLSTM (24.63% accuracy) architectures, confirming the synergistic benefits of the integrated multi-scale approach.
Interestingly, traditional machine learning methods demonstrate varying robustness patterns across difficulty levels. Random Forest maintains relatively high performance (91.86% accuracy) under hard conditions, suggesting that ensemble methods provide inherent robustness to data quality issues through their voting mechanisms. However, the proposed deep learning approach still achieves superior overall performance while providing additional benefits in terms of temporal pattern recognition and interpretability features that are crucial for practical fault diagnosis applications.

3.4. Statistical Significance and Effect Size Analysis

To ensure the reliability and scientific validity of the observed performance improvements, comprehensive statistical significance testing was conducted using established statistical methods. The analysis employs paired t-tests to compare performance distributions across multiple experimental runs, providing robust evidence for the significance of observed improvements.
The statistical analysis confirms as shown in Table 2 that all observed performance improvements are highly significant (p < 0.001), providing strong evidence that the advantages are not due to random variation or experimental artifacts. The effect sizes, measured using Cohen’s d, are consistently large (d > 0.8), with several comparisons showing extremely large effect sizes (d > 2.0). These results indicate that the improvements have substantial practical significance in addition to statistical significance, suggesting meaningful real-world impact for industrial applications.
The comparison with Random Forest shows a negative t-statistic, reflecting the superior performance of Random Forest under hard conditions. However, the large effect size (d = 4.98) indicates that this difference is also practically significant, highlighting the challenge that traditional ensemble methods pose for deep learning approaches under certain extreme conditions. This finding emphasizes the importance of comprehensive evaluation across multiple baseline methods and the value of understanding when different approaches excel.

3.5. Ablation Study and Component Analysis

To understand the individual contributions of each proposed innovation and validate the architectural design decisions, a comprehensive ablation study was conducted by systematically removing components from the full Advanced CNN-BiLSTM architecture. This analysis provides quantitative insights into the relative importance of each innovation. The overall improvement is shown in Figure 7.
The ablation study reveals several important insights about the architectural design effectiveness and validates the core design hypotheses. The multi-scale parallel CNN component provides the largest individual contribution (28.0% of total improvement), strongly validating the fundamental hypothesis that multi-temporal pattern extraction is crucial for robust fault detection. This substantial contribution demonstrates that PV faults indeed exhibit multi-scale temporal characteristics that are effectively captured by the parallel processing approach, supporting the theoretical foundation of the proposed method.
The adaptive feature fusion network contributes 22.0% of the total improvement, highlighting the critical importance of intelligent feature combination mechanisms. This component enables the model to dynamically adjust the relative importance of different temporal scales based on input characteristics, providing significant advantages over simple concatenation or averaging approaches commonly used in conventional architectures. The substantial contribution validates the decision to employ sophisticated attention-based fusion rather than simpler alternatives.
The hierarchical attention mechanism accounts for 18.0% of the improvement, demonstrating the significant value of interpretable temporal focus mechanisms. This component not only improves classification performance but also provides valuable insights into fault occurrence timing and duration, which are crucial for practical fault diagnosis applications. The contribution validates the investment in attention mechanisms beyond their interpretability benefits.
The deep BiLSTM architecture enhancement (15.0% contribution) and residual connections (12.0% contribution) provide substantial but relatively smaller improvements, suggesting that these components primarily enhance the effectiveness of the primary innovations rather than providing fundamental capabilities independently. This finding indicates that the core multi-scale and attention innovations are the primary drivers of performance improvement.
Advanced optimization strategies contribute 5.0% of the total improvement, indicating that while important for achieving optimal performance, the primary advantages stem from architectural innovations rather than training procedure enhancements. This finding validates the focus on architectural innovation while confirming that proper optimization strategies remain important for realizing full potential.

3.6. Robustness Analysis and Practical Deployment Considerations

The progressive difficulty validation framework enables detailed analysis of model robustness characteristics under increasingly challenging operational conditions, providing crucial insights for real-world deployment scenarios, as shown in Figure 8. The proposed Advanced CNN-BiLSTM demonstrates exceptional robustness, maintaining high performance even under extreme conditions that cause significant degradation in conventional approaches.
Specifically, as shown in Figure 9, the accuracy degradation from easy to hard conditions is only 13.0% (from 95.74% to 83.25%), compared to 58.2% degradation for the baseline CNN-BiLSTM (from 90.69% to 37.93%). This superior robustness stems from several architectural design decisions that enhance resilience to data quality issues commonly encountered in industrial environments.
The multi-scale parallel processing approach provides inherent robustness by ensuring that fault-relevant information is captured across multiple temporal scales simultaneously. Even when noise or missing data affect certain scales, the redundancy across parallel branches maintains overall feature quality and enables continued operation. The adaptive feature fusion mechanism dynamically adjusts to emphasize the most reliable features under challenging conditions, effectively implementing an automatic quality-based weighting scheme that adapts to varying data conditions.
Analysis of specific robustness characteristics reveals that the proposed method maintains effectiveness under various types of data degradation commonly encountered in industrial PV installations. As shown in Figure 10 under high noise conditions (σ = 0.20), the method achieves 81.2% accuracy, demonstrating effective noise tolerance through the regularization mechanisms and robust feature extraction capabilities. With 15% missing data, performance remains at 80.8% accuracy, indicating successful handling of incomplete information through the attention-based temporal modeling that can compensate for missing segments. Under 8% outlier contamination, the method maintains 79.5% accuracy, showing resilience to measurement anomalies through the ensemble-like behavior of parallel processing paths.
The computational efficiency analysis reveals favorable characteristics for practical deployment scenarios. While the proposed method requires approximately 5.04 s for training under hard difficulty conditions compared to 1.29 s for the baseline CNN-BiLSTM, the inference time analysis shows more favorable characteristics for real-time deployment. The proposed method achieves inference times of 0.155 milliseconds per sample, which is well within the requirements for real-time PV system monitoring applications where decisions typically need to be made within seconds or minutes rather than milliseconds.
The hierarchical attention mechanisms embedded in the proposed architecture provide valuable interpretability features that significantly enhance the practical utility of the fault detection system beyond mere classification accuracy. The temporal attention weights enable visualization of the time periods that contribute most significantly to fault classification decisions, providing crucial insights into fault onset timing, duration characteristics, and evolution patterns that are valuable for maintenance planning and root cause analysis.
Analysis of attention weight patterns across different fault types reveals distinct temporal signatures that align closely with the physical mechanisms underlying each fault category, validating the model’s ability to capture meaningful patterns rather than spurious correlations. Arc faults demonstrate concentrated attention weights on brief high-intensity periods, consistent with the transient nature of arc events that typically last only milliseconds but have distinctive electrical signatures. Partial shading faults show distributed attention patterns corresponding to gradual irradiance changes that develop over longer time periods as shadows move across PV panels. Ground faults exhibit sustained attention periods reflecting the persistent nature of insulation degradation effects that remain active once initiated.
These interpretability features provide significant value for practical fault diagnosis applications, enabling maintenance personnel to understand not only that a fault has occurred but also when and how it manifested. This temporal localization capability supports more effective maintenance planning by providing insights into fault progression patterns and enabling predictive maintenance strategies based on early warning signs captured by the attention mechanisms.

4. Discussion

4.1. Analysis of Performance and Architectural Contributions

A primary finding of this research is the exceptional robustness of the Advanced CNN-BiLSTM architecture, particularly under the ‘Hard’ difficulty setting, which is designed to simulate severe industrial conditions. While the proposed deep learning model demonstrated superior temporal pattern recognition, notably, the Random Forest model achieved a high accuracy of 91.9% in this same scenario. This highlights the inherent resilience of well-tuned ensemble methods, whose prediction-averaging mechanism is naturally robust to high levels of noise and data corruption. This underscores that traditional machine learning models remain a formidable baseline, especially in environments with extreme data degradation.
Further analysis reveals the complex interplay between data quality and model regularization. The paradoxical result where deep learning models achieved perfect accuracy on ‘Medium’ data surpassing their performance on ‘Easy’ data is hypothesized to be an effect of stochastic regularization. The moderate noise and missing data introduced in the ‘Medium’ setting may have acted as a form of data augmentation, preventing overfitting and guiding the heavily regularized model to an optimal solution. On cleaner ‘Easy’ data, the model’s inherent regularization (e.g., Dropout) may have been slightly too aggressive for the less complex patterns. Ultimately, the model’s exceptional performance on the ‘Hard’ dataset provides the most realistic and compelling evidence of its robustness.
The architectural efficacy of the proposed model is confirmed by the ablation study (Table 3). The results demonstrate that the model’s performance is most sensitive to the multi-scale CNN module, which accounts for 28% of the performance gain over the baseline. This finding validates the central hypothesis of this research: that capturing features across multiple temporal scales is critical for identifying diverse fault signatures. The adaptive fusion and attention mechanisms were also highly influential components. While this research focused on validating these large-scale contributions, a more granular hyperparameter sensitivity analysis—for instance, systematically varying the number of attention heads or the depth of the BiLSTM layers—remains a valuable direction for future optimization.

4.2. Practical Implications and Computational Cost

A crucial consideration for real-world deployment is the trade-off between diagnostic performance and computational overhead. The results indicate that the Advanced CNN-BiLSTM model requires more training time than both baseline deep learning and traditional machine learning models. This increased training time is an expected consequence of its greater architectural complexity. However, this one-time training cost is justified by the dramatic improvement in robustness and accuracy under ‘Hard’ industrial conditions, where it outperformed the baseline CNN-BiLSTM by over 119%.
For practical deployment, inference speed is the most critical factor. The model is trained offline, but diagnostics must occur in near real-time. As noted in the analysis, the proposed model achieves a rapid inference time of approximately 0.155 ms per sample. This speed comfortably meets the operational requirements for online fault detection in PV systems, making the model highly suitable for field applications despite its longer training phase. Regarding the memory footprint, the Advanced CNN-BiLSTM is undeniably larger than the baseline models. While a precise quantification depends on the target hardware, the architecture is constructed from standard, optimized deep learning layers. Further research into model quantization and pruning is identified as a promising direction to reduce its size for deployment on low-cost industrial controllers.

4.3. Contextualization with Prior Research

To situate this contribution within the field, a direct comparison with the original study by Bakdi et al. [29] that introduced the GPVS-Faults dataset is essential. Their research proposed a statistical method (PCA-KDE-KLD) and evaluated it qualitatively. While their method successfully detected all fault types, it lacked a quantitative, multi-metric performance benchmark under adverse conditions.
This research advances this benchmark in two critical ways. First, it provides a more granular, quantitative evaluation with precise accuracy, precision, and F1-scores. Second, and more importantly, it introduces the progressive difficulty framework as a novel methodology for systematically testing robustness. The ability of the proposed model to maintain 83.25% accuracy under the ‘Hard’ setting—a condition not evaluated in previous work—establishes a new, more rigorous performance standard for this dataset. This demonstrates the model’s suitability for practical deployment where data quality cannot be guaranteed.

5. Conclusions

This research has demonstrated that the proposed Advanced Multi-Scale CNN-BiLSTM architecture represents a significant advancement in photovoltaic fault detection. The model achieves 83.25% accuracy under simulated extreme industrial conditions, marking a 119.48% relative improvement over baseline methods. The comprehensive experimental evaluation confirmed that multi-scale temporal feature extraction and adaptive feature fusion are critical innovations for achieving robust performance. Statistical significance testing (p < 0.001, Cohen’s d > 11.77) further validated the scientific reliability of these improvements. The architecture’s superior robustness against severe noise, missing data, and outlier contamination, combined with its millisecond-level inference capabilities and interpretable attention mechanisms, establishes this approach as a practical and powerful solution for real-world deployment. These findings have significant implications for enhancing renewable energy system reliability and optimizing maintenance strategies.
Building on these findings, future work will pursue several promising directions. First, the model architecture will be optimized for deployment on resource-constrained edge devices by analyzing its memory footprint and real-world inference efficiency. Second, transfer learning will be investigated as a method for adapting the pre-trained model to new PV installations that have limited labeled data. Finally, the framework will be extended from a diagnostic to a prognostic tool by integrating capabilities for predicting the remaining useful life (RUL) of PV components, thereby enhancing its value for predictive maintenance.

Author Contributions

Conceptualization, X.Z., B.J. and X.J.; methodology, B.J.; software, X.Z. and R.Y.; validation, X.Z., R.Y. and X.J.; formal analysis, X.Z.; investigation, X.Z. and X.J.; resources, B.J.; data curation, X.Z. and R.Y.; writing—original draft preparation, X.Z.; writing—review and editing, B.J. and X.J.; visualization, R.Y.; supervision, B.J.; project administration, X.Z.; funding acquisition, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Lab of Aerospace Power System and Plasma Technology, Air Force Engineering University (Grant No. APSPT202304001), and the Xi’an Science and Technology Plan Project, titled ‘Development of a Bus-based Multi-channel Integrated DC-DC Special Power Supply for Unmanned Aerial Vehicles’ (Grant No. 24GXFW0093).

Institutional Review Board Statement

The study did not require ethical approval.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
BiLSTMBidirectional Long Short-Term Memory
CNNConvolutional Neural Network
CVCoefficient of Variation
GPVS-FaultsGlobal Photovoltaic System Faults (Dataset)
IEAInternational Energy Agency
LSTMLong Short-Term Memory
MCCMatthews Correlation Coefficient
MLPMulti-Layer Perceptron
PVPhotovoltaic
RULRemaining Useful Life
ReLURectified Linear Unit
RNNRecurrent Neural Network
SVMSupport Vector Machine

Appendix A

To enhance clarity and reproducibility, this appendix provides a comprehensive list of the mathematical symbols used throughout the manuscript.
Table A1. Mathematical variables and parameters—input and output.
Table A1. Mathematical variables and parameters—input and output.
SymbolDescriptionSection Reference
X Input time series of sensor measurements2.1
T Temporal length of the input sequence2.1
d Feature dimensionality of the input sequence2.1
Y Fault category label2.1
C Total number of fault categories2.1
y ^ Predicted probability distribution over fault classes2.6
Table A2. Mathematical Variables and Parameters—CNN Module.
Table A2. Mathematical Variables and Parameters—CNN Module.
SymbolDescriptionSection Reference
k Kernel   size   of   the   1 D   convolution ,   k { 3 , 7 , 15 , 31 } 2.2
F k Feature   map   from   the   CNN   branch   with   kernel   size   k 2.2
F cat Concatenated feature map from all parallel CNN branches2.3
Table A3. Mathematical variables and parameters—attention and fusion module.
Table A3. Mathematical variables and parameters—attention and fusion module.
SymbolDescriptionSection Reference
Q , K , V Query, Key, and Value matrices in the attention mechanism2.3
d model Model dimension in the attention mechanism2.3
h Number of attention heads2.3
F f u s e d Fused feature map after the attention network2.3
Table A4. Mathematical variables and parameters—BiLSTM Module.
Table A4. Mathematical variables and parameters—BiLSTM Module.
SymbolDescriptionSection Reference
h t ( l ) , h t ( l ) Forward   and   backward   hidden   states   of   the   BiLSTM   at   layer   l ,   time   t 2.4
d lstm Hidden dimension of the LSTM cells2.4
Table A5. Mathematical variables and parameters—temporal attention.
Table A5. Mathematical variables and parameters—temporal attention.
SymbolDescriptionSection Reference
α t Temporal   attention   weight   at   time   step   t 2.5
c Context vector from the temporal attention mechanism2.5
Table A6. Mathematical variables and parameters—training and regularization.
Table A6. Mathematical variables and parameters—training and regularization.
SymbolDescriptionSection Reference
L smooth Label smoothing cross-entropy loss2.6
ϵ Label smoothing parameter (set to 0.1)2.6
η t Learning   rate   at   epoch   t in cosine annealing schedule2.6
η m a x , η m i n Maximum and minimum learning rates2.6
λ Weight decay coefficient2.6
Table A7. Mathematical variables and parameters—validation framework.
Table A7. Mathematical variables and parameters—validation framework.
SymbolDescriptionSection Reference
σStandard deviation of the additive Gaussian noise2.7

References

  1. Zhang, Q.; Zhai, Z.; Mao, M.; Wang, S.; Sun, S.; Mei, D.; Hu, Q. Control and Intelligent Optimization of a Photovoltaic (PV) Inverter System: A Review. Energies 2024, 17, 1571. [Google Scholar] [CrossRef]
  2. Haegel, N.M.; Atwater, H., Jr.; Barnes, T.; Breyer, C.; Burrell, A.; Chiang, Y.-M.; De Wolf, S.; Dimmler, B.; Feldman, D.; Glunz, S.; et al. Terawatt-Scale Photovoltaics: Transform Global Energy. Science 2019, 364, 836–838. [Google Scholar] [CrossRef] [PubMed]
  3. de Faria, H., Jr.; Trigoso, F.B.M.; Cavalcanti, J.A.M. Review of Distributed Generation with Photovoltaic Grid-Connected Systems in Brazil: Challenges and Prospects. Renew. Sustain. Energy Rev. 2017, 75, 469–475. [Google Scholar] [CrossRef]
  4. Zhang, Z.; Qian, Z.; Chen, M.; Zhu, R.; Zhang, F.; Zhong, T.; Lin, J.; Ning, L.; Xie, W.; Creutzig, F.; et al. Worldwide Rooftop Photovoltaic Electricity Generation May Mitigate Global Warming. Nat. Clim. Change 2025, 15, 393–402. [Google Scholar] [CrossRef]
  5. El-Banby, G.M.; Moawad, N.M.; Abouzalm, B.A.; Abouzaid, W.F.; Ramadan, E.A. Photovoltaic System Fault Detection Techniques: A Review. Neural Comput. Appl. 2023, 35, 24829–24842. [Google Scholar] [CrossRef]
  6. Venkatakrishnan, G.R.; Rengaraj, R.; Tamilselvi, S.; Harshini, J.; Sahoo, A.; Saleel, C.A.; Abbas, M.; Cuce, E.; Jazlyn, C.; Shaik, S.; et al. Detection, Location, and Diagnosis of Different Faults in Large Solar PV System—A Review. Int. J. Low-Carbon Technol. 2023, 18, 659–674. [Google Scholar] [CrossRef]
  7. Pillai, D.S.; Rajasekar, N. A Comprehensive Review on Protection Challenges and Fault Diagnosis in PV Systems. Renew. Sustain. Energy Rev. 2018, 91, 18–40. [Google Scholar] [CrossRef]
  8. Lawani, A.O.; Adeleke, T.B.; Osuizugbo, I.C. Solar Photovoltaic Systems: A Review of Risks, Fault Detection, and Mitigation Strategies. In Proceedings of the 2024 IEEE 5th International Conference on Electro-Computing Technologies for Humanity (NIGERCON), Ado Ekiti, Nigeria, 26–28 November 2024; pp. 1–5. [Google Scholar]
  9. Panasetsky, D.; Sidorov, D.; Li, Y.; Ouyang, L.; Xiong, J.; He, L. Centralized Emergency Control for Multi-Terminal VSC-Based Shipboard Power Systems. Int. J. Electr. Power Energy Syst. 2019, 104, 205–214. [Google Scholar] [CrossRef]
  10. Gong, B.; An, A.; Shi, Y.; Zhang, X. Fast Fault Detection Method for Photovoltaic Arrays with Adaptive Deep Multiscale Feature Enhancement. Appl. Energy 2024, 353 Pt A, 122071. [Google Scholar] [CrossRef]
  11. Qian, H.; Lee, B.; Wu, Z.; Wang, G. Research on DC Arc Fault Detection in PV Systems Based on Adjacent Multi-Segment Spectral Similarity and Adaptive Threshold Model. Sol. Energy 2023, 264, 112011. [Google Scholar] [CrossRef]
  12. Hong, Y.-Y.; Pula, R.A. Methods of Photovoltaic Fault Detection and Classification: A Review. Energy Rep. 2022, 8, 5898–5929. [Google Scholar] [CrossRef]
  13. Islam, M.; Rashel, M.R.; Ahmed, T.; Islam, A.K.M.K.; Tlemçani, M. Artificial Intelligence in Photovoltaic Fault Identification and Diagnosis: A Systematic Review. Energies 2023, 16, 7417. [Google Scholar] [CrossRef]
  14. Gnetchejo, P.J.; Essiane, S.N.; Ele, P.; Dadjé, A.; Chen, Z. Faults Diagnosis in a Photovoltaic System Based on Multivariate Statistical Analysis. Energy Sources Part A Recovery Util. Environ. Eff. 2025, 47, 7383–7404. [Google Scholar] [CrossRef]
  15. Hichri, A.; Hajji, M.; Mansouri, M.; Bouzrara, K. Deep Learning Based Fault Diagnosis in Grid-Connected Photovoltaic Systems. In Proceedings of the 19th IEEE International Multi-Conference on Systems, Signals & Devices, Setif, Algeria, 6 May 2022; pp. 123–124. [Google Scholar]
  16. Zhang, Z.; Ma, M.; Wang, H.; Wang, H.; Ma, W.; Zhang, X. A Fault Diagnosis Method for Photovoltaic Module Current Mismatch Based on Numerical Analysis and Statistics. Sol. Energy 2021, 225, 221–236. [Google Scholar] [CrossRef]
  17. Hajji, M.; Yahyaoui, Z.; Mansouri, M.; Nounou, H.; Nounou, M. Fault Detection and Diagnosis in Grid-Connected PV Systems Under Irradiance Variations. Energy Rep. 2023, 9, 4005–4017. [Google Scholar] [CrossRef]
  18. Hashunao, S.; Mehta, R.K. Fault Analysis of Solar Photovoltaic System. In Proceedings of the 5th International Conference on Renewable Energies for Developing Countries (REDEC), Marrakech, Morocco, 29–30 June 2020; pp. 1–6. [Google Scholar]
  19. Taye, M.M. Understanding of Machine Learning with Deep Learning: Architectures, Workflow, Applications and Future Directions. Computers 2023, 12, 91. [Google Scholar] [CrossRef]
  20. Gong, B.; An, A.; Shi, Y.; Jia, W. Fault Diagnosis of Photovoltaic Array with Multi-Module Fusion Under Hyperparameter Optimization. Energy Convers. Manag. 2024, 319, 118974. [Google Scholar] [CrossRef]
  21. Zyout, I.; Oatawneh, A. Detection of PV Solar Panel Surface Defects Using Transfer Learning of the Deep Convolutional Neural Networks. In Proceedings of the Advances in Science and Engineering Technology International Conferences (ASET), Dubai, United Arab Emirates, 16 June 2020; pp. 1–4. [Google Scholar]
  22. Akhter, M.N.; Mekhilef, S.; Mokhlis, H.; Almohaimeed, Z.M.; Muhammad, M.A.; Khairuddin, A.S.M.; Akram, R.; Hussain, M.M. An Hour-Ahead PV Power Forecasting Method Based on an RNN-LSTM Model for Three Different PV Plants. Energies 2022, 15, 2243. [Google Scholar] [CrossRef]
  23. Huang, C.; Yang, M. Memory Long and Short Term Time Series Network for Ultra-Short-Term Photovoltaic Power Forecasting. Energy 2023, 279, 127961. [Google Scholar] [CrossRef]
  24. Massaoudi, M.; Chihi, I.; Abu-Rub, H.; Refaat, S.S.; Oueslati, F.S. Convergence of Photovoltaic Power Forecasting and Deep Learning: State-of-Art Review. IEEE Access 2021, 9, 136593–136615. [Google Scholar] [CrossRef]
  25. He, J.; Vashishtha, G. Construction and Simulation of a Joint Scale Model for Power Electronic Converters Based on Wavelet Decomposition and Reconstruction Algorithms. PLoS ONE 2024, 19, e0298590. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, Z.; Li, F.; Cong, R.; Bai, H.; Zhao, Y. Adaptive Feature Fusion Network Based on Boosted Attention Mechanism for Single Image Dehazing. Multimed. Tools Appl. 2022, 81, 11325–11339. [Google Scholar] [CrossRef]
  27. He, J.; Cui, J.; Zhang, G.; Xue, M.; Chu, D.; Zhao, Y. Spatial–Temporal Seizure Detection with Graph Attention Network and Bi-Directional LSTM Architecture. Biomed. Signal Process. Control 2022, 78, 103908. [Google Scholar] [CrossRef]
  28. Ahsan, S.B.; Ikhalas, M.; Khan, M.M.; Ullah, S.; Zaheer, M.Z. ARDGen: Augmentation Regularization for Domain-Generalized Medical Report Generation. In Proceedings of the Computer Vision and Pattern Recognition Conference, Vancouver, BC, Canada, 15 June 2025; pp. 6524–6533. [Google Scholar]
  29. Wang, B.; Zhao, Z.; Chen, Y.; Yu, J. A Novel Robust Point Cloud Fitting Algorithm Based on Nonlinear Gauss–Helmert Model. IEEE Trans. Instrum. Meas. 2023, 72, 1–12. [Google Scholar] [CrossRef]
  30. Bakdi, A.; Bounoua, W.; Guichi, A.; Mekhilef, S. Real-Time Fault Detection in PV Systems Under MPPT Using PMU and High-Frequency Multi-Sensor Data Through Online PCA-KDE-Based Multivariate KL Divergence. Int. J. Electr. Power Energy Syst. 2021, 125, 106457. [Google Scholar] [CrossRef]
Figure 1. Advanced CNN-BiLSTM network architecture diagram.
Figure 1. Advanced CNN-BiLSTM network architecture diagram.
Sensors 25 04474 g001
Figure 2. Conceptual diagram of the Progressive Difficulty Validation Framework.
Figure 2. Conceptual diagram of the Progressive Difficulty Validation Framework.
Sensors 25 04474 g002
Figure 3. Overall experimental flowchart.
Figure 3. Overall experimental flowchart.
Sensors 25 04474 g003
Figure 4. Feature correlation matrix.
Figure 4. Feature correlation matrix.
Sensors 25 04474 g004
Figure 5. Multi-metric performance comparison of all methods under hard difficulty conditions.
Figure 5. Multi-metric performance comparison of all methods under hard difficulty conditions.
Sensors 25 04474 g005
Figure 6. Training efficiency vs. performance.
Figure 6. Training efficiency vs. performance.
Sensors 25 04474 g006
Figure 7. Overall performance improvement.
Figure 7. Overall performance improvement.
Sensors 25 04474 g007
Figure 8. Robustness analysis across multiple dimensions.
Figure 8. Robustness analysis across multiple dimensions.
Sensors 25 04474 g008
Figure 9. Robustness performance drop (easy–hard).
Figure 9. Robustness performance drop (easy–hard).
Sensors 25 04474 g009
Figure 10. (a) Performance under varying noise levels. (b) Comprehensive radar plot comparing key performance indicators against the baseline model.
Figure 10. (a) Performance under varying noise levels. (b) Comprehensive radar plot comparing key performance indicators against the baseline model.
Sensors 25 04474 g010
Table 1. Comprehensive performance comparison across progressive difficulty levels.
Table 1. Comprehensive performance comparison across progressive difficulty levels.
DifficultyMethodAccuracyPrecisionRecallF1-ScoreKappaMCCTime (s)
EasyAdvanced CNN-BiLSTM0.9570.9750.9570.9520.9550.9579.19
CNN0.9600.9650.9600.9600.9570.9582.73
CNN-BiLSTM0.9070.8820.9070.8860.9010.9052.58
BiLSTM0.6460.5450.6460.5600.6230.6422.24
XGBoost0.9720.9720.9720.9720.9700.9700.86
MediumAdvanced CNN-BiLSTM1.0001.0001.0001.0001.0001.00010.73
CNN-BiLSTM1.0001.0001.0001.0001.0001.0004.08
CNN0.9980.9980.9980.9980.9970.9972.15
BiLSTM0.7090.6550.7090.6390.6900.7023.77
XGBoost0.9570.9570.9570.9570.9540.9541.19
HardAdvanced CNN-BiLSTM0.8330.7560.8330.7870.8100.8145.04
CNN0.6450.5050.6450.5500.5900.6261.12
CNN-BiLSTM0.3790.3800.3790.3070.2840.3451.29
BiLSTM0.2460.0850.2460.1210.1310.1961.12
Random Forest0.9190.9240.9190.9180.9110.9120.33
Table 2. Statistical significance analysis for hard difficulty conditions.
Table 2. Statistical significance analysis for hard difficulty conditions.
Comparisont-Statisticp-ValueCohen’s dEffect SizeSignificance Level
Advanced vs. CNN29.87<0.00111.78LargeHighly Significant
Advanced vs. BiLSTM130.41<0.00143.82LargeHighly Significant
Advanced vs. CNN-BiLSTM66.82<0.00126.70LargeHighly Significant
Advanced vs. Random Forest−12.45<0.0014.98LargeHighly Significant
Table 3. Ablation study results—component contribution analysis.
Table 3. Ablation study results—component contribution analysis.
ComparisonAccuracyContributionPercentageEffect SizeSignificance Level
Baseline CNN-BiLSTM0.3793--BaselineStandard sequential CNN-BiLSTM
+Multi-Scale Parallel CNN0.50620.126928.0%HighParallel CNN branches (k = 3, 7, 15, 31)
+Adaptive Feature Fusion0.60590.099722.0%HighMulti-head attention fusion
+Hierarchical Attention0.68750.081618.0%HighTemporal attention mechanism
+Deep BiLSTM Architecture0.75550.068015.0%MediumTwo-layer BiLSTM enhancement
+Residual Connections0.80990.054412.0%MediumSkip connections and normalization
+Advanced Optimization0.83250.02275.0%LowAdamW, cosine annealing, label smoothing
Total Improvement0.83250.4532119.48%AdvancedComplete proposed architecture
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Jing, B.; Jiao, X.; Yao, R. Advanced Multi-Scale CNN-BiLSTM for Robust Photovoltaic Fault Detection. Sensors 2025, 25, 4474. https://doi.org/10.3390/s25144474

AMA Style

Zhang X, Jing B, Jiao X, Yao R. Advanced Multi-Scale CNN-BiLSTM for Robust Photovoltaic Fault Detection. Sensors. 2025; 25(14):4474. https://doi.org/10.3390/s25144474

Chicago/Turabian Style

Zhang, Xiaojuan, Bo Jing, Xiaoxuan Jiao, and Ruixu Yao. 2025. "Advanced Multi-Scale CNN-BiLSTM for Robust Photovoltaic Fault Detection" Sensors 25, no. 14: 4474. https://doi.org/10.3390/s25144474

APA Style

Zhang, X., Jing, B., Jiao, X., & Yao, R. (2025). Advanced Multi-Scale CNN-BiLSTM for Robust Photovoltaic Fault Detection. Sensors, 25(14), 4474. https://doi.org/10.3390/s25144474

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop