Next Article in Journal
Evaluating Intralingual Machine Translation Quality: Application of an Adapted MQM Scheme to German Plain Language
Previous Article in Journal
Insights on the Pedagogical Abilities of AI-Powered Tutors in Math Dialogues
Previous Article in Special Issue
AI/ML Based Anomaly Detection and Fault Diagnosis of Turbocharged Marine Diesel Engines: Experimental Study on Engine of an Operational Vessel
error_outline You can access the new MDPI.com website here. Explore and share your feedback with us.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TransferLearning-Driven Large-Scale CNN Benchmarking with Explainable AI for Image-Based Dust Detection on Solar Panels

1
Department of Computer Science, National University of Computer and Emerging Sciences (NUCES-FAST), Jamrud Road 160 Industrial Estate Road, Phase 1 Hayatabad, Peshawar 25100, Pakistan
2
Interdisciplinary Center for Digital Humanities and Social Sciences, Friedrich-Alexander University, 91054 Erlangen, Germany
Information 2026, 17(1), 52; https://doi.org/10.3390/info17010052
Submission received: 8 December 2025 / Revised: 27 December 2025 / Accepted: 3 January 2026 / Published: 6 January 2026

Abstract

Solar panel power plants are typically established in regions with maximum solar irradiation, yet these conditions result in heavy dust accumulation on the panels causing significant performance degradation and reduced power output. The paper addresses this issue via an image-based dust detection solution powered by deep learning, particularly convolutional neural networks (CNNs). Most of such solutions use state-of-the-art CNNs either as backbones/features extractors, or propose custom models built upon them. Given such a reliance, future research requires a comprehensive benchmarking of CNN models to identify the ones that achieve superior performance on classifying clean vs. dusty solar panels both with respect to accuracy and efficiency. To this end, we evaluate 100 CNN models that belong to 16 families for image-based dust detection on solar panels, where the pre-trained models of these CNN architectures are used to encode solar panel images. Upon these image encodings, we then train and test a linear support vector machine (SVM) to determine the best-performing models in terms of classification accuracy and training time. The use of such a simple classifier ensures a fair comparison where the encodings do not benefit from the classifier itself and their performance reflects each CNN’s ability to capture the underlying image features. Experiments were conducted on a publicly available dust detection dataset, using stratified shuffle-split with 70–30, 80–20, and 90–10 splits, repeated 10 times. convnext_xxlarge and resnetv2_152 achieved the best classification rates of above 90%, with resnetv2_152 offering superior efficiency that is also supported by features analysis such as tSNE and UMAP, and explainableAI (XAI) such as LIME visualization. To prove their generalization capability, we tested the image encodings of resnetv2_152 on an unseen real-world image dataset captured via a drone camera, which achieved a remarkable accuracy of 96%. Consequently, our findings guide the selection of optimal CNN backbones for future image-based dust detection systems.

1. Introduction

Image classification, a subfield of computer vision, was traditionally based on hand-crafted feature extraction [1], such as color histograms [2], edge detection filters [3], local feature descriptors [4], and texture measures [5]. Once extracted from images, these features were used to train machine learning algorithms, such as support vector machines (SVMs) or k-nearest neighbors (k-NN), for classification. However, the rise of deep learning algorithms, particularly convolutional neural networks (CNNs) [1], has transformed image classification by automating feature extraction [6] and allowing hierarchical representations to be learned directly from raw image data [7]. This shift improved both the accuracy and scalability of classification frameworks [8,9].
The devastating drawbacks of fossil-fuel-based power generation [10] have led to a global shift toward renewable energy. Figure 1 illustrates this trend, showing that solar power remains the top contributor to sustainable energy. Figure 1a highlights solar photovoltaic (PV) as the dominant source of renewable energy in 2023, with an over 60% share across most regions except Africa. Figure 1b shows solar PV’s exponential growth in global capacity since 2019, while Figure 1c depicts a 44.2% drop in solar module prices in 2024, signaling significant technological advancement. These trends position solar PV as the cornerstone of renewable strategies and a key alternative to traditional power generation.
Despite the promising growth of solar PV, its operational efficiency is highly sensitive to environmental factors. For efficient power generation, solar panels are often installed in locations that receive extensive sunlight, such as arid and semi-arid regions [11]. However, in such areas, dust accumulation on the surface of solar panels is a persistent problem that reduces PV efficiency due to sunlight obstruction [12]. This issue becomes even more severe because these regions receive relatively low annual rainfall, which limits the natural cleaning of panels. Consequently, maintaining optimal solar panel performance in these environments requires innovative solutions to minimize the impact of dust and ensure consistent energy output.
To address these challenges, traditional manual cleaning methods are often employed; however, while effective, they are labor-intensive, water-intensive, and costly when performed regularly without assessing actual need. Automated cleaning systems [13], though promising, can also be resource-intensive if not optimized for need-based operation. Consequently, recent advancements [14] propose the use of automated image-based monitoring systems to identify exactly when panels require cleaning, thus ensuring efficient and timely maintenance. This approach offers a sustainable solution that reduces operational costs and resource consumption while maintaining optimal energy output. This paper proposes to extend this line of research and perform a comprehensive evaluation of the recently proposed CNN-based deep learning models for the task of dust detection on solar panels. Our approach is to encode/represent images using CNN models that are pre-trained on large-scale generic datasets such as ImageNet. These image encodings are then used for the task of image classification to detect dusty solar panels. We propose to evaluate the performance of 100 pre-trained CNN models both with respect to classification accuracy and computational efficiency. In addition to performance evaluation, we employ explainable AI (XAI) to verify whether the image classification results are driven by meaningful visual evidence on the solar panels rather than spurious background cues. This analysis helps in understanding the underlying causes of misclassifications and in assessing the reliability/quality of the image encodings. Nonetheless, such XAI-based verification is diagnostic in nature, while the design of corrective or mitigation strategies lies beyond the scope of this work.

Research Gap, Questions, Contributions, and Scope

Although CNNs have been widely employed for image-based dust detection on solar panels using standard architectures such as ResNet, VGG, and DenseNet, there exists no comprehensive benchmarking of recent state-of-the-art CNN models on publicly available solar panel datasets. As a result, the research community lacks a clear understanding of which CNN variants are most effective in terms of both classification accuracy and computational efficiency for this specific task. Furthermore, prior studies often focus on proposing new models or adopting existing ones without systematically comparing them across common evaluation settings. This limits fair comparisons and makes it difficult to understand the trade-offs involved when selecting a CNN backbone for real-world deployment. Additionally, many studies emphasize accuracy while ignoring key efficiency metrics such as inference time and FLOPs, which are essential for scalable and real-time deployment in the solar energy sector.
To address these limitations, this paper rigorously benchmarks 100 pre-trained CNN models from 16 architecture families on a real-world solar panel dust detection dataset. We assess each model based on its classification performance and computational cost to establish a reliable reference for future work in this domain.
Accordingly, the following research questions are addressed in the paper:
  • RQ-1: Which pre-trained CNN architectures achieve the highest classification accuracy for image-based dust detection on solar panels?
  • RQ-2: How do the evaluated CNN models compare in terms of computational efficiency, including training time, inference time for a single image, and FLOPs?
  • RQ-3: Which CNN models offer the best trade-off between classification performance and computational cost, making them suitable for real-time or resource-constrained deployment?
To answer these questions, the paper makes the following key contributions:
  • It presents the first large-scale benchmarking of 100 pre-trained CNN models from 16 architecture families for image-based dust detection on solar panels, providing a comprehensive comparative analysis of backbone performance.
  • It reports classification accuracy along with computational metrics such as training time and FLOPs to highlight the practical trade-offs between model performance and efficiency.
  • The benchmarking is conducted using a reproducible and modular evaluation pipeline, offering a standardized baseline for future research in solar panel monitoring using computer vision.
To frame these contributions appropriately, we briefly outline the scope and limitations of this paper:
  • We do not follow a multi-class classification approach for identifying “dirty” solar panels although the dirt may be of several types such as bird-droppings, debris, leaves, occlusion and defects, etc. Instead, we perform image-based binary classification to classify the solar panels as either “clean” or “dirty”. The details on the number of images per the dirt type is given in Section 3.
  • We do not propose a novel CNN architecture, nor do we aim to achieve the best results on the current dataset by fine-tuning a given CNN architecture to mitigate misclassifications. Rather, our focus is on benchmarking pre-trained CNN backbones, which is considered a compulsory preliminary step. This is because newly proposed CNNs are often derived from well-known architectures such as ResNet or VGG. Identifying the best backbone for a given problem can then lead to improved results through the discriminative features it extracts. Given the vast number of available CNN architectures, evaluating all of them is impractical, while evaluating too few would fail to capture meaningful comparative insights. To ensure fairness, diversity, and comprehensiveness, we therefore evaluate 100 representative models from 16 distinct CNN families, providing a solid empirical foundation for future model development in this domain.
  • We do not employ any additional sensing modalities such as infrared or electrostatic/deposition sensors—nor do we attempt to measure or estimate physical dust layer thickness μm, mass density (mg/cm2), or other quantitative deposition metrics. Our proposed classification pipeline is based on RGB images where the labels are assigned purely by visual criteria.
  • We do not conduct any end-to-end energy or cost-savings analysis of vision-based cleaning systems. Consequently, the paper does not deal with or present hardware deployment, real-time inference pipeline, or field-trial validation.
  • By clearly bounding the contribution of this paper to a comprehensive, reproducible comparison of 16 state-of-the-art CNN backbones, this paper provides a solid foundation upon which future research can achieve the following:
    integrate the top-performing models into full monitoring systems;
    extend the image dataset to accommodate the dirt types for multi-class classification;
    incorporate additional sensors;
    quantify operational and energy benefits.

2. Related Work

This section is divided into two subsections. The Section 2.1 presents an overview of recent studies that highlight the impact of dust pollution on the power generation capability of solar panels. The Section 2.2 reviews the latest image-based dust detection methods, which are essential components of a comprehensive dust monitoring and removal system for solar panels.

2.1. Impact of Dust Accumulation on Solar Panel Efficiency: A Global Perspective

To underscore the global relevance of dust accumulation on solar panel performance, we present findings from recent studies conducted in diverse regions:
  • India: In Dehradun, a 45-day exposure led to an average efficiency drop of 22.5% in polysilicon solar panels [15].
  • Jordan: At the Jordan University of Science and Technology, natural dust accumulation over three months resulted in a 13% decline in PV module output power [16].
  • Pakistan: Studies in Islamabad and Bahawalpur revealed that dust densities of 6.388 g/m2 and 4.365 g/m2 led to efficiency reductions of 15.08% and 12.61%, respectively, [17].
  • China: Research indicated that a one-micron dust layer could cause a 25.5% reduction in PV module efficiency, with a 70-day exposure leading to a 21.47% power output decrease [18].
  • Global Review: A comprehensive analysis reported that dust accumulation could reduce PV efficiency by up to 64%, with factors like tilt angle and environmental conditions playing significant roles [19].
These studies highlight the critical impact of dust on solar panel performance across various climatic regions, emphasizing the need for effective dust mitigation strategies.

2.2. Image-Based Dust Detection on Solar Panels

To mitigate the performance degradation of solar PVs caused by dust accumulation, researchers have devised solutions based on various technologies such as electrodynamic dust demoval (EDR) [20], digital image processing [21], and more recently, using image-based deep learning [14]. Since this paper extends the deep learning-based line of research, some recent methods are discussed in this section that utilizes deep learning.
MobileNetV1, V2, and V3 are used for classifying dusty and clean solar panel surfaces where MobileNetV1 achieves the best accuracy of 91.25%, demonstrating its feasibility for real-time applications [22]. A custom CNN model named “SolNet” is proposed to detect dusty solar panels and is shown to surpass state-of-the-art CNN models such as AlexNet and ResNet on a dataset that lacks challenges such as background clutter, partial occlusion and view-point changes [23]. A two-stage and end-to-end pipeline using a customized YOLOv5 model is proposed to detect and localize dust and bird droppings in UAV images of solar panels [24]. The comparison of semantic segmentation methods for dust detection is performed to demonstrate that deep learning methods outperform traditional machine learning in accuracy and robustness [25]. Addressing the unique challenges of offshore environments, a further study presents a framework that leverages image processing (HLS colorspace) and deep learning (Mask R-CNN), for effective dust detection on floating solar panels [26]. Deep learning combined with multi-view virtual reality imaging is proposed to provide a comprehensive assessment of dust accumulation patterns on solar panels [27]. Various deep learning architectures such as ResNet, EfficientNet, InceptionNetV3, MobileNetV2, and VGG are evaluated for the detection of dust on solar panels, identifying models that strike a balance between accuracy and computational efficiency [28]. CASolarNet [29] is proposed by incorporating channel attention into an EfficientNet-based architecture which is shown to enhance dust detection accuracy on solar panels. Various deep learning strategies such as EfficientNetB3, ResNet50, MobileNet, VGG19, Xception, InceptionResNetV2, VGG16, ResNet101, DenseNet201, and EfficientNetB7 are comprehensively evaluated to detect dust on solar panels, underscoring the importance of timely maintenance to maintain optimal performance [30]. A novel method for classifying photovoltaic (PV) system losses using thermographic non-destructive tests (TNDTs) combined with a CNN is proposed [31]. The method is shown to rapidly detect efficiency losses caused by dirt and malfunctions, achieving 98% accuracy with significant advantages in speed, cost reduction, and minimized electricity production losses. DVNET [32], an end-to-end solar panels dust detection model based on light transmittance estimation, is proposed. It uses image processing and attention mechanisms and is shown to accurately quantify dust density and generate transmittance maps thus demonstrating efficiency, scalability, and real-world applicability. An improved Adam optimizer for dust detection on solar photovoltaic panels is presented that integrates Warmup technology and cosine annealing to enhance training stability and generalization [33]. The proposed method is shown to outperform the standard Adam optimizer across multiple architectures including ResNet-18, VGG-16, and MobileNetV2 demonstrating superior convergence and accuracy, with potential economic and strategic benefits for solar energy maintenance. Previously, the current authors also investigated pre-trained CNN models such as AlexNet, DenseNet, ResNet, MobileNet, etc. combined with variants of SVM classifier for dust detection; in this investigation, DenseNet169 achieved notable accuracy, suggesting a cost-effective maintenance solution [14]. Since, our main focus is on the use of CNN models both custom and pre-trained models, in Table 1, we summarize the most relevant works that have utilized or evaluated multiple CNN models for the task of image-based dust detection on solar panels.
However, none of these studies have systematically compared a broad range of CNN architectures for dust detection, particularly with respect to balancing classification accuracy and computational cost. To address this gap, we present an extensive benchmark of 100 CNN models spanning 16 architecture families, evaluating each model’s dust-detection accuracy and computational efficiency.

3. Dataset

The dataset originates from a publicly available online repository [39] that comprises images depicting both clean and dusty solar panels. Overall, the dataset contains 1068 images, with 405 images representing clean solar panels and 664 images representing panels affected by dust and other contaminants. However, the originally uploaded dataset included more images than these 1068, where 1493 images depict clean solar panels while 1069 images belong to dusty solar panels. We performed a manual inspection of the complete dataset for potential duplicates, as there were images of the same scene captured at varying scales. This may have occurred because both a thumbnail and the original image were downloaded by the web crawler, each with a different name. Such cleaning is of utmost importance because the images are re-sized to a fixed size before being encoded with a pre-trained model. Hence, variants of the same image at different scales are all re-sized to the same dimensions and then encoded leading to biased classification results.
This dataset comes with immense challenges due to image variations some of which are depicted in Figure 2. The images are captured from a variety of viewpoints and under diverse daytime lighting conditions. Additionally, for the dusty panels, the dataset includes a range of images exhibiting different types and patterns of dirt accumulation. Interestingly, there are images where the cleaning process is shown, such that some solar panels are being cleaned while some of them are still dirty. Consequently, all these image variations make the image dataset challenging for recent state-of-the-art CNN models.
Based on these observations, the images in the dataset are classified purely on visual hallmarks rather than on quantifiable measurements such as dust-layer thickness in micrometers or mass density in milligrams per square centimeter. We therefore employ solely visual criteria for labeling “clean” and “dirty” panels in this Internet-sourced image dataset. The clean solar panel images predominantly depict new or freshly installed solar panels that have just undergone professional cleaning, characterized by uniform specular highlights across the entire glass surface and the absence of smudges, streaks, particulate clusters, or other irregularities under standard outdoor illumination. In contrast, the dusty solar panel images capture the real-world conditions in which a panel’s surface is occluded by environmental and biological contaminants such as dust and dirt deposits, bird droppings, fungal growths, and organic debris like leaves or pollen resulting in non-uniform reflectance and mottled shading over the affected areas. For a clearer picture, Table 2 shows the number of images in the dataset that represent each type of contaminant. This breakdown is provided for information only and is not used in any part of the experiments because the paper strictly deals with binary image-based classification of solar panels, i.e., categorizing them as either clean or dirty, regardless of the type of dirt present on the panel. Such a so-called coarse-grained binary classification serves as a first step to identify the dirty solar panels regardless of the type of the contaminant, as the cleaning procedures can be different for different dirt types. Consequently, such training on a diverse array of dirty solar panel images can ensure that the model identifies “dirtiness” in all its forms, creating a foundation that can be extended into multi-class classification framework using fine-grained classification [40].

4. Methodology

Her, we give a brief description of the proposed methodology, which involves encoding images using pre-trained CNN models and then using these encoded images to train and test a linear SVM using a stratified shuffle-split strategy as shown in Figure 3.

4.1. Image Encoding Using Pre-Trained CNN Model

In the first part of the methodology, a pre-trained CNN model is used to encode the solar panel images. Instead of training from scratch, a CNN model in its pre-trained form can be used as a features extractor, thus leveraging the knowledge that it has gained already from large-scale image datasets such as ImageNet, a process more commonly known as “transfer learning”. For a given image, the output from the final convolutional stage is a set of feature maps upon which, a global average pooling (GAP) layer is applied. This operation reduces each feature map to a single scalar by computing the average across all spatial locations. Consequently, the GAP layer transforms the set of feature maps into a compact feature vector. This vector encapsulates the most salient visual features of the image while significantly reducing the number of parameters and mitigating any overfitting that may occur due to the high dimensional feature maps. Figure 4 outlines the details of the CNN families and their variants used for image encoding, totaling 100 pre-trained models. Each CNN family alongside its variants is shown along with their corresponding feature-vector dimensions obtained via GAP, the number of floating point operations in GFLOPs, and their parameter numbers.

4.2. Classification Using a Linear SVM via Stratified Shuffle Split

The encoded image dataset is used as input to a linear support vector machine (SVM) classifier. The selection of a linear SVM is motivated by its reduced computational cost while being effective in high-dimensional space. In its linear setting, the SVM classifier learns to distinguish between the two classes, such as clean and dirty solar panels, by finding the optimal hyperplane that maximizes the margin between the two classes. The dataset is partitioned into training and test sets using a stratifiedshufflesplit strategy that ensures uniform class distribution across both the sets. This avoids biases that may result from imbalanced class representations. We also propose to use three splitting ratios which are 70% training with 30% testing, 80% training with 20% testing, and 90% training with 10% testing to analyze the impact of training set size on the achieved results. Consequently, we report the classification accuracy, precision, recall and F1-score, while also aim to calculate the training time taken by the linear SVM on image encodings generated by different CNN models.
Nonetheless, the use of a pre-trained CNN for image encoding coupled with a lightweight SVM classifier is a deliberate choice intended to minimize the local processing burden. This architecture provides a streamlined pipeline that is well-suited for future deployment on edge-computing hardware, where high-dimensional encodings can be processed with low latency to meet the demands of real-time monitoring.

5. Results and Discussion

Due to the large number of CNNs evaluated, we present the results in a systematic manner while give their details in the Appendix A. Since, we evaluate 100 CNN models that belong to 16 families, we explain and visually demonstrate the variants that are the best performers among their family members for each of the three settings of the data split. The same is done for reporting the training time where we show the variants that take the least time among family members.

5.1. Classification Accuracy

For the 100 pre-trained CNN models that span 16 families, we select, for each family, the best-performing variant based on mean classification accuracy across all the three data splits (70–30, 80–20, and 90–10). The results are shown in Figure 5 where modern architectures such as ConvNeXt, ResNetV2, and RegNetY show superior performance by achieving mean accuracies above 91%, with ConvNeXt peaking at 92.43% on the 90–10 split. Other families, including DenseNet and ConvNeXtV2, maintain competitive accuracies in the high 80s, while mid-tier performance is observed with models like ResNet, DPN, and Inception-Next, which stabilize at around 86–87%. On the lower end, older or less optimized families—such as CS3DarkNet, ShuffleNet, VGG, and even traditionally popular models like MobileNetV2, Inception, and GoogLeNet—generally achieve accuracies in the low to mid-80s, and models like Wide-ResNet, Xception, ResNeXt, and EfficientNetV2 achieve around 75–79%.
While increasing the training data consistently boosts performance across families, the relative ranking among those families remains largely stable. This highlights that while data is important, architectural design remains a crucial factor. These results clearly show that state-of-the-art CNN architectures that incorporate contemporary innovations in feature extraction are best suited as backbone feature extractors for the task of image-based dust detection on solar panels. Our results are particularly significant because many previous approaches utilize the older architectures as shown in Table 1, while these models perform adequately with sufficient data, they ultimately struggle to achieve significant performance gains.

5.2. Training Time

Figure 6 illustrates the training time performance of various model families when their image encodings are used to train a linear SVM. In this case, for each of the 16 families, the variant that achieves the lowest training time is selected. Similar to the classification accuracy, the training time is also measured for three distinct data splits (70–30, 80–20, and 90–10). These splits are denoted as “split-70”, “split-80”, and “split-90”, respectively. The values shown in the figure (e.g., 0.45, 0.54, 0.69 for one family) represent the mean training time taken by the linear SVM when trained on feature vectors derived from image encodings produced by that specific model family. Notably, the figure emphasizes the computational efficiency of certain families, such as inception_next, cs3darknet, and mobilenet_v3. These models consistently achieve lower training times even as the amount of training data increases, demonstrating their feasibility for applications where rapid model deployment or frequent retraining is essential.
In addition to the raw training time metrics, we employ a stratified shuffle-split method to partition the dataset, ensuring a consistent class distribution (e.g., clean versus dirty solar panels) across both the training and test sets. Doing so mitigates the potential bias that may occur due to class imbalance, providing a fair basis for comparing the computational efficiency of each model family. Figure 6 shows that while training time scales with data size, the rate of growth varies significantly across model families. Some architectures exhibit only a modest rise, whereas others show a much steeper increase. On the one hand, this analysis highlights the importance of efficient feature extraction models for reducing computational load; on the other hand, it identifies CNN model families that provide an optimal balance between classification performance and training time. These combined insights enable future researchers to select the CNN backbone for their custom architectures to perform image-based recognition of dusty solar panels in a more accurate and computationally efficient manner.

5.3. Overall Verdict

The extensive evaluation of 100 pre-trained CNN models across 16 families demonstrates the trade-offs between classification accuracy and training time. Models from modern families such as ConvNeXt and ResNetV2 not only consistently deliver top-tier accuracy (exceeding 91% on average across various training splits) but also demonstrate efficient learning dynamics that minimize training time. In contrast, while models such as DenseNet, ConvNeXtV2, and ResNet maintain competitive accuracy in the high 80s, their training times are moderately higher, making them better candidates for scenarios where slight increases in computational cost can be justified by incremental gains in precision. In addition to these findings, certain lightweight models like MobileNetV2, Inception, and GoogLeNet, despite achieving slightly lower accuracy in the low to mid-80s, offer significant advantages in training and inference speed, making them appealing for real-time applications or environments with limited computational resources.

5.4. Comparison of the Best-Performing Variants

The best-performing CNN variants that consistently achieve a classification accuracy of above 90% throughout all the three splits are resnetv2_152 and convnext_xxlarge. A further and detailed comparison of these two CNN models was carried out to determine a clear winner on the current dataset for image-based dust detection on solar panels. Since both models achieve the best result on the 90–10 data split, here, we also split the dataset with a 90–10 stratified strategy. The images are encoded with each of the pre-trained models while global average pooling is applied on their output feature maps to bring encodings into vectorized form. These encodings are then used to train and test a linear SVM to show the classification accuracies achieved by each of the model. The elaborate comparison of the best CNN models is explained in the following.
  • Accuracy vs. Floating Point Operations (FLOPs): Figure 7a shows the accuracy vs. FLOP comparison of the two models, where resnetv2_152 clearly outperforms convnext_xxlarge with fewer FLOPs and better classification accuracy.
  • Accuracy vs. Parameters count: Figure 7b shows the same behavior for the accuracy vs. parameter countm thus determining that resnetv2_152 is superior to convnext_xxlarge.
  • Confusion Matrices: Since the main aim is to detect dusty/dirty solar panels, convnext_xxlarge performs slightly better than resnetv2_152 at detecting dirty solar panels as shown in Figure 7c, as the former accurately detects dirty solar panels better than the latter.
  • Manifold analysis using t-SNE and UMAP: We also performed a manifold analysis for both models using t-SNE [41] and uniform manifold approximation and projection for dimension reduction (UMAP) [42] to show the effectiveness of the CNN model to discriminate among the features of both the classes. The results are shown in Figure 7d for t-SNE and Figure 7e for UMAP, where the resnetv2_152 can clearly be observed to cluster features of the same class better than convnext_xxlarge, hence providing strong evidence for its superior performance.
  • Performance across all the three splits: To show the collective performance of both the best-performing models across all three splits, we performed the experiments with each of the three splits. We show the collective behavior of both CNN models in Figure 8a. The results highlight the stability and effectiveness of both models, with convnext_xxlarge displaying a slightly higher median and lower variance.
  • Other performance metrics: The precision–recall (PR) curves for the two models under the 90–10 split are shown in Figure 8b. Both models exhibit strong precision–recall performance, with resnetv2_152 maintaining higher precision at elevated recall levels. The receiver operating characteristic (ROC) curves under the same split are shown in Figure 8c, demonstrating excellent true positive rates for both models across all false positive rates, with convnext_xxlarge showing a marginal edge in ROC performance.
In conclusion, while both resnetv2_152 and convnext_xxlarge outperform the other 98 CNN models, a detailed comparison shows that resnetv2_152 clearly outperforms convnext_xxlarge both with respect to classification accuracy and efficiency, and hence, is highly recommended for the task of image-based dust detection on solar panels.

5.5. Qualitative Analysis Using Explainable-AI

We also performed a qualitative analysis to determine the image regions that represent the dusty solar panels. To this end, we use the explainable AI (XAI)-based local interpretable model-agnostic explanations (LIME) [43] to visualize the image regions that affect the decision of an SVM to classify the image as that of a dusty solar panel. The integration of explainable AI (XAI) frameworks is increasingly vital from the perspective of industrial applications to bridge the gap between “black-box” model performance and human-interpretable logic [44]. A qualitative check on the decision taken by the SVM using the image encodings of a pre-trained CNN validates their capacity to accurately identify dust accumulation on the surface of solar panels. Nonetheless, our aim is not segmentation of the image regions that correspond to dusty solar panels, but to assess the qualitative performance of the underlying image encodings. Accordingly, our intention is not to benchmark XAI methods against human judgment, but to use them as a qualitative visual tool to demonstrate that the learned image encodings consistently highlight image regions corresponding to dust on solar panels. For these visualizations, the image encoding is performed with convnext_xxlarge and then the 90–10 split is used to train and test the SVM. We then select the test images where the label prediction is done with the highest score by the trained SVM and use these images to perform the LIME visualization. Since, we explicitly deal with image-based dust detection on solar panels, Figure 9 only depicts the results of LIME visualization on dusty solar panels, where the decisive image patches are enclosed in thick boundaries. The top two rows of Figure 9, i.e., (a) and (b), clearly show that the dusty patches have influenced the decision of the SVM to decide the labels for these images. Some of the areas of the solar panels are clean and it can clearly be observed that those image regions are not detected by the LIME visualization, which means that these patches have not or least contributed to decide the label for these solar panels. Similarly, the bottom two rows of Figure 9, i.e., (c) and (d), show the effect of the context where some of the context is captured in the case of (c) while none is captured in case of (d) as that type of context, i.e., blue and clear sky, is mostly captured with clean solar panels. In addition to LIME, Figure 10 shows the qualitative analysis performed using randomized input sampling for explanation of black-box models (RISE) [45], demonstrating the effectiveness of the proposed method to detect the image patches related to dusty parts of the solar panels.

5.6. Analyzing Misclassified Images of Dirty Solar Panels

A detailed visual inspection of the misclassified test images was performed to further investigate the causes of failure cases, particularly for dirty solar panels. The sole purpose of this inspection was to reveal the causes of misclassification, whether they are related to imaging conditions such as non-uniform illumination, varied orientation, background clutter, or due to the type of dirt or the solar panel itself. We are interested in finding dirty solar panels that should be cleaned to avoid power losses. If a clean panel is incorrectly classified as dirty, it may simply lead to an unnecessary cleaning, which, in this context, is still acceptable.
To this end, we encode the images using the best-performing pre-trained CNN model, which is resnetv2_152. This encoded image dataset is then randomly divided into training and test sets using an 80–20 strategy in a stratified manner to account for the imbalance that exists in the dataset, as there are more images of dirty solar panels than the cleaned ones. With the training dataset, a linear SVM is trained, which is then tested with the test set. The images of the dirty solar panels that are wrongly classified as the clean panels are shown in Figure 11. We are particularly interested in such examples to identify the conditions that lead to such wrong classifications.
The first misclassified example depicted in Figure 11a clearly shows that a larger area of the solar panel is clean than dirty, and hence it is classified as clean rather than dirty. Figure 11b contains many dirty solar panels. However, it may have been misclassified due to the bright and blue sky context, as the clean solar panels are often imaged with such a background, particularly in the images of advertisements. Figure 11c is most likely misclassified due to the image being taken in the evening and under dark conditions. Figure 11d,e are imaged with severe background clutter, where the solar panels cover far fewer image pixels than the background. Finally, Figure 11f,g contain clean solar panels that have been wrongly labeled in the image dataset as dirty solar panels. Hence, it is concluded from the analysis of the wrongly classified images of dirty solar panels that the misclassifications are mostly caused by the imaging conditions rather than the type of dirt present on the solar panels.

5.7. Testing on Unseen Real-World Dirty Solar Panels Image Dataset

In order to evaluate on a real-world unseen test dataset, we collect 100 images of solar panels installed on rooftops at various places in Pakistan using a drone camera. These solar panels are mostly covered with dust and in some places with bird droppings. We encoded our main dataset, i.e., the one discussed in Section 3 with resnetv2_152 and then trained a linear SVM on these encodings. All 100 images of the unseen test dataset were then encoded with resnetv2_152 and the trained SVM model was then used to predict their labels. On this dataset, we achieve a classification accuracy of 96%, with just four images are wrongly predicted as clean solar panels. All the wrong predictions and some of the correct predictions are shown in Figure 12.

6. Conclusions and Future Work

In the face of backbone architecture selection problem for deep learning-powered image-based dust detection on solar panels, we perform a comprehensive and systematic evaluation of 100 pre-trained CNN models, drawn from 16 diverse architecture families. Using a standardized pipeline comprising feature extraction via pre-trained models, vectorizing the output features map via global average pooling, and classification using a linear SVM, we assessed the models in terms of both classification accuracy and training time-based efficiency. The evaluation is conducted under three stratified data splits, 70–30, 80–20, and 90–10, to ensure robustness and generalizability across varying amounts of training data.
Our findings reveal key insights that are essential for selecting the best backbone model for deep learning-based dust detection systems for solar panels. Certain model families, such as ResNetV2, ConvNext, and RegNet stood out in terms of achieving high classification accuracy, making them highly suitable for applications where detection precision is crucial. Conversely, families like Inception_Next, CS3DarkNet, and MobileNetV3 exhibited superior training efficiency, requiring significantly less time to train the linear SVM on the encoded features. These results are particularly valuable for edge-computing scenarios or real-time applications where speed and resource constraints are critical.
Overall, our aim is to bridge a crucial gap in the literature by performing a large-scale, comparative benchmarking of CNN models to be used as backbones/feature extractors for image-based solar panel dust detection. By highlighting both the strengths and trade-offs of various architectures, we provide practical guidance to researchers and practitioners in selecting the most appropriate model based on the operational needs—whether that be maximizing accuracy, minimizing computation, or attaining a balance between the two.
In future work, we plan to extend both the datasets and the current pipeline for a multi-class classification problem to identify each of the dirt types using a novel CNN architecture based on resnetv2_152 as it achieved the best results both with respect to classification accuracy and efficiency. The future work may also explore the integration of lightweight transformers, ensemble strategies, or model pruning techniques to further improve real-time applicability while maintaining high accuracy.
Beyond the specific case of solar panels, the methodology presented here offers a versatile framework for monitoring various types of surface conditions. For instance, the proposed methodology could be adapted to target the cleaning of building facades, where identifying specific areas of pollution or staining can optimize maintenance resources. By providing a transparent image-based classification alongside qualitative analysis, our proposed pipeline ensures that such automated inspection systems remain reliable in diverse urban environments.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The image dataset used in this paper is publicly available and is mentioned in a dedicated section along with reference.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Data-Split 70–30 Results.
Table A1. Data-Split 70–30 Results.
ModelAccuracyPrecisionRecallF1 ScoreTime (s)
convnext_xxlarge91.4391.5191.4391.431.38
resnetv2_15291.3791.4491.3791.382.16
regnet_y_32gf89.8489.9589.8489.862.30
convnext_xlarge89.7589.8689.7589.770.81
densenet16189.5389.5789.5389.522.32
densenet20187.9887.9987.9887.951.93
densenet16987.8887.9387.8887.881.64
dpn13187.1387.2187.1387.121.42
mnasnet1_087.1087.1287.1087.061.26
convnextv2_tiny86.7086.8486.7086.720.31
convnext_base86.6086.6586.6086.590.80
inception_next_base86.3986.5286.3986.410.45
dpn9885.8985.9285.8985.881.79
convnext_small85.8385.9285.8385.840.60
inception_next_small85.8385.9285.8385.820.32
dpn10785.5585.6585.5585.551.29
convnextv2_nano85.5285.6685.5285.520.24
inception_next_tiny85.5185.6685.5185.500.35
densenet12185.4285.5285.4285.430.93
mobilenet_v3_large85.3985.4885.3985.380.88
resnet15285.2785.4885.2785.261.20
efficientnet_b085.2685.4185.2685.271.27
convnext_tiny84.9885.0784.9885.010.66
convnext_large84.9284.9784.9284.911.33
efficientnet_b284.9285.0084.9284.941.48
shufflenet_v2_x2_084.7784.8184.7784.751.69
regnet_x_1_6gf84.4284.5284.4284.390.90
regnet_x_800mf84.3384.3884.3384.320.64
cs3darknet_x84.2484.4384.2484.250.50
efficientnet_b184.1884.2884.1884.201.38
mnasnet0_584.0584.0484.0583.981.69
mnasnet0_7584.0584.1884.0584.061.58
mobilenet_v3_small84.0584.2484.0584.090.58
regnet_x_3_2gf84.0284.0384.0284.000.92
shufflenet_v2_x0_583.9984.0783.9983.950.47
shufflenet_v2_x1_083.9984.0583.9983.970.44
efficientnet_b383.9384.1683.9383.931.55
efficientnet_v2_s83.9383.9983.9383.911.27
dpn6883.6883.6983.6883.650.33
dpn9283.6883.9283.6883.682.20
mnasnet1_383.6183.6583.6183.581.63
mobilenetv4_conv_small83.5283.5683.5283.510.52
vgg16_bn83.5283.5883.5283.473.51
vgg1383.4983.4783.4983.425.41
efficientnet_b483.1883.3483.1883.211.88
alexnet83.1283.1383.1283.028.66
vgg11_bn82.9982.9782.9982.923.77
dpn68b82.882.9682.882.820.29
regnet_x_400mf82.4982.5282.4982.440.53
shufflenet_v2_x1_582.4682.582.4682.410.58
dpn68b82.2482.3982.2482.280.3
convnext_nano82.2182.3982.2182.260.35
vgg1182.1582.1982.1582.095.42
mobilenet_v281.8781.8781.8781.851.28
vgg13_bn81.8481.8681.8481.724.08
cs3darknet_focus_m81.7181.6681.7181.610.26
resnet1881.6581.6881.6581.590.23
resnet3481.5681.8181.5681.570.25
efficientnet_b581.581.6581.581.522.73
vgg1681.4381.3581.4381.265.21
efficientnet_b781.3481.3781.3481.333
cs3darknet_focus_l81.2881.3781.2881.280.42
densenetblur121d81.0981.2681.0981.120.41
convnext_pico80.8781.0880.8780.930.21
regnet_y_16gf80.5680.6480.5680.495.9
regnet_y_400mf80.0680.1480.0680.070.27
regnet_y_8gf80.0680.1580.0680.011.62
resnetv2_5080.0680.2280.0680.071.33
convnextv2_pico79.4179.579.4179.360.31
vgg19_bn79.3879.4779.3879.384.03
efficientnet_b679.0679.179.0679.062.82
regnet_y_800mf78.7578.8278.7578.740.39
resnet10178.7278.9278.7278.751.52
inception_v378.678.8678.678.663.2
convnextv2_huge78.4778.5378.4778.42.82
regnet_y_1_6gf78.3878.3778.3878.31.07
resnetv2_10178.3278.578.3278.351.19
vgg1978.2678.1178.2678.115.42
efficientnet_v2_m77.9778.1477.97781.76
googlenet77.8577.9377.8577.841.24
resnet5077.677.6677.677.61.65
regnet_y_3_2gf77.0177.0277.0176.971.03
wide_resnet50_277.0177.1677.0177.041.68
wide_resnet101_276.6776.7676.6776.631.77
convnextv2_base76.676.7776.676.60.41
xception6576.5476.5576.5476.51.19
resnext101_32x8d76.1476.1376.1476.081.79
regnet_x_32gf75.987675.9875.914.55
convnextv2_large75.9575.8175.9575.840.88
regnet_x_16gf75.4875.4675.4875.453.18
resnext50_32x4d75.3375.4575.3375.331.6
convnextv2_femto74.1574.3874.1574.210.43
convnextv2_atto74.0574.1174.0573.970.75
inception_v473.8374.1273.8373.881.12
regnet_x_8gf73.5873.5573.5873.532.75
mobilenetv4_conv_medium73.1873.9273.1873.350.75
efficientnet_v2_l69.7569.8469.7569.661.99
xception7169.5969.569.5969.391.56
xception4167.4867.4867.4867.361.62
mobilenetv4_conv_large66.1466.0666.1466.041.2
Table A2. Data-Split 80–20 Results.
Table A2. Data-Split 80–20 Results.
ModelAccuracyPrecisionRecallF1 ScoreTime (s)
convnext_xxlarge91.9692.0491.9691.981.84
resnetv2_15291.9692.0891.9691.982.68
regnet_y_32gf89.8689.9289.8689.863.85
convnext_xlarge89.6789.7789.6789.691.08
densenet16189.289.3789.289.232.94
densenet16987.9487.9587.9487.922.06
densenet20187.9488.0187.9487.962.42
dpn13187.6287.8787.6287.661.97
convnext_base86.9687.0486.9686.981.03
convnextv2_tiny86.9687.3886.9687.040.39
convnextv2_nano86.9287.186.9286.960.32
mnasnet1_086.8786.9286.8786.861.59
convnext_small86.8286.986.8286.830.73
inception_next_base86.6486.8386.6486.680.54
mobilenet_v3_large86.2686.3586.2686.281.09
dpn10786.1286.2586.1286.141.75
resnet15286.1286.2886.1286.161.59
dpn9886.0786.2386.0786.12.37
inception_next_small86.0786.2586.0786.090.42
regnet_x_3_2gf85.9386.0285.9385.921.13
densenet12185.5685.7285.5685.591.11
efficientnet_b385.5685.7285.5685.531.91
efficientnet_b285.4785.6485.4785.511.8
convnext_tiny85.2385.4885.2385.310.83
mnasnet0_585.1985.2585.1985.181.96
mnasnet0_758585.158585.021.71
efficientnet_b084.9185.0484.9184.881.56
inception_next_tiny84.9185.0184.9184.940.46
convnext_large84.8185.1284.8184.891.68
dpn9284.6384.6884.6384.612.81
cs3darknet_x84.4984.7784.4984.560.71
efficientnet_b484.3984.5984.3984.432.41
regnet_x_800mf84.2584.3584.2584.240.8
mnasnet1_384.1184.1684.1184.081.91
mobilenetv4_conv_small84.0284.0484.02840.76
regnet_x_1_6gf83.9284.1583.9283.971.05
shufflenet_v2_x2_083.9284.0283.9283.942.24
efficientnet_b183.8383.9183.8383.851.61
mobilenet_v3_small83.8384.0983.8383.890.76
dpn6883.788483.7883.840.39
efficientnet_v2_s83.7883.9983.7883.851.57
vgg16_bn83.6483.6683.6483.614.07
vgg11_bn83.683.6983.683.594.58
convnext_nano83.5584.2283.5583.680.41
shufflenet_v2_x1_083.5583.6883.5583.570.56
alexnet83.3683.3283.3683.2810.84
vgg13_bn83.3683.4483.3683.325.28
dpn68b83.2283.5183.2283.290.37
mobilenet_v283.0883.1383.0883.041.6
regnet_x_400mf82.982.9382.982.880.67
shufflenet_v2_x0_582.8582.8982.8582.830.58
dpn68b82.882.9982.882.850.38
vgg1382.882.9282.882.826.61
shufflenet_v2_x1_582.6682.8882.6682.690.61
vgg1182.4882.5782.4882.446.6
cs3darknet_focus_l82.2482.582.2482.280.56
efficientnet_b582.2482.3282.2482.253.29
densenetblur121d82.282.4182.282.260.45
cs3darknet_focus_m82.1582.2682.1582.160.44
efficientnet_b781.9682.2681.9682.023.8
regnet_y_16gf81.8781.9581.8781.847.3
resnet1881.7381.9481.7381.750.31
regnet_y_8gf81.4581.5781.4581.472.45
convnext_pico81.3581.4981.3581.370.27
resnet3481.1781.5281.1781.270.3
resnetv2_5081.1281.4181.1281.21.77
vgg1681.0781.1281.0781.046.37
efficientnet_b680.8480.9180.8480.823.53
convnextv2_pico80.7580.9180.7580.790.4
regnet_y_800mf80.7580.9280.7580.720.45
inception_v380.1480.380.1480.154.03
resnetv2_10179.6779.8879.6779.721.66
vgg19_bn79.4479.6679.4479.485.09
regnet_y_400mf78.9779.278.9779.040.31
convnextv2_huge78.9378.8378.9378.813.38
convnextv2_base78.6578.7378.6578.640.58
resnet10178.678.7978.678.651.98
resnet5078.5578.7778.5578.612.29
regnet_y_3_2gf78.2778.6378.2778.361.36
googlenet77.9978.1377.9977.991.46
vgg1977.9977.8977.9977.886.53
convnextv2_large77.9478.0677.9477.91.13
efficientnet_v2_m77.978.1877.977.912.1
regnet_y_1_6gf77.3877.577.3877.371.22
wide_resnet50_276.8777.1376.8776.922.18
resnext101_32x8d76.5476.8176.5476.592.39
regnet_x_16gf76.1276.2176.1276.14.08
xception6575.8476.1375.8475.891.46
resnext50_32x4d75.776.0575.775.822.07
convnextv2_femto75.6575.9875.6575.730.7
convnextv2_atto75.5675.5375.5675.51.13
wide_resnet101_275.3375.5275.3375.322.39
mobilenetv4_conv_medium74.8675.2574.8674.991.02
regnet_x_32gf74.6374.4774.6374.515.8
inception_v474.2574.7574.2574.41.43
regnet_x_8gf73.3673.6673.3673.463.2
efficientnet_v2_l69.5869.4369.5869.452.53
xception7169.1669.1869.1669.132.13
mobilenetv4_conv_large68.5568.4468.5568.461.64
xception4168.1368.1168.1368.012.21
Table A3. Data-Split 90–10 Results.
Table A3. Data-Split 90–10 Results.
ModelAccuracyPrecisionRecallF1 ScoreTime (s)
convnext_xxlarge92.4392.4992.4392.432.11
resnetv2_15291.7891.9291.7891.83.21
regnet_y_32gf90.5690.6490.5690.575.79
convnext_xlarge90.2890.590.2890.31.3
densenet16189.8190.0289.8189.823.51
densenet16988.688.7388.688.592.54
densenet20188.3288.5788.3288.342.93
convnextv2_tiny87.6688.0487.6687.670.47
dpn13187.6687.7887.6687.652.19
resnet15287.4887.6587.4887.52.01
inception_next_tiny87.2987.5487.2987.310.58
convnextv2_nano87.287.3887.287.180.39
inception_next_small86.6486.6586.6486.590.54
convnext_base86.5486.686.5486.51.21
dpn9886.5486.7686.5486.543.11
inception_next_base86.5486.7986.5486.510.69
dpn10786.2686.3686.2686.222.35
mnasnet0_7586.2686.486.2686.171.97
convnext_large86.1786.386.1786.162.08
mnasnet1_086.1786.3286.1786.122.02
regnet_x_3_2gf86.1786.4186.1786.151.36
efficientnet_b286.0786.1886.0786.082.26
convnext_small85.8986.0285.8985.880.88
efficientnet_b085.8986.185.8985.891.91
efficientnet_b185.898685.8985.842.02
dpn6885.786.0685.785.740.61
alexnet85.6185.7385.6185.5512.25
regnet_x_800mf85.4285.685.4285.380.96
efficientnet_b385.3385.5285.3385.312.36
efficientnet_b485.2485.4285.2485.232.96
convnext_tiny85.2385.4785.2385.170.97
cs3darknet_x85.0585.1685.0585.060.87
efficientnet_v2_s85.0585.3485.0585.081.96
regnet_x_1_6gf84.9585.284.9584.961.29
mobilenet_v3_large84.7784.984.7784.71.32
vgg16_bn84.6784.7284.6784.624.59
shufflenet_v2_x1_084.5884.6684.5884.550.67
shufflenet_v2_x2_084.5884.7584.5884.512.66
mnasnet1_384.3984.3784.3984.352.11
dpn9284.384.4284.384.243.06
mobilenetv4_conv_small84.384.4884.384.31.01
densenet12184.1184.1284.1184.031.36
mobilenet_v3_small84.0284.2784.0284.040.92
mnasnet0_583.8383.8283.8383.732.06
efficientnet_b583.7483.9583.7483.753.83
efficientnet_b783.7483.9683.7483.734.61
shufflenet_v2_x0_583.5583.9483.5583.610.75
regnet_x_400mf83.1883.3783.1883.160.82
dpn68b82.9983.2682.9982.960.45
vgg11_bn82.9983.0682.9982.835.41
vgg13_bn82.9983.0482.9982.836.23
mobilenet_v282.982.9582.982.781.97
vgg1182.982.8782.982.777.64
regnet_y_16gf82.4382.5982.4382.358.9
shufflenet_v2_x1_582.3482.4682.3482.290.75
convnextv2_pico82.2482.4982.2482.260.53
densenetblur121d82.2482.5882.2482.210.56
resnet3482.2482.782.2482.250.37
resnetv2_5082.2482.8282.2482.252.17
regnet_y_8gf82.1582.2182.1582.083.34
resnet1882.0682.2982.0682.070.37
cs3darknet_focus_l81.788281.7881.720.56
convnext_pico81.6881.7681.6881.520.37
cs3darknet_focus_m81.5981.8781.5981.430.41
inception_v381.581.681.581.444.64
regnet_y_800mf81.481.9581.481.360.52
vgg1381.2181.381.2181.187.81
convnext_nano80.9381.3980.9381.040.49
dpn68b80.9381.1480.9380.960.44
vgg1680.7580.9880.7580.567.65
resnet5080.2880.5480.2880.282.92
regnet_y_400mf80.0980.3980.0980.130.42
convnextv2_huge79.6379.6979.6379.584.08
efficientnet_b679.5379.6979.5379.554.26
vgg19_bn79.3579.4679.3579.166.19
googlenet79.2579.4479.2579.241.8
resnetv2_10179.2579.6779.2579.222.03
efficientnet_v2_m78.678.678.678.492.36
convnextv2_base78.578.8878.578.470.73
convnextv2_large77.8577.9577.8577.651.53
regnet_y_3_2gf77.2977.3277.2977.251.61
vgg1977.277.1977.276.917.83
resnet10177.177.4777.177.112.38
wide_resnet101_277.0176.9377.0176.753.06
regnet_y_1_6gf76.7377.0676.7376.71.44
xception6576.6376.6276.6376.451.83
resnext101_32x8d76.4576.5876.4576.363.02
resnext50_32x4d76.4576.5976.4576.422.43
regnet_x_32gf76.3676.1876.3676.077.22
convnextv2_atto75.8975.8475.8975.771.67
wide_resnet50_275.875.975.875.752.68
regnet_x_16gf75.4275.3475.4275.345.01
convnextv2_femto75.1475.3575.1475.161.07
regnet_x_8gf74.9575.0774.9574.843.71
inception_v474.3974.4474.3974.371.87
mobilenetv4_conv_medium74.3974.5774.3974.391.38
mobilenetv4_conv_large69.0668.6569.0668.682.26
efficientnet_v2_l68.8869.0768.8868.623.18
xception7167.8567.7967.8567.72.86
xception4165.9865.5465.9865.552.89

References

  1. Toennies, K.D. An Introduction to Image Classification: From Designed Models to End-to-End Learning; Springer: Berlin/Heidelberg, Germany, 2024. [Google Scholar]
  2. Liu, G.H.; Yang, J.Y. Content-based image retrieval using color difference histogram. Pattern Recognit. 2013, 46, 188–198. [Google Scholar] [CrossRef]
  3. Castleman, K.R. Digital image processing; Prentice Hall Press: Upper Saddle River, NJ, USA, 1996. [Google Scholar]
  4. Lowe, D.G. Object recognition from local scale-invariant features. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–25 September 1999; Volume 2, pp. 1150–1157. [Google Scholar]
  5. Guo, Z.; Zhang, L.; Zhang, D. A completed modeling of local binary pattern operator for texture classification. IEEE Trans. Image Process. 2010, 19, 1657–1663. [Google Scholar] [CrossRef]
  6. Anwar, A.; Anwar, H.; Anwar, S. Towards Low-Cost Classification for Novel Fine-Grained Datasets. Electronics 2022, 11, 2701. [Google Scholar] [CrossRef]
  7. Torralba, A.; Isola, P.; Freeman, W.T. Foundations of Computer Vision; MIT Press: Cambridge, MA, USA, 2024. [Google Scholar]
  8. Anwar, H.; Anwar, S.; Zambanini, S.; Porikli, F. Deep ancient Roman Republican coin classification via feature fusion and attention. Pattern Recognit. 2021, 114, 107871. [Google Scholar] [CrossRef]
  9. Imran, M.; Anwar, H.; Tufail, M.; Khan, A.; Khan, M.; Ramli, D.A. Image-Based Automatic Energy Meter Reading Using Deep Learning. Comput. Mater. Contin. 2023, 75, 203–216. [Google Scholar] [CrossRef]
  10. Yi, J.; Zhang, G.; Yu, H.; Yan, H. Advantages, challenges and molecular design of different material types used in organic solar cells. Nat. Rev. Mater. 2024, 9, 46–62. [Google Scholar] [CrossRef]
  11. Pourasl, H.H.; Barenji, R.V.; Khojastehnezhad, V.M. Solar energy status in the world: A comprehensive review. Energy Rep. 2023, 10, 3474–3493. [Google Scholar] [CrossRef]
  12. Ding, R.; Cao, Z.; Teng, J.; Cao, Y.; Qian, X.; Yue, W.; Yuan, X.; Deng, K.; Wu, Z.; Li, S.; et al. Self-Powered Autonomous Electrostatic Dust Removal for Solar Panels by an Electret Generator. Adv. Sci. 2024, 11, 2401689. [Google Scholar] [CrossRef] [PubMed]
  13. Mondal, A.K.; Bansal, K. A brief history and future aspects in automatic cleaning systems for solar photovoltaic panels. Adv. Robot. 2015, 29, 515–524. [Google Scholar] [CrossRef]
  14. Alatwi, A.M.; Albalawi, H.; Wadood, A.; Anwar, H.; El-Hageen, H.M. Deep Learning-Based Dust Detection on Solar Panels: A Low-Cost Sustainable Solution for Increased Solar Power Generation. Sustainability 2024, 16, 8664. [Google Scholar] [CrossRef]
  15. Rathod, A.P.S.; Singh, R.K.; Kumar, S. Effect of Dust Accumulation on Efficiency of Solar Panels in Clement Town Region (Dehradun) India: An Empirical Study. NanoWorld J. 2023, 9, S1–S6. [Google Scholar] [CrossRef]
  16. Shariah, A.; Al-Ibrahim, E. Impact of Dust and Shade on Solar Panel Efficiency and Development of a Simple Method for Measuring the Impact of Dust in any Location. J. Sustain. Dev. Energy Water Environ. Syst. 2023, 11, 1110448. [Google Scholar]
  17. Rashid, M.; Yousif, M.; Rashid, Z.; Muhammad, A.; Altaf, M.; Mustafa, A. Effect of dust accumulation on the performance of photovoltaic modules for different climate regions. Heliyon 2023, 9, e23069. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, J.; Hu, W.; Wen, Y.; Zhang, F.; Li, X. Dust deposition characteristics on photovoltaic arrays investigated through wind tunnel experiments. Sci. Rep. 2025, 15, 1582. [Google Scholar] [CrossRef]
  19. Yakubu, S.; Samikannu, R.; Gawusu, S.; Wetajega, S.D.; Okai, V.; Shaibu, A.K.S.; Workneh, G.A. A holistic review of the effects of dust buildup on solar photovoltaic panel efficiency. Sol. Compass 2025, 13, 100101. [Google Scholar] [CrossRef]
  20. Kawamoto, H. Electrodynamic dust removal technologies for solar panels: A comprehensive review. J. Electrost. 2025, 134, 104045. [Google Scholar] [CrossRef]
  21. Dantas, G.M.; Mendes, O.L.C.; Maia, S.M.; de Alexandria, A.R. Dust detection in solar panel using image processing techniques: A review. Res. Soc. Dev. 2020, 9, e321985107. [Google Scholar] [CrossRef]
  22. Abdelsattar, M.; Rasslan, A.A.A.; Emad-Eldeen, A. Detecting dusty and clean photovoltaic surfaces using MobileNet variants for image classification. SVU-Int. J. Eng. Sci. Appl. 2025, 6, 9–18. [Google Scholar]
  23. Onim, M.S.H.; Sakif, Z.M.M.; Ahnaf, A.; Kabir, A.; Azad, A.K.; Oo, A.M.T.; Afreen, R.; Hridy, S.T.; Hossain, M.; Jabid, T.; et al. SolNet: A convolutional neural network for detecting dust on solar panels. Energies 2022, 16, 155. [Google Scholar] [CrossRef]
  24. Naeem, U.; Chadda, K.; Vahaji, S.; Ahmad, J.; Li, X.; Asadi, E. Aerial Imaging-Based Soiling Detection System for Solar Photovoltaic Panel Cleanliness Inspection. Sensors 2025, 25, 738. [Google Scholar] [CrossRef]
  25. Cruz-Rojas, T.; Franco, J.A.; Hernandez-Escobedo, Q.; Ruiz-Robles, D.; Juarez-Lopez, J.M. A novel comparison of image semantic segmentation techniques for detecting dust in photovoltaic panels using machine learning and deep learning. Renew. Energy 2023, 217, 119126. [Google Scholar] [CrossRef]
  26. Gao, X.; Wang, T.; Liu, M.; Lian, J.; Yao, Y.; Yu, L.; Li, Y.; Cui, Y.; Xue, R. A framework to identify guano on photovoltaic modules in offshore floating photovoltaic power plants. Sol. Energy 2024, 274, 112598. [Google Scholar] [CrossRef]
  27. Oulefki, A.; Trongtirakul, T.; Agaian, S.; Benbelkacem, S.; Zenati, N. Multi-view VR imaging for enhanced analysis of dust accumulation on solar panels. Sol. Energy 2024, 279, 112708. [Google Scholar] [CrossRef]
  28. Bassil, J.; Noura, H.; Salman, O.; Chahine, K.; Guizani, M. Deep learning image classification models for solar panels dust detection. In Proceedings of the 2024 International Wireless Communications and Mobile Computing (IWCMC), Ayia Napa, Cyprus, 3–7 June 2024; pp. 1516–1521. [Google Scholar]
  29. Mohammed, H.M.; Alawi, A.E.B. CASolarNet: Channel Attention EfficientNet-based Model for Solar Panel Dust Detection. In Proceedings of the 2024 4th International Conference on Emerging Smart Technologies and Applications (eSmarTA), Sana’a, Yemen, 28–30 October 2024; pp. 1–4. [Google Scholar]
  30. Sefer, T.; Kaya, M. Detection of Dust on Solar Panels with Deep Learning. Kahramanmaraş Sütçü İmam Üniversitesi Mühendislik Bilim. Derg. 2024, 27, 1451–1464. [Google Scholar] [CrossRef]
  31. Cipriani, G.; D’Amico, A.; Guarino, S.; Manno, D.; Traverso, M.; Di Dio, V. Convolutional neural network for dust and hotspot classification in PV modules. Energies 2020, 13, 6357. [Google Scholar] [CrossRef]
  32. Chen, L.; Fan, S.; Sun, S.; Cao, S.; Sun, T.; Liu, P.; Gao, H.; Zhang, Y.; Ding, W. A detection model for dust deposition on photovoltaic (PV) panels based on light transmittance estimation. Energy 2025, 322, 135284. [Google Scholar] [CrossRef]
  33. Shao, Y.; Zhang, C.; Xing, L.; Sun, H.; Zhao, Q.; Zhang, L. A new dust detection method for photovoltaic panel surface based on Pytorch and its economic benefit analysis. Energy AI 2024, 16, 100349. [Google Scholar] [CrossRef]
  34. Noura, H.N.; Chahine, K.; Bassil, J.; Abou Chaaya, J.; Salman, O. Efficient combination of deep learning models for solar panel damage and soiling detection. Measurement 2025, 251, 117185. [Google Scholar] [CrossRef]
  35. Ahmed, S.; Rashid, H.; Qadir, Z.; Tayyab, Q.; Senjyu, T.; Elkholy, M. Deep Learning-Based Recognition and Classification of Soiled Photovoltaic Modules Using HALCON Software for Solar Cleaning Robots. Sensors 2025, 25, 1295. [Google Scholar] [CrossRef] [PubMed]
  36. Prova, N.N.I. Improved Solar Panel Efficiency through Dust Detection Using the InceptionV3 Transfer Learning Model. In Proceedings of the 2024 8th International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud)(I-SMAC), Dharan, Nepal, 5–7 September 2024; pp. 260–268. [Google Scholar]
  37. Tan, Y.; Liao, K.; Bai, X.; Deng, C.; Zhao, Z.; Zhao, B. Denoising Convolutional Neural Networks Based Dust Accumulation Status Evaluation of Photovoltaic Panel. In Proceedings of the 2019 IEEE International Conference on Energy Internet (ICEI), Nanjing, China, 27–31 May 2019; pp. 560–566. [Google Scholar]
  38. Zhang, W.; Archana, V.; Gandhi, O.; Rodríguez-Gallegos, C.D.; Quan, H.; Yang, D.; Tan, C.W.; Chung, C.; Srinivasan, D. SoilingEdge: PV soiling power loss estimation at the edge using surveillance cameras. IEEE Trans. Sustain. Energy 2023, 15, 556–566. [Google Scholar] [CrossRef]
  39. Sai, H. Solar Panel Dust Detection. 2023. Available online: https://www.kaggle.com/datasets/hemanthsai7/solar-panel-dust-detection (accessed on 13 June 2024).
  40. Zambanini, S.; Kampel, M. Coarse-to-fine correspondence search for classifying ancient coins. In Proceedings of the ACCV, Daejeon, Republic of Korea, 5–9 November 2012; pp. 25–36. [Google Scholar]
  41. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  42. Healy, J.; McInnes, L. Uniform manifold approximation and projection. Nat. Rev. Methods Prim. 2024, 4, 82. [Google Scholar] [CrossRef]
  43. Buckley, S.J.; Ringdal, K.; Naumann, N.; Dolva, B.; Kurz, T.H.; Howell, J.A.; Dewez, T.J. LIME: Software for 3-D visualization, interpretation, and communication of virtual geoscience models. Geosphere 2019, 15, 222–235. [Google Scholar] [CrossRef]
  44. Adadi, A.; Berrada, M. Peeking inside the black-box: A survey on explainable artificial intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  45. Petsiuk, V.; Das, A.; Saenko, K. RISE: Randomized Input Sampling for Explanation of Black-box Models. In Proceedings of the British Machine Vision Conference (BMVC), Newcastle, UK, 3–6 September 2018. [Google Scholar]
Figure 1. Global renewable-based power generation (a) The share of renewable energy capacity with respect to various technologies region-wise (b) The annual addition of new renewable energy-based power generation installations from 2010 to 2023 where a sharp increase can be noticed for solar PV from 2019 until 2023 (c) The yearly global percentage price drop of solar panels belonging to various types in the last five years https://www.irena.org (accessed on 29 November 2025).
Figure 1. Global renewable-based power generation (a) The share of renewable energy capacity with respect to various technologies region-wise (b) The annual addition of new renewable energy-based power generation installations from 2010 to 2023 where a sharp increase can be noticed for solar PV from 2019 until 2023 (c) The yearly global percentage price drop of solar panels belonging to various types in the last five years https://www.irena.org (accessed on 29 November 2025).
Information 17 00052 g001
Figure 2. The challenges in the image dataset. The first row shows an interesting challenge where the cleaning process is depicted such that some of the solar panels are cleaned while others are dirty. These images are placed into the dirty class. The second row shows the variations found in the dirt itself. For instance, dust, bird droppings and fungus, etc., are the commonly found dirt on solar panels. The third row yet again shows an interesting aspect of the clean solar panels as they are imaged with clear sky and a green environment that induce a strong bias in the images. If such an environment is also found in dirty solar panels, this may lead to misclassifications. The final row shows the variations in panels orientations. Overall, despite being fewer in number, the images in this dataset are very challenging due to these factors. Both the dataset and code can be accessed on https://github.com/hafeez-anwar/SolClean-Code (accessed on 2 January 2026).
Figure 2. The challenges in the image dataset. The first row shows an interesting challenge where the cleaning process is depicted such that some of the solar panels are cleaned while others are dirty. These images are placed into the dirty class. The second row shows the variations found in the dirt itself. For instance, dust, bird droppings and fungus, etc., are the commonly found dirt on solar panels. The third row yet again shows an interesting aspect of the clean solar panels as they are imaged with clear sky and a green environment that induce a strong bias in the images. If such an environment is also found in dirty solar panels, this may lead to misclassifications. The final row shows the variations in panels orientations. Overall, despite being fewer in number, the images in this dataset are very challenging due to these factors. Both the dataset and code can be accessed on https://github.com/hafeez-anwar/SolClean-Code (accessed on 2 January 2026).
Information 17 00052 g002
Figure 3. The complete methodology. The solar panel images belong to two classes namely clean and dirty. These images are then encoded using a pre-trained CNN model that generates a dense multi-dimensional feature map. These feature maps are then converted to 1D vectors by applying GlobalAveragePooling. Once encoded, the images are then used to train and test a linear SVM using a StratifiedShuffleSplit strategy that takes the class distribution into account thus accounting for class imbalance. The evaluate splitting percentages are 70–30, 80–20, and 90–10, where for each of the split percentage, the experiments are run 10 times due to the random split, and the average results are reported.
Figure 3. The complete methodology. The solar panel images belong to two classes namely clean and dirty. These images are then encoded using a pre-trained CNN model that generates a dense multi-dimensional feature map. These feature maps are then converted to 1D vectors by applying GlobalAveragePooling. Once encoded, the images are then used to train and test a linear SVM using a StratifiedShuffleSplit strategy that takes the class distribution into account thus accounting for class imbalance. The evaluate splitting percentages are 70–30, 80–20, and 90–10, where for each of the split percentage, the experiments are run 10 times due to the random split, and the average results are reported.
Information 17 00052 g003
Figure 4. Specifications of all the evaluated CNN models such as the encoded vector size, floating point operations (FLOPs), number of parameters, and the time in seconds they take to process a single image also called the inference time.
Figure 4. Specifications of all the evaluated CNN models such as the encoded vector size, floating point operations (FLOPs), number of parameters, and the time in seconds they take to process a single image also called the inference time.
Information 17 00052 g004
Figure 5. Classification accuracies achieved across all the three splitting strategies by the best variant of the family.
Figure 5. Classification accuracies achieved across all the three splitting strategies by the best variant of the family.
Information 17 00052 g005
Figure 6. Training Time taken across all the three splitting strategies by the best variant of the family.
Figure 6. Training Time taken across all the three splitting strategies by the best variant of the family.
Information 17 00052 g006
Figure 7. Detailed comparison of the bests.
Figure 7. Detailed comparison of the bests.
Information 17 00052 g007
Figure 8. Comparative performance analysis of the top two pre-trained models—resnetv2_152 and convnext_xxlarge—on image-based dust detection in solar panels using linear SVM classification in terms of results stability, PR, and ROC curves.
Figure 8. Comparative performance analysis of the top two pre-trained models—resnetv2_152 and convnext_xxlarge—on image-based dust detection in solar panels using linear SVM classification in terms of results stability, PR, and ROC curves.
Information 17 00052 g008
Figure 9. (ad) LIME visualization on images that are encoded with convnext_xxlarge.
Figure 9. (ad) LIME visualization on images that are encoded with convnext_xxlarge.
Information 17 00052 g009
Figure 10. (ad) RISE visualization on images that are encoded with resnetv2_152l.
Figure 10. (ad) RISE visualization on images that are encoded with resnetv2_152l.
Information 17 00052 g010
Figure 11. (ag) Images of dirty solar panels wrongly classified as clean.
Figure 11. (ag) Images of dirty solar panels wrongly classified as clean.
Information 17 00052 g011
Figure 12. Correctly and wrongly classified dirty solar panels in the unseen test dataset.
Figure 12. Correctly and wrongly classified dirty solar panels in the unseen test dataset.
Information 17 00052 g012
Table 1. Summary of Research Papers on Dust Detection on Solar Panels Using Deep Learning.
Table 1. Summary of Research Papers on Dust Detection on Solar Panels Using Deep Learning.
S.No.YearRef.CNN Models Evaluated
12025[34]ViTs and EfficientNet
22025[35]ANN
32025[22]MobileNetV1, V2, V3
42024[14]DenseNet169, VGG16, ResNet50, and 17 others
52024[36]InceptionV3
62024[28]ResNet50, VGG16, InceptionV3
72024[29]EfficientNet (with Channel Attention)
82024[33]Custom CNN Model
92023[23]SolNet, VGG16, ResNet50, InceptionV3, MobileNetV2
102023[37]DnCNN, VGG16, AlexNet, ResNet
112023[38]ResNet50, MobileNet
122023[25]UNet
Table 2. Number of dirty solar panel images categorized by the type of contaminant.
Table 2. Number of dirty solar panel images categorized by the type of contaminant.
Type of DirtNumber of Images
Bird droppings69
Debris7
Defect9
Dust524
Leaves4
Occlusions (human or cleaning tools)45
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Anwar, H. TransferLearning-Driven Large-Scale CNN Benchmarking with Explainable AI for Image-Based Dust Detection on Solar Panels. Information 2026, 17, 52. https://doi.org/10.3390/info17010052

AMA Style

Anwar H. TransferLearning-Driven Large-Scale CNN Benchmarking with Explainable AI for Image-Based Dust Detection on Solar Panels. Information. 2026; 17(1):52. https://doi.org/10.3390/info17010052

Chicago/Turabian Style

Anwar, Hafeez. 2026. "TransferLearning-Driven Large-Scale CNN Benchmarking with Explainable AI for Image-Based Dust Detection on Solar Panels" Information 17, no. 1: 52. https://doi.org/10.3390/info17010052

APA Style

Anwar, H. (2026). TransferLearning-Driven Large-Scale CNN Benchmarking with Explainable AI for Image-Based Dust Detection on Solar Panels. Information, 17(1), 52. https://doi.org/10.3390/info17010052

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop