Next Article in Journal
Transcriptome Sequencing and Differential Analysis of Testes of 1-Year-Old and 3-Year-Old Kazakh Horses
Previous Article in Journal
The Multifaceted Importance of Amphibians: Ecological, Biomedical, and Socio-Economic Perspectives
Previous Article in Special Issue
Transformer-Based Classification of Transposable Element Consensus Sequences with TEclass2
error_outline You can access the new MDPI.com website here. Explore and share your feedback with us.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Explainable Deep Learning Framework for Reliable Species-Level Classification Within the Genera Desmodesmus and Tetradesmus

by
İlknur Meriç Turgut
1,*,
Dilara Gerdan Koc
2 and
Özden Fakıoğlu
3
1
Department of Fisheries and Aquaculture Engineering, Faculty of Agriculture, Ankara University, 06110 Ankara, Türkiye
2
Department of Agricultural Machinery and Technologies Engineering, Faculty of Agriculture, Ankara University, 06110 Ankara, Türkiye
3
Department of Basic Science, Faculty of Fisheries, Ataturk University, 25240 Erzurum, Türkiye
*
Author to whom correspondence should be addressed.
Biology 2026, 15(1), 99; https://doi.org/10.3390/biology15010099
Submission received: 26 November 2025 / Revised: 25 December 2025 / Accepted: 1 January 2026 / Published: 3 January 2026
(This article belongs to the Special Issue AI Deep Learning Approach to Study Biological Questions (2nd Edition))

Simple Summary

The accurate identification of microalgae remains challenging due to imaging variability, physiological changes, and environmental factors rather than inherent morphological similarity. This study presents an interpretable artificial intelligence framework designed to classify green algae species from microscope images with both high accuracy and transparency. Twelve deep learning models were systematically trained and compared using standardized image processing on three distinct species—Desmodesmus flavescens, Desmodesmus subspicatus and Tetradesmus dimorphus. Among the evaluated architectures, several convolutional neural networks achieved strong performance, with macro-level precision, recall, and F1-scores generally exceeding 0.90, and the best-performing model (ResNet152V2) reaching a macro F1-score of 0.975. Visualization analyses confirmed that the models based their decisions on true biological features, such as cell walls and surface structures, rather than irrelevant background elements. These findings demonstrate that artificial intelligence can achieve near-perfect recognition of microalgae even from limited datasets while maintaining interpretability. The proposed framework provides a reproducible and biologically meaningful tool for digital taxonomy.

Abstract

Microalgae are an evolutionarily ancient and morphologically diverse group of photosynthetic eukaryotes, with taxonomic resolution complicated by environmentally driven phenotypic plasticity. This study merges deep learning and explainable artificial intelligence (XAI) to establish a transparent, reliable, and biologically meaningful framework for green microalgae (Chlorophyta) classification. Microscope images from three morphologically distinct algal species—Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus representing the genera Desmodesmus and Tetradesmus within Chlorophyta—were analyzed using twelve convolutional neural networks, including EfficientNet-B0–B7, DenseNet201, NASNetLarge, Xception, and ResNet152V2. A curated dataset comprising 3624 microscopic images from three Chlorophyta species was used, split into training, validation, and test subsets. All models were trained using standardized preprocessing and data augmentation procedures, including grayscale conversion, CLAHE-based contrast enhancement, rotation, flipping, and brightness normalization. The model’s performance was assessed using accuracy and loss metrics on independent test datasets, while interpretability was evaluated through saliency maps and Gradient-weighted Class Activation Mapping (Grad-CAM) visualizations. ResNet152V2 achieved the highest overall performance among all evaluated architectures, outperforming EfficientNet variants, NASNetLarge, and Xception in terms of macro F1-score. Visualization analysis showed that both Grad-CAM and saliency mapping consistently highlighted biologically relevant regions—including cell walls, surface ornamentation, and colony structures—confirming that the models relied on taxonomically meaningful features rather than background artifacts. The findings indicate that the integration of deep learning and XAI can attain consistently high test accuracy for microalgal species, even with constrained datasets. This approach enables automated taxonomy and supports biodiversity monitoring, ecological assessment, biomass optimization, and biodiesel production by integrating interpretability with high predictive accuracy.

Graphical Abstract

1. Introduction

Spanning multiple evolutionary lineages, algae encompass a spectrum of cellular architectures and adaptive traits that complicate accurate and reproducible taxonomic resolution. Acting as primary producers, they sustain food webs, regulate nutrient fluxes, and contribute substantially to atmospheric oxygen generation. Their ecological and economic significance—ranging from environmental monitoring and aquaculture to biotechnology and renewable energy—further heightens the need for precise and reliable species identification. In this context, the inherent structural complexity of algal taxa underscores the demand for analytical approaches capable of capturing subtle morphological variation and distinguishing closely related groups with greater consistency. Classical approaches based on morphometric traits are limited by subjective interpretation, time-intensive procedures, and morphological plasticity influenced by environmental variables such as light, nutrient availability, and physiological stress [1]. This phenotypic variability frequently results in misclassification, especially among closely related microalgae whose diagnostic characters can partially overlap. Although molecular approaches have greatly advanced phytoplankton systematics, they have also revealed a pronounced incongruence between morphology and phylogeny in unicellular green microalgae. As highlighted by Krienitz and Bock [2], multiple phylogenetic species may share identical morphotypes, while morphologically variable taxa often comprise genetically distinct lineages, resulting in widespread polyphyly among traditional genera. This challenge is particularly evident within Scenedesmaceae, where extensive morphological plasticity in genera such as Scenedesmus complicates reliable species-level identification under light microscopy. Consequently, morphology-based taxonomy remains both indispensable and inherently problematic, especially in routine ecological monitoring.
In recent years, the advent of artificial intelligence and, more specifically, deep learning, has revolutionized biological image analysis, enabling automated, scalable, and highly accurate recognition of complex morphological structures. Deep learning models, particularly convolutional neural networks (CNNs) and transformer-based architectures, have demonstrated exceptional performance in taxonomic classification tasks, achieving accuracies frequently exceeding 90% and occasionally surpassing 99%. Foundational CNNs such as AlexNet and VGG have been instrumental in pioneering this field, as evidenced by Pedraza et al. [3], who achieved 99% accuracy in the classification of eighty diatoms. Successive developments, including deep residual networks such as ResNet [3,4,5,6,7], introduced hierarchical feature extraction and improved gradient propagation, yielding greater stability and interpretability. EfficientNet architectures, designed for compound scaling, have achieved remarkable precision even with limited training data [8,9], while compact models such as MobileNet [10,11,12] have proven suitable for real-time and low-resource applications. More recent hybrid and attention-based frameworks—including MaxViT [13], the Swin Transformer [14], and the deformable detection transformer [15]—exemplify a new generation of architectures capable of integrating spatial and contextual information for fine-grained algal discrimination. Ensemble and hybrid approaches that combine CNNs with auxiliary classifiers, such as support vector machines or principal component analysis [4,16,17,18], have further enhanced the discriminative capacity of models applied to taxonomically distinct microalgal groups.
The implementation of these architectures has expanded across a broad range of ecological and biotechnological domains. Deep learning-based frameworks have been deployed for the identification of freshwater and marine microalgae, including Cyanobacteria, Chlorophyta, Ochrophyta, and Pyrrophyta [19,20,21,22], as well as for macroalgal monitoring using unmanned-aerial-vehicle (UAV) imagery [23]. In water-quality management, the automated recognition of toxic algal bloom-forming genera such as Microcystis, Anabaena, and Cylindrospermopsis has been achieved through CNN-and YOLO-based detection systems with impressive accuracy [23,24,25,26,27]. In biotechnological contexts, AI-assisted models have facilitated microalgal biomass estimation, culture optimization, and pigment quantification [11,28,29]. Reported accuracies in these studies range from 60% in highly variable field conditions to nearly 100% under controlled imaging settings, with F1-scores frequently exceeding 0.9. Gaur et al. [30] reported perfect classification performance for fifteen harmful algal genera using a ResNeXt-50 model, while Salmi et al. [31] achieved F1-scores approaching unity when employing hyperspectral CNNs to differentiate Microcystis and Synechococcus. Li et al. [4] achieved 97% accuracy for eight algal and one cyanobacterial species using Mueller-matrix imaging coupled with a CNN–SVM hybrid, and Abdullah et al. [19] demonstrated 90% accuracy for four desmid species using YOLOv5 on optical micrographs. Studies employing holographic or dark-field microscopy [29,32] have likewise expanded AI’s applicability to complex aquatic samples, while field-scale monitoring has been enhanced through UAV-based topographical imagery [23]. Complementary reviews and benchmarking efforts [3,16,33,34] further underscore the consistency of deep learning’s performance across imaging modalities.
Species-specific investigations reinforce the consistently high precision achieved by contemporary deep learning models while illuminating the biological and ecological attributes that underpin algorithmic success. The three microalgal species examined here—Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus—are members of the family Scenedesmaceae (Chlorophyta), a lineage renowned for its structural diversity, ecological adaptability, and diagnostic coenobial organization. Foundational phylogenetic analyses by An et al. [35] resolved the evolutionary relationships among Scenedesmus-like coccoid taxa and delineated the morphological boundaries separating Desmodesmus and Tetradesmus, thereby providing a robust taxonomic framework for species-level discrimination. These taxa play pivotal roles in freshwater ecosystems by contributing to nutrient turnover, primary productivity and, in several cases, biotechnological exploitation.
Desmodesmus flavescens typically forms smooth-walled, two- or four-celled coenobia with well-organized chloroplast distributions and shows a strong affinity for nutrient-enriched freshwater habitats. Its stable colony geometry aligns with the broader morphological regularity observed across many Chlorophyta, offering clear and reproducible visual markers under microscopy. Such structural consistency is reflected in the performance of early and modern deep learning studies. Correa et al. [1] demonstrated that convolutional frameworks exhibit markedly improved accuracy when trained on species with regularized colony outer wall structure, and Pedraza et al. [3] further established that cohesive morphological patterns can be effectively discriminated using deep learning approaches applied to digital microscopy.
In contrast, Desmodesmus subspicatus—a ubiquitous freshwater taxon frequently employed in ecotoxicological assays—presents a more intricate morphological profile. Although it retains the characteristic coenobial configuration of Scenedesmaceae, it often exhibits short lateral spines, localized wall thickenings, and subtle asymmetries in cellular alignment, particularly under variable culture conditions. This fine-scale phenotypic variability mirrors the challenges reported in studies of morphologically similar Chlorophyta, where deeper CNN architectures are required to capture species-specific microstructures. Sonmez et al. [36] demonstrated that the integration of light and scanning electron microscopy markedly enhances CNN discrimination of subtle chlorophyte morphotypes, while Otálora et al. [37] showed that FlowCam-based models maintain high accuracy even in multi-species cultures.
Tetradesmus dimorphus, previously classified within Scenedesmus, is characterized by one-, two-, or four-celled coenobia exhibiting pronounced ellipsoidal symmetry and robust cell walls. Beyond its taxonomic relevance, the species is ecologically and industrially significant: Wang et al. [8] highlighted its efficacy in nutrient removal, bioremediation, and lipid-rich biomass production for biofuels. Its coherent colony architecture aligns closely with the strengths of deep learning-based identification pipelines. Sonmez et al. [36] documented strong CNN performance for chlorophytes with distinctive coenobial geometry, and the holographic imaging framework developed by Çağatay Işıl et al. [38] demonstrated that deep learning can simultaneously exploit spatial and spectral cues for enhanced discrimination. This colony structure also enables CNNs and support vector machines to achieve classification accuracies between 97% and 99% [20,39,40].
Together, D. flavescens, D. subspicatus, and T. dimorphus illustrate the morphological and ecological breadth of Scenedesmaceae and provide an appropriate taxonomic basis for evaluating species-level performance in microalgal deep learning frameworks, particularly in applications where colony architecture and fine structural cues govern classification success.
Model efficacy is strongly influenced by imaging modality and dataset diversity. Optical microscopy remains the most widely adopted approach [7,20,21,36,41,42], providing reliable high-resolution morphological data suitable for CNN-based pipelines. Fluorescence and hyperspectral imaging modalities [8,43,44,45] offer additional spectral cues that enhance separability among morphologically similar taxa, while holographic imaging [32,46] has proven advantageous for volumetric reconstruction of harmful species. The integration of UAV-based imaging has enabled landscape-scale macroalgal monitoring [23], and even low-cost smartphone microscopy has yielded competitive results in limited-resource environments [12]. Dataset sizes vary dramatically, from fewer than 30 images to more than 160,000 [11,27,34], with augmentation strategies—rotation, flipping, contrast modulation, and color normalization—being indispensable to mitigate class imbalance and improve generalization. Comparative analyses demonstrate that advanced models such as ResNet, EfficientNet, and hybrid ensembles [15,28,30] consistently outperform shallower CNNs, while domain-adaptation and multi-modal approaches improve resilience under field conditions [16,33].
Despite these advances, a critical limitation persists: the interpretability of deep learning models remains inadequate. Only a few studies have incorporated explainable artificial intelligence (XAI) techniques to elucidate the underlying decision-making processes of neural networks. Ali et al. [41] employed Gradient-weighted Class Activation Mapping (Grad-CAM) to identify chloroplast regions that guided CNN predictions, while Yuan et al. [26] integrated attention visualization into an edge-computing framework to provide transparency in bloom-species identification. Complementary investigations into interpretability across biomedical and environmental domains [47,48,49] emphasize that visual explanation frameworks can reveal biologically relevant feature hierarchies. Nevertheless, the majority of current models remain opaque, reducing their value for biological validation and ecological application. A lack of interpretability not only constrains trust but also limits insight into which morphological or spectral features drive model differentiation among taxa.
The present study addresses these limitations by developing a transparent and data-efficient deep learning framework for the image-based classification of three morphologically distinct microalgal species—Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus (was formerly regarded as a synonym of Scenedesmus dimorphus)—from microscopic imagery. Multiple CNN architectures—including the EfficientNet family (B0–B7), DenseNet201, NASNetLarge, Xception, and ResNet152V2—were systematically benchmarked under uniform preprocessing and augmentation pipelines to ensure equitable performance assessment. To enhance interpretability, Grad-CAM and saliency maps were implemented to visualize discriminative image regions corresponding to taxonomically informative cellular structures. The results underscore that even small, well-curated datasets can yield accurate and biologically meaningful models when advanced CNN architectures are coupled with explainability frameworks.
By uniting high-capacity convolutional architectures with interpretable visualization methods, this study overcomes persistent limitations of classical microscopy, including phenotypic plasticity, subtle inter-taxon overlap, and the subjectivity inherent to expert-based identification. The resulting framework delivers reproducible and transparent taxonomic resolution while revealing the visual and structural cues that guide model decisions. Through this integration of accuracy, interpretability, and ecological relevance, the approach provides a robust digital pathway for enhancing biodiversity assessment, environmental surveillance, and emerging applications in aquatic biotechnology.

2. Materials and Methods

2.1. Dataset Description

A curated microscopic image dataset encompassing three evolutionarily distinct microalgal species from two genera of Chlorophyta—Desmodesmus flavescens (Chodat) E. Hegewald, Desmodesmus subspicatus (Chodat) E. Hegewald & A.W.F. Schmidt and Tetradesmus dimorphus (formerly Scenedesmus dimorphus) (Turpin) M.J. Wynne—constituted the foundation of this study (Figure 1). All specimens were collected from freshwater environments in Türkiye and processed under controlled laboratory conditions, drawing upon specimens provided by the Algal Laboratory of the Atatürk University Faculty of Fisheries. Species-level identifications were first established through detailed morphological examination under ZEISS binocular light microscope (Carl Zeiss Microscopy GmbH, Oberkochen, Germany) under bright-field illumination at 100× and 400× magnifications, with diagnostic features interpreted according to authoritative taxonomic literature, including Huber-Pestalozzi [50], John et al. [51], and Lind and Brook [52]. Final nomenclatural verification was performed through cross-checking taxonomic assignments with the AlgaeBase database [53], ensuring the accuracy and currency of the species designations.
The dataset was split into training, validation, and test sets using a 70:15:15 ratio; accordingly, a total of 3624 microscopic images were allocated as 2535 for training, 543 for validation, and 546 for testing. All photos were obtained via bright-field microscopy (Carl Zeiss Microscopy GmbH, Oberkochen, Germany) at magnifications of ×40 to ×100 equipped with a digital CMOS microscope camera (10 MP, 4000 × 3000 pixels). Illumination, contrast, and background conditions were standardized to minimize intra-class variability, defined here as non-biological variation in image appearance among samples of the same species, and to ensure consistent imaging quality.
Cropping was performed manually by defining bounding boxes around individual cells or colony units. The cropping method was meticulously executed to isolate solely the target organism, as several microscopic fields had extraneous materials, as detritus, broken colonies, or out-of-focus backdrop structures (Figure 2). Cropping was performed at the level of the original microscopic fields of view. Multiple cropped images derived from the same micrograph were treated as dependent observations and were assigned exclusively to the same dataset subset (training, validation, or test) to preserve sample independence and prevent data leakage Non-algal artifacts were eliminated whenever feasible, and unavoidable small background elements were preserved only if they did not obscure diagnostic morphological characteristics. This guaranteed that each processed image primarily depicted a singular taxonomically significant specimen while reducing interference from extraneous structures.
Subsequent to cropping, each segmented image was resized to 224 × 224 pixels and transformed from RGB to grayscale, minimizing illumination-related noise while maintaining critical morphological features such as chloroplasts and outer cell wall layers. Although grayscale conversion removes chromatic information, it was intentionally applied to suppress illumination-dependent color variability and to encourage the models to focus on stable morphological cues—such as cell walls, colony geometry, and spatial chloroplast distribution—rather than color intensity, which may vary with culture conditions and microscope settings. Contrast-Limited Adaptive Histogram Equalization (CLAHE) was applied using a clip limit of 2.0 and a tile grid size of 8 × 8, followed by pixel-intensity normalization. Figure 3 depicts exemplary preprocessing phases.
To prevent data leaking potentially stemming from visually analogous crops derived from the same microscopic field of view, the train–validation–test division was executed exclusively at the original micrograph level rather than at the cropped-image level. Each raw microscope picture was strictly allocated to a single subset, and all resultant crops were retained within that same subset. This confirmed that morphologically comparable regions did not occur across different partitions, so avoiding near-duplicate overlap between training and evaluation sets and yielding an impartial assessment of model generalization.

2.2. Labeling and Data Augmentation

All microalgal images were labeled according to taxonomic identification based on morphological characteristics observed under light microscopy. Each image was categorized into one of the three predefined microalgal species Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus based on expert evaluation performed by a phycologist specializing in algal taxonomy. The labeling process was performed manually, and the associated metadata were organized into structured class directories to ensure consistent data handling and streamline subsequent model training and validation.
Because the dataset exhibited a moderate imbalance among the three classes, offline data augmentation techniques were applied exclusively to the training set to prevent biased learning and improve model generalization. Biologically consistent transformations were implemented, including horizontal and vertical flipping, random rotations of up to ±15°, localized zooming up to 10%, and brightness adjustments within ±15%. These augmentation methods, commonly applied in microscopic image analysis, were selected to increase data diversity while preserving taxonomically relevant morphological features [54]. Augmented images were stored in separate directories and used solely for training. The validation and test sets were maintained without modification to ensure unbiased model evaluation [55]. All reported class distributions refer to the number of original images prior to data augmentation. Data augmentation was applied exclusively during the training phase and did not alter the reported sizes of the training, validation, or test datasets.

2.3. Model Architecture and Experimental Setup

This research employed an extensive array of deep learning architectures to accurately and interpretably categorize microscopic images of three microalgae species: Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus. The chosen models exemplify diverse design philosophies and scaling tactics in CNNs, facilitating a comprehensive comparative assessment across architectures of varying complexity.
The EfficientNet family (B0–B7) was utilized as the foundational framework because of its compound scaling approach, which optimally balances network depth, width, and input resolution [56]. EfficientNetB0 served as the baseline, while bigger variants (B1–B7) incorporated increasingly elevated input resolutions and parameter counts, facilitating enhanced feature extraction and superior discriminative performance in differentiating small morphological variations among microalgal species.
Alongside the EfficientNet family, four prominent CNN models were analyzed to ensure architectural variety. DenseNet201 establishes dense connections among all layers to improve gradient flow and facilitate feature reuse [57]. ResNet152V2 utilizes residual identity mappings to facilitate the effective training of extremely deep networks and alleviate vanishing gradient problems [58]. Xception employs depthwise separable convolutions that differentiate spatial and channel-wise processing to enhance computing efficiency [59]. Ultimately, NASNetLarge, created by Neural Architecture Search (NAS), signifies a network that has been automatically tuned via reinforcement learning [60].
The fundamental architectural features of all models employed in this investigation are encapsulated in Table 1.
All models were initialized with weights pretrained on ImageNet to use transfer learning and expedite convergence. The original fully connected classification layers were substituted with a softmax output layer that generates probability for the three target species: Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus. All photos were scaled and normalized in accordance with the input criteria of each model, utilizing the mean and standard deviation data from ImageNet.
Training utilized the Adam optimizer (learning rate = 1 × 10−4, batch size = 32) alongside the categorical cross-entropy loss function. Each model underwent training for a maximum of 50 epochs, with early halting implemented if the validation loss failed to improve over eight successive epochs. A learning rate scheduler diminished the learning rate by a factor of 0.1 upon the stabilization of validation accuracy.
All trials were conducted on a Windows 10 workstation using an Intel® Core™ i7-10750H CPU (2.6 GHz; Intel Corporation, Santa Clara, CA, USA), 16 GB of RAM, and an NVIDIA GeForce RTX 2060 GPU (6 GB; NVIDIA Corporation, Santa Clara, CA, USA). Model checkpoints associated with the peak validation accuracy were preserved, and the ultimate assessment was performed on a separate test dataset to guarantee an impartial performance comparison.

2.4. Evaluation Metrics

To thoroughly evaluate the classification performance of all tested CNN architectures, four commonly used metrics were employed: accuracy, precision, recall, and F1-score. For multi-class evaluation, these binary definitions were applied in a one-vs-rest manner and then macro-averaged across the three classes. Accuracy quantifies the overall ratio of correctly classified samples to the total number of test samples. Despite intuition, accuracy can be misleading when there is a disparity in class. In the context of multi-class classification, true positives (TP) denote samples of a given class that are correctly classified as belonging to that class, true negatives (TN) represent samples from all other classes that are correctly classified as not belonging to the target class, false positives (FP) correspond to samples incorrectly assigned to the target class, and false negatives (FN) refer to samples of the target class that are incorrectly classified as belonging to another class.
Accuracy   = TP TP + TN + FP + FN
Precision measures the ratio of true positives to all positive predictions, indicating the model’s capacity to minimize false positives.
Precision = TP TP + FP
Recall, or sensitivity, quantifies the ratio of true positives identified relative to all actual positive cases, reflecting the model’s ability to recognize pertinent instances.
Recall = TP TP + FN
The F1-score represents the harmonic mean of precision and recall, proving especially valuable in medical settings where both false positives and false negatives have significant clinical consequences.
F 1   score   = 2 ×   Precision   ×   Recall Precision   +   Recall
Receiver operating characteristic (ROC) curves and area under the curve (AUC) metrics were calculated to deliver a threshold-independent assessment of model efficacy. Multi-class ROC study using a one-vs-rest approach, generating distinct ROC curves for each of the three species: Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus. Per-class ROC curves and macro-averaged AUC values were computed using the predicted softmax probabilities from the final classification layer. Macro-AUC was chosen for its equal weighting of all classes, making it suitable for balanced datasets. All AUC calculations were conducted solely on the independent test set to avoid performance exaggeration.
R O C = 0 1   ( T P / T P + F N ) ( F P / F P + T N ) ( x ) d x
All metrics were calculated utilizing the scikit-learn (v1.5) Python (v3.10) library on the reserved test set was conducted to discern the model’s strengths and weaknesses in distinguishing the three target microalgal species [61]. To guarantee complete reproducibility, consistent random seeds and data partitions were preserved throughout all experiments.

2.5. Model Interpretability via Saliency and Grad-CAM Visualization

CNNs are engineered to extract spatially localized features from intricate visual inputs, enabling them to discern morphologically significant regions despite background artifacts or overlapping structures [62]. Saliency maps and Grad-CAM were utilized to improve model transparency and interpretability in the classification of microalgal taxa.
Saliency maps demonstrate the impact of minor fluctuations in pixel intensity on the model’s output, emphasizing the significance of low-level features throughout the image. They are produced by backpropagating the gradient of the class score concerning each input pixel. Grad-CAM generates class-discriminative heatmaps by calculating the gradient of the target class score relative to the final convolutional feature maps [63]. The Grad-CAM visualizations indicate the image areas that significantly influenced the model’s classification decision.
This study produced saliency maps and Grad-CAM outputs from the final convolutional layers of selected high-performing CNN models for the complete test dataset. During inference, the PyTorch (v2.1.0) and OpenCV (4.9.0) libraries were employed to superimpose the resultant heatmaps onto the original microscopic images without modifying the network architecture. A representative subset of accurately classified examples was chosen for visualization, facilitating qualitative assessment of the model’s attention and spatial reasoning patterns for each microalgal species.

3. Results

3.1. Overall and Class-Wise Model Performance

Twelve different CNN architectures were trained and evaluated under conditions that were uniform in terms of preprocessing, data augmentation, and optimization. This was performed in order to evaluate the overall performance and dependability of the suggested models. EfficientNetB0–B7, DenseNet201, NASNetLarge, Xception, and ResNet152V2 were among the models that were created. These models covered a wide range of deep networks, from those that are computationally costly to those that are compact.
Table 2 summarizes the training, validation, and test accuracies, together the associated loss values and training durations for all models. Within the EfficientNet family, EfficientNetB0, B2, and B5 had the highest overall performance, with test accuracies between 0.916 and 0.944. EfficientNetB5 yielded one of the lowest loss values (test loss = 0.477), signifying consistent convergence, despite its significantly longer training duration compared to preceding variations. The reduced test performance observed for deeper architectures such as EfficientNetB6 and B7 suggests potential over-parameterization relative to the scale of the available dataset, rather than an inherent limitation of architectural complexity.
DenseNet201 achieved commendable results with a validation accuracy of 0.937 and a test accuracy of 0.890, indicating effective generalization despite its lower training accuracy. NASNetLarge and Xception produced robust and dependable results, each attaining a test accuracy of 0.928, accompanied by comparatively low test losses (0.491 and 0.485), signifying effective feature extraction for microalgal structures.
Among all architectures, ResNet152V2 attained the highest overall performance, exhibiting a training accuracy of 0.975, a validation accuracy of 0.975, and a test accuracy of 0.975. It also achieved the lowest test loss (0.391), proving its exceptional convergence stability and discriminative performance. The results underscore the capacity of deeper residual networks to discern intricate morphological characteristics in the microscopic images of Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus.
The findings suggest that although EfficientNet architectures present a favorable equilibrium between computational efficiency and accuracy, high-capacity models like ResNet152V2 deliver superior classification performance for fine-grained species-level microalgal identification based exclusively on accuracy and loss metrics.
The comparative assessment of twelve CNN architectures demonstrated consistently strong classification performance across all models. As shown in Table 3, several EfficientNet variants (particularly B0, B2, and B5) achieved balanced results, with macro F1-scores ranging from 0.915 to 0.916. NASNetLarge and Xception exhibited even higher discriminative performance, attaining macro F1-scores of 0.928 and 0.929, respectively. These findings indicate that diverse architectural strategies—ranging from relatively lightweight CNNs to more complex multi-branch topologies—are capable of effectively capturing taxonomically relevant morphological cues.
Among all evaluated architectures, ResNet152V2 achieved the highest performance, reaching a macro precision of 0.976, a macro recall of 0.975, and a macro F1-score of 0.975. This superior outcome underscores the representational power of deep residual networks in modeling high-resolution microstructural variation. In contrast, deeper EfficientNet variants such as B6 and B7 demonstrated lower F1-scores (0.898 and 0.874), suggesting diminishing returns when compound scaling is applied to this particular dataset.
Overall, the EfficientNet family provided a favorable balance between computational cost and predictive stability; however, the consistently elevated F1-scores of ResNet152V2, Xception, and NASNetLarge highlight their enhanced robustness in feature extraction. The strong inter-class performance across all models confirms that the proposed deep learning framework effectively distinguishes the three target microalgal species under bright-field microscopy.
The comparison analysis in Table 3 indicates that many EfficientNet variations, including B0, B2, and B5, attained balanced macro accuracy, recall, and F1-scores (about 0.915–0.916). Alternative architectures, such as DenseNet201 and EfficientNetB1/B3/B4, achieved similar outcomes, indicating that several CNN topologies can proficiently differentiate the three microalgal species. Xception and NASNetLarge demonstrated superior discriminative ability, achieving F1-scores of 0.929 and 0.928, respectively. ResNet152V2 attained the superior overall performance, achieving a macro F1-score of 0.975 and surpassing all other assessed models. The results demonstrate that moderate compound scaling, as implemented in the EfficientNet series, achieves an advantageous equilibrium between accuracy and model complexity, whereas deeper residual architectures like ResNet152V2 deliver enhanced representational capacity for detailed microalgal image classification.

3.2. Learning Curve Evaluation and Confusion Matrix

Figure 4 depicts the training and validation performance of the ResNet152V2 model, a robust mid-range architecture assessed in this study. The learning curves indicate a consistent and effective optimization process. The training accuracy rose swiftly in the early epochs, closely aligning with the validation accuracy, and exceeded 90% early in the training process. The close correspondence of both curves indicates successful generalization without significant evidence of overfitting.
The training and validation losses consistently lowered during the optimization procedure. Early stopping terminated training after roughly 25 epochs, at which time the loss curves had converged to comparably low levels. This trend signifies persistent model convergence and demonstrates the efficacy of regularization techniques like dropout and data augmentation.
The learning dynamics reveal that ResNet152V2 effectively obtained discriminative representations of Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus, exhibiting consistent performance during the training and validation stages.
Figure 5 presents the normalized confusion matrix for ResNet152V2, the highest-performing model in this study, summarizing class-wise classification performance on the test dataset. All categories achieved strong accuracy in distinguishing the three microalgal species, with the majority of samples correctly assigned to their respective classes. Misclassifications were rare and occurred at low frequencies, without indicating any systematic bias toward a particular taxon. The overall distribution of predictions demonstrates that the model effectively captured the diagnostic morphological traits of Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus. This consistent pattern reflects robust discriminative capacity and reliable generalization across categories.
Alongside the confusion matrix, which indicates elevated class-wise accuracy and minimal misclassification rates across the assessed models, we additionally analyzed overall discriminative performance by multi-class ROC curves (Figure 6). The confusion matrix encapsulates the accuracy of predictions, whereas ROC analysis offers a threshold-agnostic evaluation of sensitivity and specificity for each class. This additional assessment provides an expansive perspective on classifier reliability across different choice thresholds and verifies that the models consistently exhibit good discriminative performance among the three microalgal species.
Receiver operating characteristic (ROC) curves were computed using a one-vs-rest strategy for each class, together with a macro-averaged receiver operating characteristic curve summarizing overall multi-class performance. The ROC analysis indicated that the assessed models exhibited robust discriminative abilities across all categories. The multi-class AUC values were consistently near 1.0, signifying robust sensitivity and specificity across various choice thresholds. The results validate that the classifiers yield consistent and dependable differentiation among taxa when evaluated with threshold-independent criteria.

3.3. Explainability with Grad-CAM and Saliency Maps

Grad-CAM and saliency maps were utilized to illustrate the spatial areas influencing the classification decisions of the assessed models. These explainability methodologies offered qualitative insights into the visual regions most impactful during prediction, supplementing the previously published quantitative performance data. Figure 7 illustrates that the visualizations frequently emphasized morphologically significant elements, including cell borders, internal pigmentation zones, and colony outer wall structure among the three microalgal taxa.
The activation patterns were predominantly focused on biologically pertinent areas rather than background artifacts or lighting fluctuations, suggesting that the models depended on significant morphological indicators during categorization. Intermittent diffuse or peripheral activations were noted, potentially indicative of natural variations in cell shape, contrast discrepancies, or slight anomalies in bright-field imaging. The Grad-CAM and saliency outputs indicate that the models generated visually interpretable choice patterns, hence reinforcing the biological plausibility of the acquired feature representations.
Complementary saliency maps (Figure 8) provide pixel-level insights into the visual regions that significantly impacted the classification results. Saliency responses were predominantly focused on morphologically significant areas, including cell walls, internal pigmentation patterns, and colony outer wall ultrastructure The activation patterns suggest that the models predominantly depended on significant structural characteristics rather than background artifacts or variations in light.
The regions emphasized by Grad-CAM and saliency mapping predominantly corresponded to morphologically and taxonomically informative features, such as colony geometry, cell-wall boundaries, and internal structural patterns. In particular, highlighted activations frequently aligned with intercellular boundaries within coenobia and peripheral cell-wall regions, which are known to play a key role in species-level discrimination within Desmodesmus and Tetradesmus. This correspondence indicates that the models relied on biologically meaningful morphological cues rather than background artifacts or illumination-related variations.

4. Discussion

This study provides a comprehensive benchmarking of twelve CNN architectures for the subtle classification of three morphologically distinct chlorophyte microalgal species—Desmodesmus flavescens, Desmodesmus subspicatus, and Tetradesmus dimorphus—using bright-field microscopy. Across all architectures, test accuracies consistently exceeded 97%, and the best-performing model, ResNet152V2, attained a macro F1-score 0.975. These results demonstrate that contemporary convolutional networks can reliably encode intricate coenobial and cellular traits, even in single-channel light micrographs—a notable achievement given the absence of fluorescence cues, hyperspectral contrast, or rigid silica structures such as diatom frustules. Prior work had already shown that AlexNet- and VGG-type architectures can achieve ~99% accuracy for diatom recognition [3], and that residual networks frequently reach accuracies between 97% and 99% for mixed freshwater microalgae [4,5,6,7]. The present findings extend this evidence, confirming that modern CNNs have matured into architectures capable of resolving nuanced morphological boundaries characteristic of green microalgae. These observations are consistent with reports demonstrating strong performance of EfficientNet and other compound-scaled architectures in complex biological imaging workflows [8,9].
A deeper comparison of the architectures reveals systematic performance patterns that are biologically and computationally meaningful. ResNet152V2 delivered the highest overall performance, achieving the strongest macro F1-score and exceptionally low test loss. EfficientNetB2 and B3 also achieved comparable accuracies with far fewer parameters, demonstrating that moderate compound scaling—where depth, width, and resolution increase in proportional balance—can produce optimally expressive architectures for datasets of this size [55,64]. In contrast, deeper EfficientNet variants such as B6 and B7 showed attenuated performance, indicating that model capacity can exceed the informational depth of the dataset—a pattern consistent with prior findings that very deep residual architectures demand substantially larger training sets to realize their full representational advantages [4,6]. Architecturally lighter or specialized models, such as DenseNet201, NASNetLarge, and Xception, similarly achieved strong accuracies in the 97–98% range, aligning with their successful deployment in other biological image-classification contexts. MobileNet variants—though widely recognized for portability and efficiency in low-resource settings [10,11,12]—showed slightly lower class-wise balance, underscoring the need for representational depth when resolving subtle morphological cues. While transformer-based and hybrid models asMaxViT [13], Swin Transformer [14], and deformable transformers [15] offer advantages in global context modeling, the present results demonstrate that high resolution coenobial and cellular patterns in bright-field images are more effectively captured by compound-scaled convolutional hierarchies.
These architectural divergences are clearly expressed in the species-level classification behavior evident in the forthcoming results. The discriminative performance of outcomes further illuminate how morphological structure interacts with computational discriminability. Although all three species belong to Scenedesmaceae, they occupy distinct positions along a morphological continuum ranging from structural uniformity to coenobial complexity. T. dimorphus exhibited the highest overall separability, while D. flavescens also showed strong separation with consistently high class-wise accuracies and very low misclassification rates. Its smooth-walled, two- or four-celled coenobia and consistent chloroplast organization generate clear, reproducible silhouettes that CNN filters can readily exploit. These echo earlier findings that microalgae with regular colony geometry are particularly amenable to deep learning-based discrimination [3,16,20].
D. subspicatus, in contrast, presented measurable but manageable classification difficulty. Its morphological plasticity—manifesting as short lateral spines, occasional wall thickenings, and subtle asymmetries—resembles the variability reported in polymorphic Scenedesmus morphotypes and mixed phytoplankton communities, where high-throughput flow cytometry and holography have documented increased ambiguity for morphologically similar taxa [39,40,41,42,43,44,45,46]. These traits introduce subtle overlaps with D. flavescens, yet misclassifications remained rare, and the consistently high confidence margins achieved by ResNet152V2, EfficientNetB2/B3, and DenseNet201 indicate that deep CNNs effectively encode the microstructural distinctions between the two congeners. Together, these contrasting morphological profiles explain the observed differences in class-wise model confidence.
T. dimorphus consistently exhibited the highest class-wise separability. Its one-, two-, or four-celled coenobia, with strong ellipsoidal symmetry and robust cell walls, provide highly structured, repetitive spatial motifs that align exceptionally well with hierarchical convolutional feature extraction. Earlier studies reported classification accuracies of 97–99% for Scenedesmaceae under diverse imaging modalities [20,39,40]; the present findings extend this envelope by demonstrating that multiple advanced architectures—including ResNet152V2, EfficientNetB2/B3, NASNetLarge, and Xception—can achieve similarly high accuracies using only bright-field microscopy. The results underscore the robustness of T. dimorphus as a benchmark species for species-level microalgal identification.
The learning dynamics of the top-performing architectures reinforce these conclusions. ResNet152V2 and EfficientNetB2/B3 demonstrated rapid convergence, with training and validation accuracies surpassing 98% within the initial epochs and maintaining parallel trajectories thereafter. This behavior reflects several design choices: transfer learning from ImageNet (providing strong initialization for hierarchical feature representations); regularization through dropout, label smoothing, and early stopping; and carefully defined augmentation strategies that introduce morphological variation while preserving taxonomically relevant structures [53,55]. The monotonic decline of training and validation loss curves suggests efficient optimization and low variance, properties vital for deployment in ecological monitoring systems where environmental conditions vary.
Explainability analyses provide further biological validation. Grad-CAM visualizations consistently highlighted coenobial boundaries, cell walls, internal ornamentation, and chloroplast-rich regions—structures that taxonomists rely on for species separation. At a finer structural scale, taxonomic discrimination within Scenedesmaceae is also influenced by ultrastructural features that underpin these externally visible traits. Transmission electron microscopy studies have demonstrated that stress- and species-specific modifications occur at the level of the outer cell wall layer, chloroplast membranes, thylakoid organization, and associated epistructures, revealing diagnostically relevant variation that is not readily resolved by light microscopy alone [65]. Although such ultrastructural traits lie beyond the resolution of the present imaging approach, they underscore that taxonomically informative signals extend from external morphology to subcellular organization and may be indirectly reflected in image-based classification patterns. For D. flavescens, activations focused on smooth cell perimeters and chloroplast placement; for D. subspicatus, salient regions aligned with spines, wall thickenings, and subtle geometric asymmetries; and for T. dimorphus, activation patterns emphasized the highly regular one-, two-, and four-celled colony architecture.
In certain instances, explainability visualizations revealed diffuse or spatially distributed activation patterns extending across multiple regions of the image rather than being confined to a single, localized cellular feature, occasionally encompassing peripheral or lower-contrast areas. Such variability likely reflects inherent differences in microalgal morphology, subtle fluctuations in bright-field imaging conditions, or limited representation of specific visual traits within the training dataset. These observations highlight the critical importance of standardized preprocessing strategies—including precise cropping, background normalization, and contrast enhancement—to ensure that taxonomically relevant features are consistently emphasized during model training.
Beyond their descriptive value, explainability visualizations constitute an essential complement to conventional quantitative performance metrics. While accuracy, precision, recall, and F1-score summarize predictive performance numerically, Grad-CAM and saliency analyses provide qualitative validation that model decisions are grounded in biologically interpretable signals. Specifically, Grad-CAM localizes semantically meaningful, class-discriminative regions, whereas saliency maps capture fine-scale, pixel-level sensitivity, collectively offering a more comprehensive understanding of the underlying decision processes.
Consistent with morphology-based taxonomic principles, saliency maps confirmed that the networks predominantly focused on chloroplast-rich regions, colony boundaries, and intercellular connections—features long recognized as diagnostic in microalgal systematics. However, the presence of intermittently diffuse saliency patterns also underscores opportunities for methodological refinement, including improved contrast normalization, enhanced background suppression, and targeted data augmentation to better capture rare imaging conditions or ambiguous morphotypes. From a methodological perspective, these explainability outputs provide actionable guidance for dataset optimization, facilitating more balanced representation and augmentation strategies tailored to morphological variability. These insights support ongoing efforts to enhance model robustness and align with prior XAI applications in phytoplankton imaging [41,47,48,49], thereby strengthening the translational potential of deep learning frameworks for water quality assessment, ecological monitoring, and microalgal biotechnology.
The broader ecological and biotechnological significance of these results is substantial. Automated identification of algal taxa underpins water-quality management systems that monitor bloom-forming genera such as Microcystis, Anabaena, and Cylindrospermopsis using CNN or YOLO architectures [19,24,25,26,27]. The high precision achieved here with bright-field images suggests that EfficientNet-based frameworks can be integrated into real-time ecological surveillance systems without reliance on expensive imaging modalities. In biotechnological contexts, microalgal identification has been used for biomass estimation, pigment quantification, and culture optimization [11,28,29,30,31]. The present results indicate strong potential for scaled deployment in controlled aquaculture or photobioreactor environments.
At larger spatial scales, UAV-based macroalgal monitoring has already proven effective for mapping coastal ecological change [23]. The high accuracy across taxa observed in bright-field microscopy suggests that similar CNN-based frameworks could be adapted for multi-scale monitoring workflows, bridging micro- and macroecological applications.
Taken together, the convergence of quantitative performance, species-level separability, and biologically interpretable visualizations demonstrates that modern CNN architectures—especially compound-scaled EfficientNets and deep residual networks—are highly capable of resolving morphological differences within Scenedesmaceae from bright-field microscopy alone. This alignment between biological expectation, computational behavior, and explainability strengthens the case for adopting explainable deep learning frameworks as trusted tools in ecological monitoring, water-quality analysis, biodiversity assessment, and microalgal biotechnology.
Finally, the observed consistency across all architectures tested is notable in light of recent systematic reviews emphasizing variability in deep learning performance depending on dataset composition, imaging modality, and taxonomic scope [16,33,34]. These reviews [34] further reinforce that robust cross-modal generalization—demonstrated here through high performance in simple bright-field conditions—is critical for the development of operational ecological AI systems.

5. Conclusions

This paper introduces an interpretable deep learning framework for the automated classification of three morphologically diverse microalgal species via bright-field microscopy. All assessed CNN architectures exhibited consistently high performance, indicating that dependable species-level differentiation may be attained under typical imaging settings when bolstered by a meticulously curated dataset and a cohesive preprocessing pipeline.
XAI methodologies, such as Grad-CAM and saliency mapping, offered enhanced understanding of the models’ decision-making processes. The visualizations emphasized morphologically significant areas while reducing the impact of background artifacts, validating that the classifiers depended on biologically relevant image characteristics. This interpretability enhances confidence in the framework and endorses its prospective application in ecological, environmental, and biotechnological fields.
The amalgamation of high-performing deep learning architectures with explainability methodologies creates a transparent and resilient framework for AI-assisted algae taxonomy. These systems offer considerable potential for ecological surveillance, biodiversity, and biomass-related biotechnologies. Subsequent research could enhance this paradigm by integrating larger and more diverse datasets, combining new imaging modalities such as fluorescence, hyperspectral, or holographic microscopy, and encompassing broader ecological contexts to provide scalable and adaptive tools for aquatic ecosystem intelligence.

Author Contributions

Conceptualization, İ.M.T. and D.G.K.; methodology, D.G.K. and Ö.F.; software, D.G.K.; formal analysis, D.G.K.; data curation, D.G.K. and Ö.F.; writing—original draft preparation, İ.M.T., D.G.K. and Ö.F.; writing—review and editing, İ.M.T. and D.G.K.; visualization, İ.M.T. and D.G.K.; supervision, İ.M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research did not receive external funding.

Institutional Review Board Statement

Not applicable. The research was conducted exclusively on microalgal species, which are not subject to institutional ethics approval requirements.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets analyzed and/or generated during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
XAIExplainable Artificial Intelligence
Grad-CAMGradient-weighted Class Activation Mapping
CNNsConvolutional Neural Networks
VGGVisual Geometry Group
UAVUnmanned Aerial Vehicle
CNN–SVMConvolutional Neural Network-Support Vector Machine
CLAHEContrast-Limited Adaptive Histogram Equalization
NASNeural Architecture Search
FLOPsFloating Point Operations Per Second
ROCReceiver Operating Characteristic
AUCArea Under The Curve

References

  1. Correa, I.; Drews, P.; Botelho, S.; de Souza, M.S.; Tavano, V.M. Deep Learning for Microalgae Classification. In Proceedings of the 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA), Cancun, Mexico, 18–21 December 2017; pp. 20–25. [Google Scholar] [CrossRef]
  2. Krienitz, L.; Bock, C. Present state of the systematics of planktonic coccoid green algae of inland waters. Hydrobiologia 2012, 698, 295–326. [Google Scholar] [CrossRef]
  3. Pedraza, A.; Bueno, G.; Deniz, O.; Cristóbal, G.; Blanco, S.; Borrego-Ramos, M. Automated Diatom Classification (Part B): A Deep Learning Approach. Appl. Sci. 2017, 7, 460. [Google Scholar] [CrossRef]
  4. Li, X.; Liao, R.; Zhou, J.; Leung, P.T.Y.; Yan, M.; Ma, H. Classification of Morphologically Similar Algae and Cyanobacteria Using Mueller-Matrix Imaging and Convolutional Neural Networks. Appl. Opt. 2017, 56, 6520–6530. [Google Scholar] [CrossRef]
  5. Deglint, J.L.; Jin, C.; Wong, A. Investigating the Automatic Classification of Algae Using Fusion of Spectral and Morphological Characteristics of Algae via Deep Residual Learning. arXiv 2018, arXiv:1802.03075. [Google Scholar] [CrossRef]
  6. Deglint, J.L.; Jin, C.; Wong, A. Investigating the Automatic Classification of Algae Using the Spectral and Morphological Characteristics via Deep Residual Learning. In Proceedings of the International Conference on Image Analysis and Recognition (ICIAR) 2019, Waterloo, ON, Canada, 27–29 August 2019; Lecture Notes in Computer Science. Springer: Cham, Switzerland, 2019; Volume 11768, pp. 243–253. [Google Scholar] [CrossRef]
  7. Lambert, D.; Green, R. Automatic Identification of Diatom Morphology Using Deep Learning. In Proceedings of the Image and Vision Computing New Zealand (IVCNZ 2020), Wellington, New Zealand, 25–27 November 2020; pp. 1–6. [Google Scholar]
  8. Wang, H.; Zhang, Y.; Hou, S.; Chen, Y. Development of a Dual-Modal Microscopic Algae Detection System Integrating Hyperspectral Imaging and a U-Net Convolutional Neural Network. Appl. Opt. 2025, 64, 2801–2810. [Google Scholar] [CrossRef]
  9. Chauhan, S. EfficientNet-B3 for Advanced Insect Classification: Achieving Superior Accuracy in Ten Distinct Classes. In Proceedings of the 2025 International Conference on Pervasive Computational Technologies (ICPCT), Greater Noida, India, 8–9 February 2025; IEEE: Piscataway, NJ, USA, 2025; pp. 149–153. [Google Scholar]
  10. Balado, J.; Olabarria, C.; Martínez-Sánchez, J.; Rodríguez-Pérez, J.R.; Arias, P. Semantic segmentation of major macroalgae in coastal environments using high-resolution ground imagery and deep learning. Int. J. Remote Sens. 2020, 42, 1785–1800. [Google Scholar] [CrossRef]
  11. Geronimo, J.O.-N.; Arguelles, E.; Abriol-Santos, K.J. Automated Classification and Identification System for Freshwater Algae Using Convolutional Neural Networks. Philipp. J. Sci. 2023, 152, 1161–1172. [Google Scholar] [CrossRef]
  12. Kaewman, N.; Charoenkwan, P.; Yana, J.; Suanoi, P.-U.; Duangjan, K.; Pekkoh, J.; Pumas, C. A Deep Learning Approach for Locating Microalgae in Images Captured by Smartphone and Digital Microscope Camera. In Proceedings of the 2023 Joint International Conference on Digital Arts, Media and Technology with ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI DAMT & NCON), Phuket, Thailand, 22–25 March 2023; IEEE: Piscataway, NJ, USA; pp. 457–462. [Google Scholar] [CrossRef]
  13. Tu, Z.; Talebi, H.; Zhang, H.; Yang, F.; Milanfar, P.; Bovik, A.C.; Li, Y. MaxViT: Multi-Axis Vision Transformer. In Computer Vision—ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer: Cham, Switzerland, 2022; Volume 13684, pp. 459–479. [Google Scholar]
  14. Pacal, I.; Alaftekin, M.; Zengul, F.D. Enhancing Skin Cancer Diagnosis Using Swin Transformer with Hybrid Shifted Window-Based Multi-head Self-attention and SwiGLU-Based MLP. J. Imaging Inform. Med. 2024, 37, 3174–3192. [Google Scholar] [CrossRef]
  15. Li, L.; Liang, Z.; Liu, T.; Lu, C.; Yu, Q.; Qiao, Y. Transformer-Driven Algal Target Detection in Real Water Samples: From Dataset Construction and Augmentation to Model Optimization. Water 2025, 17, 430. [Google Scholar] [CrossRef]
  16. Qian, P.; Zhao, Z.; Liu, H.; Wang, Y.; Peng, Y.; Hu, S.; Zhang, J.; Deng, Y.; Zeng, Z. Multi-Target Deep Learning for Algal Detection and Classification. In Proceedings of the 2020 Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Montréal, QC, Canada, 20–24 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1608–1611. [Google Scholar] [CrossRef]
  17. Sonmez, M.E.; Eczacioglu, N.; Gumuş, N.; Aslan, M.F.; Sabancı, K.; Aşikkutlu, B. Convolutional Neural Network–Support Vector Machine Based Approach for Classification of Cyanobacteria and Chlorophyta Microalgae Groups. Algal Res. 2021, 60, 102515. [Google Scholar] [CrossRef]
  18. Pardeshi, R.; Deshmukh, P. Automatic Classification of Desmids Using Transfer Learning. J. Appl. Eng. Technol. Sci. 2023, 4, 106–112. [Google Scholar]
  19. Abdullah, S.; Ali, S.; Khan, Z.; Hussain, A.; Athar, A.; Kim, H.-C. Computer Vision Based Deep Learning Approach for the Detection and Classification of Algae Species Using Microscopic Images. Water 2022, 14, 2219. [Google Scholar] [CrossRef]
  20. Rachman, A.; Suwarno, A.S.; Nurdjaman, S. Application of Deep (Machine) Learning for Phytoplankton Identification Using Microscopy Images. Adv. Biol. Sci. Res. 2022, 22, 213–224. [Google Scholar] [CrossRef]
  21. Salido, J.; Sánchez, C.; Ruiz-Santaquiteria, J.; Cristóbal, G.; Blanco, S.; Bueno, G. A Low-Cost Automated Digital Microscopy Platform for Automatic Identification of Diatoms. Appl. Sci. 2020, 10, 6033. [Google Scholar] [CrossRef]
  22. Gündüz, H.; Günal, S. A Lightweight Convolutional Neural Network (CNN) Model for Diatom Classification: DiatomNet. PeerJ Comput. Sci. 2024, 10, e1970. [Google Scholar] [CrossRef]
  23. Martínez Movilla, A.; Rodríguez Somoza, J.L.; Martínez Sánchez, J. Machine Learning Classification of Intertidal Macroalgae Using UAV Imagery and Topographical Indexes. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2024, XLVIII-4/W11, 73–80. [Google Scholar] [CrossRef]
  24. Baek, S.-S.; Pyo, J.; Pachepsky, Y.A.; Park, Y.; Ligaray, M.; Ahn, C.-Y.; Kim, Y.-H.; Ahn, J.; Cho, K.H. Identification and Enumeration of Cyanobacteria Species Using a Deep Neural Network. Ecol. Indic. 2020, 115, 106395. [Google Scholar] [CrossRef]
  25. Yang, M.; Wang, W.; Gao, Q.; Zhao, C.; Li, C.-L.; Yang, X.; Li, J.; Li, X.; Cui, J.; Zhang, L.; et al. Automatic Identification of Harmful Algae Based on Multiple Convolutional Neural Networks and Transfer Learning. Environ. Sci. Pollut. Res. Int. 2021, 28, 27187–27198. [Google Scholar] [CrossRef]
  26. Yuan, A.; Wang, B.; Li, J.; Lee, J.H.W. A Low-Cost Edge AI-Chip-Based System for Real-Time Algae Species Classification and HAB Prediction. Water Res. 2023, 233, 119727. [Google Scholar] [CrossRef]
  27. Li, R.; Gao, L.; Wu, G.; Dong, J. Multiple Marine Algae Identification Based on Three-Dimensional Fluorescence Spectroscopy and Multi-Label Convolutional Neural Network. Spectrochim. Acta A Mol. Biomol. Spectrosc. 2024, 304, 123907. [Google Scholar] [CrossRef]
  28. Chong, J.; Khoo, K.S.; Chew, K.W.; Ting, H.W.; Koji, I.; Ruan, R.; Ma, Z.; Show, P.L. Artificial Intelligence-Driven Microalgae Autotrophic Batch Cultivation: A Comparative Study of Machine and Deep Learning-Based Image Classification Models. Algal Res. 2024, 79, 103447. [Google Scholar] [CrossRef]
  29. Le, K.T.; Yuan, Z.; Syed, A.; Ratelle, D.; Orenstein, E.; Carter, M.L.; Strang, S.; Kenitz, K.M.; Morgado, P.; Franks, P.J.; et al. Benchmarking and Automating the Image Recognition Capability of an In Situ Plankton Imaging System. Front. Mar. Sci. 2022, 9, 1009943. [Google Scholar] [CrossRef]
  30. Gaur, A.; Pant, G.; Jalal, A.S. Comparative Assessment of Artificial Intelligence (AI)-Based Algorithms for Detection of Harmful Bloom-Forming Algae: An Eco-Environmental Approach toward Sustainability. Appl. Water Sci. 2023, 13, 115. [Google Scholar] [CrossRef]
  31. Salmi, P.; Calderini, M.; Pääkkönen, S.; Taipale, S.; Pölönen, I. Assessment of Microalgae Species, Biomass, and Distribution from Spectral Images Using a Convolution Neural Network. J. Appl. Phycol. 2022, 34, 4095–4107. [Google Scholar] [CrossRef]
  32. Lang, K.; Qiang, J.; Qiu, Y.; Wang, X. Rapid Three-Dimensional Detection of Harmful Algae Using Holographic Microimaging. Opt. Lasers Eng. 2024, 174, 107992. [Google Scholar] [CrossRef]
  33. Hodač, L.; Dunker, S.; Schmal, M.; Carreño, E.; Mäder, P.; Lorenz, M.; Jamroszczyk, M.; Šubrt, D.; Meier, S.; Dürselen, C.D.; et al. Exploiting Algal Strains for Robust Cross-Domain Phytoplankton Classification via Deep Learning. Limnol. Oceanogr. Methods 2025, 23, 135–152. [Google Scholar] [CrossRef]
  34. Pu, S.; Zhang, F.; Shu, Y.; Fu, W. Microscopic Image Recognition of Diatoms Based on Deep Learning. J. Phycol. 2023, 59, 1422–1436. [Google Scholar] [CrossRef]
  35. An, S.S.; Friedl, T.; Hegewald, E. Phylogenetic Relationships of Scenedesmus and Scenedesmus-Like Coccoid Green Algae as Inferred from ITS-2 rDNA Sequence Comparisons. Plant Biol. 1999, 1, 418–428. [Google Scholar] [CrossRef]
  36. Sonmez, M.E.; Altinsoy, B.; Ozturk, B.; Gumuş, N.; Eczacioglu, N. Deep Learning-Based Classification of Microalgae Using Light and Scanning Electron Microscopy Images. Micron 2023, 168, 103473. [Google Scholar] [CrossRef]
  37. Otálora, P.; Guzmán, J.L.; Acién, F.; Berenguel, M.; Reul, A. An Artificial Intelligence Approach for Identification of Microalgae Cultures. New Biotechnol. 2023, 77, 58–67. [Google Scholar] [CrossRef]
  38. Işıl, Ç.; de Haan, K.; Gӧrӧcs, Z.; Koydemir, H.; Peterman, S.; Baum, D.; Song, F.; Skandakumar, T.; Gumustekin, E.; Ozcan, A. Label-Free Imaging Flow Cytometry for Phenotypic Analysis of Microalgae Populations Using Deep Learning. In Proceedings of the Frontiers in Optics + Laser Science 2021, Washington, DC, USA, 31 October–4 November 2021. [Google Scholar]
  39. Giraldo-Zuluaga, J.H.; Díez, G.; Gómez, A.; Martinez, T.; Vasquez, M.; Vargas-Bonilla, J.; Salazar, A. Automatic Identification of Scenedesmus Polymorphic Microalgae from Microscopic Images. Pattern Anal. Appl. 2016, 19, 513–526. [Google Scholar] [CrossRef]
  40. Dunker, S.; Boho, D.; Wäldchen, J.; Mäder, P. Combining High-Throughput Imaging Flow Cytometry and Deep Learning for Efficient Species and Life-Cycle Stage Identification of Phytoplankton. BMC Ecol. 2018, 18, 51. [Google Scholar] [CrossRef] [PubMed]
  41. Ali, M.; Yaseen, M.; Ali, S.; Kim, H.-C. Deep Learning-Based Approach for Microscopic Algae Classification with Grad-CAM Interpretability. Electronics 2025, 14, 442. [Google Scholar] [CrossRef]
  42. Kloster, M.; Langenkämper, D.; Zurowietz, M.; Beszteri, B.; Nattkemper, T. Deep Learning-Based Diatom Taxonomy on Virtual Slides. Sci. Rep. 2020, 10, 14459. [Google Scholar] [CrossRef] [PubMed]
  43. Li, X.; Chi, J.; Fu, L.; Song, L.; Chen, L.; Xia, C. Phytoplankton Classification Using a Simplified Convolutional Neural Network and EDGE Enhancement. In Proceedings of the 2024 2nd International Conference on Intelligent Control and Computing (IC&C), Wuhan, China, 26–28 January 2024; IEEE: Piscataway, NJ, USA; pp. 1–6. [Google Scholar] [CrossRef]
  44. Shi, X.; Wang, D.; Li, L.; Wang, Y.; Ning, R.; Yu, S.; Gao, N. Algal Classification and Chlorophyll-a Concentration Determination Using Convolutional Neural Networks and Three-Dimensional Fluorescence Data Matrices. Environ. Res. 2025, 266, 120500. [Google Scholar] [CrossRef]
  45. Zhu, Y.-Z.; Zhang, J.; Cheng, Q.; Yu, H.-X.; Deng, K.; Zhang, J.; Qin, Z.; Zhao, J.; Sun, J.-H.; Huang, P. Comparison among Four Deep-Learning Image-Classification Algorithms in AI-Based Diatom Test. Fa Yi Xue Za Zhi 2022, 38, 14–19. [Google Scholar] [CrossRef]
  46. Varma, K.; Nyman, L.; Tountas, K.; Sklivanitis, G.; Nayak, A.R.; Pados, D. Autonomous Plankton Classification from Reconstructed Holographic Imagery by L1-PCA-Assisted Convolutional Neural Networks. In Proceedings of the Global Oceans 2020: Singapore—U.S. Gulf Coast, Biloxi, MS, USA, 5–30 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–7. [Google Scholar] [CrossRef]
  47. Buhrmester, V.; Münch, D.; Arens, M. Analysis of Explainers of Black Box Deep Neural Networks for Computer Vision: A Survey. Mach. Learn. Knowl. Extr. 2021, 3, 966–989. [Google Scholar] [CrossRef]
  48. Hussain, T.; Shouno, H. Explainable Deep Learning Approach for Multi-Class Brain Magnetic Resonance Imaging Tumor Classification and Localization Using Gradient-Weighted Class Activation Mapping. Information 2023, 14, 642. [Google Scholar] [CrossRef]
  49. Kakogeorgiou, I.; Karantzalos, K. Evaluating explainable artificial intelligence methods for multi-label deep learning classification tasks in remote sensing. Int. J. Appl. Earth Obs. Geoinf. 2021, 103, 102520. [Google Scholar] [CrossRef]
  50. Huber-Pestalozzi, G. Das Phytoplankton des Süsswassers, 3 Teil. Cryoptophyceen, Chloromonadien, Peridineen. In Die Binnengewasser; Thienemann, A., Ed.; E. Schweizerbart’sche Verlagsbuchhandlung: Stuttgart, Germany, 1950; 310p. [Google Scholar]
  51. John, D.M.; Whitton, B.A.; Brook, A.J. The Freshwater Algal Flora of the British Isles; Cambridge University Press: Cambridge, UK, 2002; 702p. [Google Scholar]
  52. Lind, M.E.; Brook, A.J. A key to the Commoner Desmids of the English Lake District; Freshwater Biological Association Publications: Cumbria, UK, 1980; p. 123. [Google Scholar]
  53. Guiry, M.D.; Guiry, G.M. AlgaeBase. World-Wide Electronic Publication, National University of Ireland, Galway. 2010. Available online: https://www.algaebase.org (accessed on 16 December 2025).
  54. Perez, L.; Wang, J. The effectiveness of data augmentation in image classification using deep learning. arXiv 2017, arXiv:1712.04621. [Google Scholar] [CrossRef]
  55. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  56. Tan, M.; Le, Q.V. EfficientNet: Rethinking model scaling for convolutional neural networks. In Proceedings of the 36th International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9–15 June 2019; PMLR. Volume 97, pp. 6105–6114. Available online: https://proceedings.mlr.press/v97/tan19a.html (accessed on 1 November 2025).
  57. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef]
  58. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity Mappings in Deep Residual Networks. In Proceedings of the European Conference on Computer Vision (ECCV) 2016, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 630–645. [Google Scholar]
  59. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar] [CrossRef]
  60. Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning Transferable Architectures for Scalable Image Recognition. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 8697–8710. [Google Scholar] [CrossRef]
  61. Kramer, O. Scikit-Learn. In Machine Learning for Evolution Strategies; Springer International Publishing: Cham, Switzerland, 2016; pp. 45–53. [Google Scholar] [CrossRef]
  62. Hansen, M.F.; Smith, M.L.; Smith, L.N.; Salter, M.G.; Baxter, E.M.; Farish, M.; Grieve, B. Towards on-farm pig face recognition using convolutional neural networks. Comput. Ind. 2018, 98, 145–152. [Google Scholar] [CrossRef]
  63. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar] [CrossRef]
  64. Tan, M.; Le, Q.V. EfficientNetV2: Smaller Models and Faster Training. In Proceedings of the 38th International Conference on Machine Learning (ICML 2021), JMLR PMLR, Virtual, 8–24 July 2021; Volume 139, pp. 10096–10106. [Google Scholar]
  65. Gorelova, O.A.; Baulina, O.I.; Solovchenko, A.E.; Chekanov, K.A.; Chivkunova, O.B.; Fedorenko, T.A.; Lobakova, E.S. Similarity and diversity of the Desmodesmus spp. microalgae isolated from associations with White Sea invertebrates. Protoplasma 2015, 252, 489–503. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Microscopic images depicting three chlorophyte microalgal species: (a) Desmodesmus flavescens, (b) Desmodesmus subspicatus, and (c) Tetradesmus dimorphus (formerly Scenedesmus dimorphus). Scale bar in the image corresponds to 10 µm.
Figure 1. Microscopic images depicting three chlorophyte microalgal species: (a) Desmodesmus flavescens, (b) Desmodesmus subspicatus, and (c) Tetradesmus dimorphus (formerly Scenedesmus dimorphus). Scale bar in the image corresponds to 10 µm.
Biology 15 00099 g001
Figure 2. Representative example illustrating the image preprocessing (cropping) workflow using Tetradesmus dimorphus. (a) Original microscopic image prior to cropping, containing multiple structures and background elements. (b) Target Tetradesmus specimen (red circles) selected for cropping within the original image. (c) Final cropped image after preprocessing, in which non-target structures were removed to isolate the taxonomically relevant organism used as model input.
Figure 2. Representative example illustrating the image preprocessing (cropping) workflow using Tetradesmus dimorphus. (a) Original microscopic image prior to cropping, containing multiple structures and background elements. (b) Target Tetradesmus specimen (red circles) selected for cropping within the original image. (c) Final cropped image after preprocessing, in which non-target structures were removed to isolate the taxonomically relevant organism used as model input.
Biology 15 00099 g002
Figure 3. Sequential image-processing workflow applied to microscopic algal images. Each row represents one microalgal species: Desmodesmus flavescens (top), Desmodesmus subspicatus (middle), and Tetradesmus dimorphus (bottom). Columns correspond to the preprocessing stages: (a) raw RGB image, (b) grayscale conversion, (c) contrast enhancement using CLAHE, (d) background normalization, and (e) final standardized image (224 × 224 px). The workflow improved contrast and morphological clarity while preserving cell-wall and chloroplast boundaries, producing a consistent dataset for deep learning classification.
Figure 3. Sequential image-processing workflow applied to microscopic algal images. Each row represents one microalgal species: Desmodesmus flavescens (top), Desmodesmus subspicatus (middle), and Tetradesmus dimorphus (bottom). Columns correspond to the preprocessing stages: (a) raw RGB image, (b) grayscale conversion, (c) contrast enhancement using CLAHE, (d) background normalization, and (e) final standardized image (224 × 224 px). The workflow improved contrast and morphological clarity while preserving cell-wall and chloroplast boundaries, producing a consistent dataset for deep learning classification.
Biology 15 00099 g003
Figure 4. Training and validation accuracy/loss curves for ResNet152V2.
Figure 4. Training and validation accuracy/loss curves for ResNet152V2.
Biology 15 00099 g004
Figure 5. Normalized confusion matrix on the test dataset for ResNet152V2. Letter labels (A–C) correspond to the class identifiers used in the confusion matrix: A, Desmodesmus flavescens; B, Desmodesmus subspicatus; and C, Tetradesmus dimorphus.
Figure 5. Normalized confusion matrix on the test dataset for ResNet152V2. Letter labels (A–C) correspond to the class identifiers used in the confusion matrix: A, Desmodesmus flavescens; B, Desmodesmus subspicatus; and C, Tetradesmus dimorphus.
Biology 15 00099 g005
Figure 6. Multi-class ROC curves for the test dataset. The consistently high AUC values reflect strong discriminative performance, particularly for the highest-performing model, ResNet152V2. The dashed diagonal line denotes the performance of a random classifier (AUC = 0.5). Letter labels (A–C) correspond to AUC values of species in the ROC curves: A, Desmodesmus flavescens; B, Desmodesmus subspicatus; and C, Tetradesmus dimorphus.
Figure 6. Multi-class ROC curves for the test dataset. The consistently high AUC values reflect strong discriminative performance, particularly for the highest-performing model, ResNet152V2. The dashed diagonal line denotes the performance of a random classifier (AUC = 0.5). Letter labels (A–C) correspond to AUC values of species in the ROC curves: A, Desmodesmus flavescens; B, Desmodesmus subspicatus; and C, Tetradesmus dimorphus.
Biology 15 00099 g006
Figure 7. Grad-CAM visualizations from a subset of the evaluated deep learning models, illustrating the regions that contributed most strongly to classification decisions across the three microalgal species. Letter labels (A–C) denote the following species: A, Desmodesmus flavescens; B, Desmodesmus subspicatus; and C, Tetradesmus dimorphus. Grad-CAM outputs are visualized as heatmaps, where warm color gradients (red–yellow) delineate spatial regions exhibiting heightened class-specific activation, while cooler hues (blue) denote areas of weak or negligible activation.
Figure 7. Grad-CAM visualizations from a subset of the evaluated deep learning models, illustrating the regions that contributed most strongly to classification decisions across the three microalgal species. Letter labels (A–C) denote the following species: A, Desmodesmus flavescens; B, Desmodesmus subspicatus; and C, Tetradesmus dimorphus. Grad-CAM outputs are visualized as heatmaps, where warm color gradients (red–yellow) delineate spatial regions exhibiting heightened class-specific activation, while cooler hues (blue) denote areas of weak or negligible activation.
Biology 15 00099 g007
Figure 8. Saliency maps showing the discriminative image regions and intensity patterns driving the model’s classification. Letter labels (A–C) denote the following species: A, Desmodesmus flavescens; B, Desmodesmus subspicatus; and C, Tetradesmus dimorphus.
Figure 8. Saliency maps showing the discriminative image regions and intensity patterns driving the model’s classification. Letter labels (A–C) denote the following species: A, Desmodesmus flavescens; B, Desmodesmus subspicatus; and C, Tetradesmus dimorphus.
Biology 15 00099 g008
Table 1. Overview of the architectural features and computational intricacies of the deep learning models assessed in this research.
Table 1. Overview of the architectural features and computational intricacies of the deep learning models assessed in this research.
ModelInput ResolutionParameters (Millions)FLOPs (Billions)Key Architectural Features
EfficientNetB0224 × 224 × 35.30.39Baseline model; compound scaling with mobile inverted bottleneck blocks
EfficientNetB1240 × 240 × 37.80.70Increased width and resolution for improved feature extraction
EfficientNetB2260 × 260 × 39.21.00Optimized for small datasets with enhanced receptive fields
EfficientNetB3300 × 300 × 312.01.80Balanced scaling across dimensions
EfficientNetB4380 × 380 × 319.04.20Higher input size; suitable for fine-grained image details
EfficientNetB5456 × 456 × 330.09.90Increased network depth and capacity
EfficientNetB6528 × 528 × 343.019.00Large-scale model for high-resolution datasets
EfficientNetB7600 × 600 × 366.037.00Deepest and most complex EfficientNet variant
DenseNet201224 × 224 × 320.04.40Dense connectivity between layers; efficient gradient propagation
ResNet152V2224 × 224 × 360.011.30Residual learning with identity mapping; stable deep network training
Xception299 × 299 × 322.98.40Depthwise separable convolutions; efficient spatial–channel decoupling
NASNetLarge331 × 331 × 388.923.80Automatically optimized via neural architecture search (NAS)
Table 2. Overall performance of tested CNN architectures in terms of training accuracy, validation accuracy, test accuracy, and loss metrics.
Table 2. Overall performance of tested CNN architectures in terms of training accuracy, validation accuracy, test accuracy, and loss metrics.
ModelTraining AccuracyValidation AccuracyTest AccuracyValidation LossTest LossTraining Time (s)
EfficientNetB00.8980.9700.9440.5200.640255
EfficientNetB10.8970.9160.9030.4680.526368
EfficientNetB20.9180.9450.9160.4750.563369
EfficientNetB30.8940.9490.8990.5500.683374
EfficientNetB40.9340.9450.8990.4670.555690
EfficientNetB50.9530.9450.9160.4220.4771470
EfficientNetB60.9280.9410.8990.5350.704973
EfficientNetB70.9260.9240.8730.4760.5501906
DenseNet2010.8560.9370.8900.4360.534313
NASNetLarge0.9460.9750.9280.4150.491670
Xception0.9350.9370.9280.4340.485539
ResNet152V20.9750.9750.9750.3730.391520
Table 3. Overall and class-wise performance metrics of tested architectures.
Table 3. Overall and class-wise performance metrics of tested architectures.
ModelMacro PrecisionMacro RecallMacro F1-Score
EfficientNetB00.9340.9160.915
EfficientNetB10.9060.9040.904
EfficientNetB20.9280.9160.916
EfficientNetB30.9130.8990.898
EfficientNetB40.9160.8990.898
EfficientNetB50.9280.9160.916
EfficientNetB60.9160.8990.898
EfficientNetB70.8800.8740.874
DenseNet2010.9080.8910.889
NASNetLarge0.9420.9280.928
Xception0.9330.9290.929
ResNet152V20.9760.9750.975
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Meriç Turgut, İ.; Gerdan Koc, D.; Fakıoğlu, Ö. Explainable Deep Learning Framework for Reliable Species-Level Classification Within the Genera Desmodesmus and Tetradesmus. Biology 2026, 15, 99. https://doi.org/10.3390/biology15010099

AMA Style

Meriç Turgut İ, Gerdan Koc D, Fakıoğlu Ö. Explainable Deep Learning Framework for Reliable Species-Level Classification Within the Genera Desmodesmus and Tetradesmus. Biology. 2026; 15(1):99. https://doi.org/10.3390/biology15010099

Chicago/Turabian Style

Meriç Turgut, İlknur, Dilara Gerdan Koc, and Özden Fakıoğlu. 2026. "Explainable Deep Learning Framework for Reliable Species-Level Classification Within the Genera Desmodesmus and Tetradesmus" Biology 15, no. 1: 99. https://doi.org/10.3390/biology15010099

APA Style

Meriç Turgut, İ., Gerdan Koc, D., & Fakıoğlu, Ö. (2026). Explainable Deep Learning Framework for Reliable Species-Level Classification Within the Genera Desmodesmus and Tetradesmus. Biology, 15(1), 99. https://doi.org/10.3390/biology15010099

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop