Next Article in Journal
Temporal Variation and Human Host Predominance in Aedes aegypti from Coastal and Western Kenya: Insights from Pooled Blood Meal Metagenomics
Previous Article in Journal
Neurological Complications Associated with Respiratory Syncytial Virus Infections: A Scoping Review of Prospective Clinical Trials Conducted in Populations up to 17 Years of Age
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Complete Transfer Learning-Based Pipeline for Discriminating Between Select Pathogenic Yeasts from Microscopy Photographs

by
Ryan A. Parker
1,
Danielle S. Hannagan
2,
Jan H. Strydom
1,
Christopher J. Boon
2,
Jessica Fussell
2,
Chelbie A. Mitchell
2,
Katie L. Moerschel
2,
Aura G. Valter-Franco
2 and
Christopher T. Cornelison
2,*
1
School of Data Science and Analytics, Kennesaw State University, Kennesaw, GA 30144, USA
2
BioInnovation Laboratory, Department of Molecular and Cellular Biology, College of Science and Mathematics, Kennesaw State University, Kennesaw, GA 30144, USA
*
Author to whom correspondence should be addressed.
Pathogens 2025, 14(5), 504; https://doi.org/10.3390/pathogens14050504
Submission received: 28 March 2025 / Revised: 1 May 2025 / Accepted: 14 May 2025 / Published: 21 May 2025
(This article belongs to the Section Fungal Pathogens)

Abstract

:
Pathogenic yeasts are an increasing concern in healthcare, with species like Candida auris often displaying drug resistance and causing high mortality in immunocompromised patients. The need for rapid and accessible diagnostic methods for accurate yeast identification is critical, especially in resource-limited settings. This study presents a convolutional neural network (CNN)-based approach for classifying pathogenic yeast species from microscopy images. Using transfer learning, we trained the model to identify six yeast species from simple micrographs, achieving high classification accuracy (93.91% at the patch level, 99.09% at the whole image level) and low misclassification rates across species, with the best performing model. Our pipeline offers a streamlined, cost-effective diagnostic tool for yeast identification, enabling faster response times in clinical environments and reducing reliance on costly and complex molecular methods.

1. Introduction

Fungal pathogens pose a growing threat to public health, especially in immunocompromised individuals and healthcare settings where vulnerable patients are at heightened risk. The challenge of managing fungal infections is compounded by factors such as rising resistance to antifungal treatments, limited access to rapid and accurate diagnostic methods, and a general lack of awareness surrounding fungal pathogens [1].
One example of a particularly problematic fungal pathogen is Candida auris, first identified in 2009, which is now associated with severe infections across more than 35 countries [1,2,3,4]. C. auris exemplifies the challenges posed by pathogenic yeasts due to its multidrug resistance, and persistence in healthcare environments [5]. Similar to other pathogenic yeast species, its resilience and high mortality rates underscore the urgent need for more effective and accessible diagnostic tools [1,6,7,8].
Identifying Candida species accurately is a critical challenge, as several pathogenic yeasts, including Candida auris, Candida glabrata, and Candida haemulonii, share morphological similarities that complicate diagnosis. This leads to common misidentification between many of these species using current methods, especially when C. auris is involved (Table 1) [6,9]. When misdiagnoses occur, improper treatment can be prescribed, which decrease patient prognosis, as well as increase the risk of outbreaks. Current methods for distinguishing these species—such as whole-genome sequencing, biochemical assays, PCR, and MALDI-TOF MS fingerprinting—are highly accurate but often prohibitively expensive and inaccessible in many healthcare settings [4,6]. MALDI-TOF, in particular, can have costs exceeding USD 200,000, when considering setup and database access costs and even ignoring maintenance requirements. This makes this technology inaccessible to many low and middle-income regions. Alternative diagnostic methods using machine learning to classify pathogenic yeast species from microscopy images have shown promise but remain underexplored [10].
Convolutional neural networks (CNN) are a popular neural network architecture in computer vision that are well known to exceed human abilities in image classification tasks. A few successful architectures include Google Net, which can classify up to 1000 image classes or CIFAR-10 which can classify up to 10 classes of tiny images with over 95% accuracy [12,13]. These networks are composed of two parts: convolutional filters and the classifier. The convolutional filters operate by passing over the pixels of an image to generate features, such as edges or sharpened regions, a process known as convolution (Figure 1) [14,15]. The classifier then takes these features as input and solves the relation between those features and the ‘class’ of the target. This involves complex mathematics, the depth of which is beyond the scope of this work. It should suffice to say that the kernel involves activation functions which impart non-linear transformation on numerical combinations of the input features to approximate the true and very non-linear polynomial function that defines the relation. The output is a set of probabilities that the instance belongs to each class, with the predicted class being the one with the highest value. For this to work, however, the model must first be trained, a process where the model is exposed to class-labeled images, from which it can learn the parameters of the best filters to use and the mathematical function to approximate the true relation. Though this is a very powerful technique, it is quite demanding, requiring significant computational resources and large quantities of high-quality images for training.
Fortunately, much of this burden can be reduced using existing convolutional filters. There is a limited set of possible image features, and sufficiently complex networks, such as those previously mentioned, can extract the majority of these. In a process called transfer learning, these extant networks are utilized by replacing the classifier but retaining the pretrained filters. This reduces training time and allows faster deployment of a model, with far less experimentation with architecture required. After the classifier has been trained relatively well, the network can be fine-tuned by allowing very small changes to the filters to be made, a process which optimizes the network to the image set and maximizes performance. Here, we seek to leverage existing convolutional neural networks but topped with a custom classifier and fine-tuned to a novel image set of micrographs, to distinguish between six related species of yeasts quickly and accurately, as a proof of concept to enable greater effort in such classification.

2. Materials and Methods

2.1. Culture Preparation

Five isolates from the CDC antibiotic resistant bank and one from the NRRL culture collection were used [16,17]. These were Candida krusei (AR-0397), Candida glabrata (AR-0319), Saccharomyces cerevisiae (AR-0400), Candida haemulonii (AR-0393), Candida albicans (NRRL Y-12983) and Candida auris (AR-0384). Saccharomyces cerevisiae was included to serve as a genetically and visually distinct baseline, while the others were chosen due to clinical relevance or their propensity to be confused with C. auris using standard methods. Stock plates using 3% Sabouraud Dextrose Agar (Thermo Fisher Scientific, Item Number: DF0109-17-1, Location: Waltham, MA, USA) of each were prepared from long-term storage stocks stored at −80 °C. Broth cultures of each species were prepared by transferring 2 to 3 colonies from stock plates into 10 mL of Sabouraud Dextrose Broth (Sigma-Aldrich, Item Number: S3306-500G, Location: Waltham, MA, USA). Cultures were incubated for 48 h at 37.0 °C with shaking, after which the cultures were stored at 3 °C for no more than three days, to minimize cell degradation.

2.2. Microscope Slide Preparation and Imaging

Before microscope slide preparation, each culture was concentrated by centrifuging for one minute at 3000 rpm (1693 g, as calculated). For each culture, 5 mL of supernatant was removed, and the pellet was resuspended in the tube by shaking and tapping the tube for 2–3 min. Slides were prepared with 10 µL of the concentrate, with no stains or other visualization aids. Images were taken at 40× magnification using a camera-adapted microscope (Microscope: Motic BA410E Cytology Microscope Motic Instruments Inc., Richmond, BC, Canada; Camera: Canon EOS Rebel T6i DSLR 126571, Cannon Inc., Tokyo, JapanUtility Software (Canon Inc., Tokyo, Japan) to generate an image set of 1000 images per species for further processing. While a standardized image capture protocol was followed, small amounts of variation were allowed in the quality of images, lighting and cell density, to enhance robustness of the dataset in reflecting real-world variability.

2.3. Image Processing and Dataset

Each 6000 × 3368 pixel image was cut up into 512 × 512 pixel patches from the source image (Figure 2 and Figure 3) with excess discarded to meet image dimension criteria. This drastically increased the number of training samples for the machine learning algorithms. To remove potential blank patches, an elementary blank detection algorithm based on thresholding the standard deviation of pixel values was implemented. This approach allows for blank patches with low standard deviation to be dropped, based on a threshold of 2.44 standard deviations from the mean. Due to varied growth rates between species, there were varied amounts of blank images between species. To retain class balance for machine learning training, the minimal number of patches was chosen and used to construct our dataset. This resulted in a dataset with 26,585 images per Candida species. Of these images, 21,268 were used to train the model (the training set), 2659 were used for model tuning (the validation set) and the remaining 2658 were used to evaluate the finished model’s performance (the test set). Together, this was a train, validation, test split of about 80:10:10.

2.4. Model Training

Three models were trained and tested on the dataset generated. The first model was a from-scratch CNN developed using hyperband optimization to find the optimal hyperparameters [18]. Two transfer learning models, one using the VGG16 architecture as the base model, and another using the MobileNet architecture, were trained [19,20]. Both initialized the base models with the ImageNet weights [21]. Model training and fine-tuning were implemented in Python using TensorFlow (v2.12, Google LLC, Mountain View, CA, USA) and Keras (v2.12, originally developed by François Chollet, now maintained by Google LLC, Mountain View, GA, USA). For brevity’s sake, only the architecture details of the best performing model—the VGG16-based network [19,22], originally introduced by Simonyan and Zisserman, are presented (Table 2 and Table 3, Figure 4) [19,23]. For the classifier, a flatten layer was used to convert the input into a one-dimensional array for faster processing. To reduce overfitting, two dropout layers were added, one after the input layer and one following the second dense layer. These layers randomly deactivate nodes from the previous layer during training, to prevent overreliance on specific nodes that leads to overfitting. There were three dense layers (fully connected layers that work to approximate the objective function), two with 256 nodes, and one with 128 nodes. The second layer used the rectified linear unit (ReLU) as the activation function (Equation (1), Figure 5a) [22,24]. The other two used a custom-made function, a parametric form of the hard swish function (Equation (2), Figure 5b) [25,26]. The parametric function contained parameters that the model could learn, six for each node that used the function. Finally, the output layer used a SoftMax activation, which is a multinomial logistic regression that determines the final probability of each class and returns the highest as the prediction. As this was a transfer learning model, the training was performed in stages. First, the base model weights were locked and only the classifier was trained. Then, starting with the final layer of the base model, each convolutional layer was unlocked for fine-tuning, one at a time. In all of these instances, the collection of training images was randomly sorted into batches of sixteen. Each batch was passed through the model, with the weights being updated after each batch. After all the training batches were passed through, then the validation set was passed through to determine the out-of-sample loss (categorical cross-entropy, similar to mean squared error, but better suited to classification problems) and accuracy. This was repeated up to ten times (epochs), for each round of training, possibly stopping early if the loss ceased to improve. After the entirety of the model training was completed, the test set was used to evaluate the model’s final performance. A similar procedure was used to train the other two models, and the model exhibiting the highest accuracy on the test set was chosen for deployment into the final pipeline.
The Rectified Linear Unit.
F x = max 0 , x
The Parametric Hard Swish Function (αi are learnable parameters).
F x = α 1 x + α 2 min max α 3 x + α 4 , x + α 5 , α 6
After the winning model was selected, it was integrated into a pipeline for classifying a complete micrograph (Figure 6). The pipeline begins by taking the input image and creating patches, as previously discussed. Each non-blank patch is passed through the trained model for classification. Finally, a hard voting scheme is used to classify the overall picture. Cumulatively, after the model classifies each patch, the species which received the highest probability in the greatest number of patches is selected as the prediction. To test the overall performance of the pipeline and to detect potential sources of error, the entire image library was passed through the finished pipeline.

3. Results

Each model achieved an overall accuracy above 70% on the image patches (Table 4), with the best performing model being the VGG16-based model, at ~93.6% accuracy. The MobileNet based model exhibited the highest range in class accuracy, showing about a 0.15 difference between its lowest, ~0.65 on C. glabrata to ~0.80 on its highest, C. krusei. The VGG16-based model presented a much tighter range of 0.035. The lowest species accuracy was C. haemulonii at 0.917 and the highest was C. albicans at 0.954. The training curves of the VGG16-based model presented as expected, with both in and out-of-sample accuracy increasing consistently throughout training (Figure 7). Note that the dips in accuracy at regular intervals are normal in the fine-tuning process. As new layers are unlocked for tuning, the combination of the additional parameters and the relatively high learning rate at the start of the training cycle causes temporary disturbances in model performance. Validation accuracy and training accuracy follow an increasing trend, with validation accuracy being slightly lower than that of the training accuracy, however being more resilient to the reductions in accuracy caused by unlocking additional layers. The confusion matrix demonstrated no signs of chronic confusion. The most prominent confusion occurred between C. auris and C. haemulonii, C. haemulonii and C. glabrata, and S. cerevisiae and C. albicans (Table 5). The assembled pipeline presented very high accuracy. Over 98% of images were classified correctly. C. krusei misclassified only one image. The lowest accuracy was for C. haemulonii, at 0.9520 (Table 2, final row). Once more, there were no signs of chronic confusion. The confusion present follows the same patterns described for the isolated model (Table 6).

4. Discussion

The results of this study demonstrate that the model and pipeline successfully distinguish among multiple pathogenic yeast species from microscopy images. Of the three models tested, the VGG16-based classifier performed best, likely due to its extensive convolutional filters that capture a variety of spatial and coloration patterns in microscopic images (Figure 8). In comparison, the MobileNet-based model is optimized for mobile applications, leading to a trade-off in accuracy. The hyperband model required substantial experimentation to optimize, but ultimately did not outperform the VGG16-based approach. Deeper analysis of these two models was not performed, as the performance gap was simply too large. It should also be noted that classical statistics were not performed on the results. Big data have the benefit of avoiding prior assumptions on data distributions, instead learning them directly from the data. Additionally, the use of validation and test sets allows us to directly evaluate out of sample performance, instead of needing to rely on traditional metrics like confidence intervals and p-values.
It is interesting to note that C. haemulonii and C. auris are among the most commonly misidentified species, which is not surprising, considering that this has been reported using other detection methodologies [9,10]. In general, the model does misclassify a small number of image patches. Upon further inspection, it was noted that the image quality in many of these cases was less than ideal. Cell debris or other imperfections were found in a small proportion of the images generated using the imaging methodology (Figure 9). During training, such debris would have little impact, assuming it is a relatively rare occurrence. During inference, however, such debris could pass the blank checking of the preprocessing phase. A weakness of these models is that they will issue a prediction, even if the confidence is low. Thus, given a non-blank patch with only cell debris or proliferous atypical cells, the model will give its best guess, which, in this case, is little better than a random guess. However, the hard-voting scheme protects from this when it occurs in a small number of patches. If the whole image is flawed, however, the problem persists. The simplest solution is to set a confidence threshold and either ignore the patches with low confidence or reject the entire image if the overall confidence falls below the chosen level. Another noteworthy trend amongst the misclassified images was low cell density. Each image could be split into as many as 66 patches. Images with less than 10 usable patches displayed a far higher rate of misclassification, especially when they also contained visual imperfections. This makes sense as it is a basic rule of statistics that lower precision is associated with smaller sample size. In either case, this demonstrates the importance of input image quality and standardization of the culturing, slide preparation and image capture processes.
Despite these minor and correctable or circumventable issues, these results demonstrate a proof-of-concept that a properly trained model and pipeline can, with the assistance of a camera-enabled compound microscope and a GPU-enabled computer, be used to effectively distinguish Candida species in a timely manner, allowing for targeted treatment to be administered rapidly. Though these organisms may not be easily distinguishable by the human eye, the machine learning models are able to pick up on subtle details that differentiate these organisms. In fact, they find underlying patterns in the image at different resolutions and consider them in an integrated way to uncover trends that can be used to discriminate between the classes. While this could have value in a clinical setting in its current form, a major improvement in this process would be enabling it to perform on micrographs prepared directly from patient skin samples, such as those from the groin, axilla, or other common colonization sites. The current requirement is that a pure culture be obtained, a process that requires at least 24–48 h. If direct patient samples can be used, the classification could be carried out in the time it takes for the preparation of the slide, capturing the image, and inferencing through the pipeline, all together in less than 30 min. It is possible that this model could perform well on such samples, but this is uncertain. Microorganisms will exhibit different morphological characteristics in different environments, and these changes may be pronounced enough to undermine the model’s predictive capabilities. If so, an obvious solution would be to retrain the model on patient sample micrograms. However, these images contain significant noise from host cells, bacteria, etc. These can prove troublesome for pure classification models.
A better choice is an object segmentation model, such as a variant Mask-RCNN [26]. This architecture consists of three different networks: a region proposal network for finding areas likely to contain cells, a classifier for identifying the type of cell, and a masking network for highlighting the pixels of the cells. The problem with this approach is that the images are typically down sampled or patched into windows. The former erases critical details and the latter leads to inaccurate masking on patch borders. Thus, we suggest modifying the approach to use a sliding window and incomplete bounding box suppression. This should allow the pipeline to detect and identify each instance from the initial image, while ignoring the background noise, all with only small increases to computational time. By integrating the current architecture into a Mask RCNN-based pipeline, this would enable the identification of the organisms in debris-ridden images, or even those of mixed cultures.

5. Conclusions

This study demonstrates a successful application of convolutional neural networks (CNNs) for classifying select pathogenic yeast species from microscopy images, offering a streamlined, cost-effective approach for yeast identification. The high accuracy achieved by this pipeline highlights the potential of CNN-based models to support faster, accessible diagnostic tools in clinical environments where traditional methods may be prohibitively costly or time-consuming. Considering that the cost of MALDI-TOF can exceed USD 200,000, while our camera and microscope are priced at a total of USD 4400. This makes such an option very attractive for resource-limited settings. While microscopy-based diagnostics are not currently common in clinical settings, the barrier for entry is low and many facilities use microscopy for the preliminary detection of yeasts before submitting for proper identification. The additional burden required to utilize this sort of technology would be as little as taking a photo and uploading it. With an inference time per image of about 1–2 s on a GPU-enabled system, this could be especially valuable in high-volume scenarios, such as outbreaks or triage, where immediate results are crucial to positive outcomes.
While this work provides a strong proof of concept, future efforts will aim to expand the model’s applicability by including a wider range of yeast species (such as C. tropicalis and C. parapsilosis) and isolates, as well as clade-level identification in C. auris, enabling more comprehensive clinical utility. Because microorganisms exhibit phenotypic plasticity in different conditions, the expanding study should include variability in factors like media type and incubation length to capture a more robust sampling of this diversity. Additionally, integrating the current classifier into an object segmentation framework, such as a Mask-RCNN variant, offers potential for isolating and identifying cells within complex or noisy samples. With modifications like a sliding window and incomplete bounding box suppression, this approach could improve identification in debris-ridden images and mixed cultures without sacrificing critical morphological details. Such advancements would further enhance the pipeline’s adaptability and robustness, supporting diverse diagnostic needs in clinical mycology.

Author Contributions

Conceptualization, R.A.P., J.H.S. and C.T.C.; data curation, R.A.P., J.H.S., C.J.B., J.F., C.A.M., K.L.M. and A.G.V.-F.; formal analysis, R.A.P. and J.H.S.; investigation, R.A.P., D.S.H., J.H.S., C.J.B., J.F., C.A.M., K.L.M., A.G.V.-F. and C.T.C.; methodology, R.A.P., J.H.S. and C.T.C.; project administration, R.A.P., J.H.S. and C.T.C.; resources, R.A.P., J.H.S. and C.T.C.; software, R.A.P., J.H.S. and C.J.B.; supervision, R.A.P., J.H.S. and C.T.C.; validation, R.A.P. and J.H.S.; Visualization, R.A.P., D.S.H. and J.H.S.; writing—original draft, R.A.P., D.S.H., J.H.S., C.J.B., J.F., C.A.M., K.L.M. and A.G.V.-F.; writing—review and editing, R.A.P., D.S.H., J.H.S., C.J.B., J.F., C.A.M., K.L.M., A.G.V.-F. and C.T.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are not being made publicly available, but can be made available to individuals upon request.

Acknowledgments

The authors would like to thank Ramazan Aygun and Minjae Woo, for general advice on the application of convolutional neural networks; Joe Sexton, for providing excellent information on the state of the art in Candida auris identification; the CDC AR bank and NRRL culture collection, for providing the isolates used; and Kyle Gabriel and the rest of the Bioinnovation lab, for providing and maintaining a supportive and productive research space.

Conflicts of Interest

The authors declare no conflicts of interest. No funding was provided and thus no funders had any role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Parker, R.A.; Gabriel, K.T.; Graham, K.; Butts, B.K.; Cornelison, C.T. Antifungal Activity of Select Essential Oils against Candida auris and Their Interactions with Antifungal Drugs. Pathogens 2022, 11, 821. [Google Scholar] [CrossRef] [PubMed]
  2. Satoh, K.; Makimura, K.; Hasumi, Y.; Nishiyama, Y.; Uchida, K.; Yamaguchi, H. Candida auris sp. nov., a novel ascomycetous yeast isolated from the external ear canal of an inpatient in a Japanese hospital. Microbiol. Immunol. 2009, 53, 41–44. [Google Scholar] [CrossRef] [PubMed]
  3. Kordalewska, M.; Perlin, D.S. Identification of Drug Resistant Candida auris. Front. Microbiol. 2019, 10, 1918. [Google Scholar] [CrossRef] [PubMed]
  4. Černáková, L.; Roudbary, M.; Brás, S.; Tafaj, S.; Rodrigues, C.F. Candida auris: A Quick Review on Identification, Current Treatments, and Challenges. Int. J. Mol. Sci. 2021, 22, 4470. [Google Scholar] [CrossRef] [PubMed]
  5. Jackson, B.R.; Chow, N.; Forsberg, K.; Litvintseva, A.P.; Lockhart, S.R.; Welsh, R.; Vallabhaneni, S.; Chiller, T. On the Origins of a Species: What Might Explain the Rise of Candida auris? J. Fungi 2019, 5, 58. [Google Scholar] [CrossRef] [PubMed]
  6. Fasciana, T.; Cortegiani, A.; Ippolito, M.; Giarratano, A.; Di Quattro, O.; Lipari, D.; Graceffa, D.; Giammanco, A. Candida auris: An Overview of How to Screen, Detect, Test and Control This Emerging Pathogen. Antibiotics 2020, 9, 778. [Google Scholar] [CrossRef] [PubMed]
  7. Li, Y.; Wu, Y.; Gao, Y.; Niu, X.; Li, J.; Tang, M.; Fu, C.; Qi, R.; Song, B.; Chen, H.; et al. Machine-learning based prediction of prognostic risk factors in patients with invasive candidiasis infection and bacterial bloodstream infection: A singled centered retrospective study. BMC Infect. Dis. 2022, 22, 150. [Google Scholar] [CrossRef] [PubMed]
  8. Wu, T.T.; Xiao, J.; Sohn, M.B.; Fiscella, K.A.; Gilbert, C.; Grier, A.; Gill, A.L.; Gill, S.R. Machine learning approach identified Multi-Platform Factors for caries prediction in Child-Mother dyads. Front. Cell. Infect. Microbiol. 2021, 11, 727630. [Google Scholar] [CrossRef] [PubMed]
  9. CDC. Identification of Candida auris|Candida auris|Fungal Diseases. Available online: https://www.cdc.gov/fungal/candida-auris/identification.html (accessed on 28 August 2024).
  10. Fernández-Manteca, M.G.; Ocampo-Sosa, A.A.; Ruiz de Alegría-Puig, C.; Pía Roiz, M.; Rodríguez-Grande, J.; Madrazo, F.; Calvo, J.; Rodríguez-Cobo, L.; López-Higuera, J.M.; Fariñas, M.C.; et al. Automatic classification of Candida species using Raman spectroscopy and machine learning. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2023, 290, 122270. [Google Scholar] [CrossRef]
  11. Jeffery-Smith, A.; Taori, S.; Schelenz, S.; Jeffery, K.; Johnson, E.; Borman, A.M.; Candida auris Incident Management Team; Manuel, R.; Brown, C.S. Candida auris: A Review of the Literature. Clin. Microbiol. Rev. 2017, 31, e00029-17. [Google Scholar] [CrossRef] [PubMed]
  12. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, A.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  13. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images. University of Toronto. 2009. Available online: https://www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf (accessed on 9 September 2023).
  14. Shaikh, J. Essentials of Deep Learning: Visualizing Convolutional Neural Networks in Python. Analytics Vidhya. 2018. Available online: https://www.analyticsvidhya.com/blog/2018/03/essentials-of-deep-learning-visualizing-convolutional-neural-networks (accessed on 1 April 2024).
  15. Lee, H.; Grosse, R.; Ranganath, R.; Ng, A.Y. Unsupervised learning of hierarchical representations with convolutional deep belief networks. Commun. ACM 2011, 54, 95–103. [Google Scholar] [CrossRef]
  16. CDC. How the AR Isolate Bank Helps Combat Antibiotic Resistance. Centers for Disease Control and Prevention. 2020. Available online: https://www.cdc.gov/drugresistance/resistance-bank/index.html (accessed on 8 August 2023).
  17. Welcome to the ARS Culture Collection (NRRL). ARS Culture Collection (NRRL). 2022. Available online: https://nrrl.ncaur.usda.gov/ (accessed on 8 August 2023).
  18. Li, L.; Jamieson, K.; DeSalvo, G.; Rostamizadeh, A.; Talwalkar, A. Hyperband: A novel Bandit-Based approach to hyperparameter optimization. arXiv 2016, arXiv:1603.06560. [Google Scholar]
  19. Simonyan, K.; Zisserman, A. Very deep convolutional networks for Large-Scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  20. Howard, A.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, A. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  21. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; Li, F. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar] [CrossRef]
  22. Fukushima, K. Cognitron: A self-organizing multilayered neural network. Biol. Cybern. 1975, 20, 121–136. [Google Scholar] [CrossRef] [PubMed]
  23. Džakula, N.B.; Bezdan, T. Convolutional Neural Network Layers and Architectures. In Proceedings of the Sinteza 2019—International Scientific Conference on Information Technology and Data Related Research, Novi Sad, Serbia, 20 April 2019. [Google Scholar] [CrossRef]
  24. Papers with Code—ReLU Explained. Available online: https://paperswithcode.com/method/relu (accessed on 28 August 2024).
  25. Howard, A.; Pang, R.; Adam, H.; Le, Q.V.; Sandler, M.; Chen, B.; Wang, W.; Chen, L.-C.; Tan, M.; Chu, G.; et al. Searching for MobileNetV3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar] [CrossRef]
  26. Papers with Code—Hard Swish Explained. Available online: https://paperswithcode.com/method/hard-swish (accessed on 28 August 2024).
  27. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar] [CrossRef]
Figure 1. An example representation of a convolutional neural network, adapted from [14,15]. A car is used as an example, as features of a car are more intuitive than those of yeast cells would be.
Figure 1. An example representation of a convolutional neural network, adapted from [14,15]. A car is used as an example, as features of a car are more intuitive than those of yeast cells would be.
Pathogens 14 00504 g001
Figure 2. Image preprocessing step, uniformly cutting source image (S. cerevisiae in this example) into 66 patches, with potentially blank patches to be removed by a blank detection algorithm that used standard deviation thresholding (highlighted in red).
Figure 2. Image preprocessing step, uniformly cutting source image (S. cerevisiae in this example) into 66 patches, with potentially blank patches to be removed by a blank detection algorithm that used standard deviation thresholding (highlighted in red).
Pathogens 14 00504 g002
Figure 3. Sample patches of each species, representing some of the variation in appearance (clustering, budding, etc.) captured between patches of the same species. Note that each of these patches was correctly identified.
Figure 3. Sample patches of each species, representing some of the variation in appearance (clustering, budding, etc.) captured between patches of the same species. Note that each of these patches was correctly identified.
Pathogens 14 00504 g003
Figure 4. A visual representation of the VGG16 base model. Adapted from [23].
Figure 4. A visual representation of the VGG16 base model. Adapted from [23].
Pathogens 14 00504 g004
Figure 5. Activation functions used in neural networks. (a) The Rectified Linear Unit. (b) The non-parametric Hard Swish Function. Taken from [25,27].
Figure 5. Activation functions used in neural networks. (a) The Rectified Linear Unit. (b) The non-parametric Hard Swish Function. Taken from [25,27].
Pathogens 14 00504 g005
Figure 6. Pipeline of pre-processing and CNN model flow chart. The process begins with microscopy image acquisition, followed by patch extraction and filtering of low-content regions. Patched containing cells are passed through a convolutional neural network (CNN) for classification. Final species prediction is made via a hard-voting scheme across all patches in the image. This pipeline supports rapid, low-cost fungal diagnostics based on microscope images.
Figure 6. Pipeline of pre-processing and CNN model flow chart. The process begins with microscopy image acquisition, followed by patch extraction and filtering of low-content regions. Patched containing cells are passed through a convolutional neural network (CNN) for classification. Final species prediction is made via a hard-voting scheme across all patches in the image. This pipeline supports rapid, low-cost fungal diagnostics based on microscope images.
Pathogens 14 00504 g006
Figure 7. The accuracy curves of the VGG16-based model over the course of training. Each Epoch represents a complete pass over the dataset. Note that the periodic dips in accuracy occur when an additional layer of the base model is unlocked for training.
Figure 7. The accuracy curves of the VGG16-based model over the course of training. Each Epoch represents a complete pass over the dataset. Note that the periodic dips in accuracy occur when an additional layer of the base model is unlocked for training.
Pathogens 14 00504 g007
Figure 8. VGG16-based model feature maps which show the intensity of each channel in the RGB specification to show the detail contributed by each channel. Note that the darker a spot, the more intensely that specific pixel contributed to the learning of the model. Microscopy images acquired at 40× magnification, patches are 224 × 224 pixels.
Figure 8. VGG16-based model feature maps which show the intensity of each channel in the RGB specification to show the detail contributed by each channel. Note that the darker a spot, the more intensely that specific pixel contributed to the learning of the model. Microscopy images acquired at 40× magnification, patches are 224 × 224 pixels.
Pathogens 14 00504 g008
Figure 9. Examples of images with cell debris. Top row, left-to-right, (a): C. albicans, (b): C. auris, (c): C. glabrata. Bottom row, left-to-right, (d): C. haemulonii, (e): C. krusei, (f): S. cerevisiae. Microscopy images acquired at 40× magnification, patches are 224 × 224 pixels.
Figure 9. Examples of images with cell debris. Top row, left-to-right, (a): C. albicans, (b): C. auris, (c): C. glabrata. Bottom row, left-to-right, (d): C. haemulonii, (e): C. krusei, (f): S. cerevisiae. Microscopy images acquired at 40× magnification, patches are 224 × 224 pixels.
Pathogens 14 00504 g009
Table 1. Misdiagnosis of Candida auris using different identification platforms. Information current as of 27 June 2024, from CDC.gov.
Table 1. Misdiagnosis of Candida auris using different identification platforms. Information current as of 27 June 2024, from CDC.gov.
Identification PlatformDiagnosis of Candida auris
API 20CRhodotorula glutinis
Candida sake
API ID 32CCandida intermedia
Candida sake
Saccharomyces kluyveri
BD Phoenix Yeast Identification SystemCandida haemulonii
Candida catenulata
MicroScanCandida famata
Candida lusitaniae
Candida guilliermondii
Candida parapsilosis
RapID Yeast PlusCandida parapsilosis
Vitek 2 YSTCandida haemulonii
Candida duobushaemulonii
Vitek MS MALDI-TOF (with older libraries)Candida lusitaniae
Candida haemulonii
Adapted from [9,11].
Table 2. The architecture of VGG16 base model, VGG16, used as the base model for the VGG16-based network.
Table 2. The architecture of VGG16 base model, VGG16, used as the base model for the VGG16-based network.
LayerOutput ShapeParam #
InputLayer(None, 224, 224, 3)0
Conv2D(None, 224, 224, 64)1792
Conv2D(None, 224, 224, 64)36,928
MaxPooling2D(None, 112, 112, 64)0
Conv2D(None, 112, 112, 128)73,856
Conv2D(None, 112, 112, 128)147,584
MaxPooling2D(None, 56, 56, 128)0
Conv2D(None, 56, 56, 256)295,168
Conv2D(None, 56, 56, 256)590,080
Conv2D(None, 56, 56, 256)590,080
MaxPooling2D(None, 28, 28, 256)0
Conv2D(None, 28, 28, 512)1,180,160
Conv2D(None, 28, 28, 512)2,359,808
Conv2D(None, 28, 28, 512)2,359,808
MaxPooling2D(None, 14, 14, 512)0
Conv2D(None, 14, 14, 512)0
Conv2D(None, 14, 14, 512)2,359,808
Conv2D(None, 14, 14, 512)2,359,808
MaxPooling2D(None, 7, 7, 512)0
Table 3. The VGG16-based model architecture.
Table 3. The VGG16-based model architecture.
LayerOutput ShapeParam #
VGG16 Base Model(None, 7, 7, 512)14,714,688
Flatten(None, 25088)0
Dropout (50%)(None, 25088)0
Dense (256, PH-Swish)(None, 256)6,422,784
Dense (256, ReLU)(None, 256)65,792
Dropout (50%)(None, 256)0
Dense (128, PH-Swish)(None, 128)32,896
Output (6, Softmax)(None, 6)774
Table 4. Accuracy of our convolutional neural network models to predict the identity of select Candida species and Saccharomyces cerevisiae from the test set of micrograph image patches, as well as the complete pipeline on whole test images.
Table 4. Accuracy of our convolutional neural network models to predict the identity of select Candida species and Saccharomyces cerevisiae from the test set of micrograph image patches, as well as the complete pipeline on whole test images.
ModelCandida
albicans
Candida aurisCandida glabrataCandida
haemulonii
Candida kruseiSaccharomyces
cerevisiae
Overall
Hyperband CNN0.88660.83800.823640.84420.92490.88050.8652
VGG16-Based CNN0.95440.92000.95290.91670.94280.94820.9391
MobileNet-Based CNN0.71730.69310.65720.73860.80930.78650.7337
Completed1.00001.00001.00000.99140.96360.98250.9909
Pipeline
(Whole Images)
Table 5. Confusion matrix for VGG16-based model with learned weights on image patches.
Table 5. Confusion matrix for VGG16-based model with learned weights on image patches.
Predicted
Actual
Candida
albicans
Candida aurisCandida
glabrata
Candida
haemulonii
Candida kruseiSaccharomyces cerevisiae
Candida albicans2447748174693
Candida auris125531570172
Candida glabrata11512466115312
Candida haemulonii10138472431248
Candida krusei1124211258921
Saccharomyces cerevisiae842108672487
Table 6. Confusion matrix for integrated pipeline on whole images.
Table 6. Confusion matrix for integrated pipeline on whole images.
Predicted
Actual
Candida albicansCandida
auris
Candida glabrataCandida
haemulonii
Candida kruseiSaccharomyces cerevisiae
Candida albicans10100001
Candida auris0690000
Candida glabrata0042100
Candida haemulonii00011500
Candida krusei0000530
Saccharomyces cerevisiae0000256
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Parker, R.A.; Hannagan, D.S.; Strydom, J.H.; Boon, C.J.; Fussell, J.; Mitchell, C.A.; Moerschel, K.L.; Valter-Franco, A.G.; Cornelison, C.T. A Complete Transfer Learning-Based Pipeline for Discriminating Between Select Pathogenic Yeasts from Microscopy Photographs. Pathogens 2025, 14, 504. https://doi.org/10.3390/pathogens14050504

AMA Style

Parker RA, Hannagan DS, Strydom JH, Boon CJ, Fussell J, Mitchell CA, Moerschel KL, Valter-Franco AG, Cornelison CT. A Complete Transfer Learning-Based Pipeline for Discriminating Between Select Pathogenic Yeasts from Microscopy Photographs. Pathogens. 2025; 14(5):504. https://doi.org/10.3390/pathogens14050504

Chicago/Turabian Style

Parker, Ryan A., Danielle S. Hannagan, Jan H. Strydom, Christopher J. Boon, Jessica Fussell, Chelbie A. Mitchell, Katie L. Moerschel, Aura G. Valter-Franco, and Christopher T. Cornelison. 2025. "A Complete Transfer Learning-Based Pipeline for Discriminating Between Select Pathogenic Yeasts from Microscopy Photographs" Pathogens 14, no. 5: 504. https://doi.org/10.3390/pathogens14050504

APA Style

Parker, R. A., Hannagan, D. S., Strydom, J. H., Boon, C. J., Fussell, J., Mitchell, C. A., Moerschel, K. L., Valter-Franco, A. G., & Cornelison, C. T. (2025). A Complete Transfer Learning-Based Pipeline for Discriminating Between Select Pathogenic Yeasts from Microscopy Photographs. Pathogens, 14(5), 504. https://doi.org/10.3390/pathogens14050504

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop