Next Article in Journal
Hashimoto’s Thyroiditis and Female Fertility: Endocrine, Immune, and Microbiota Perspectives in Assisted Reproduction—A Narrative Review
Previous Article in Journal
Serum NT-ProBNP/Chloride Ratio Predicts Adverse Cardiovascular Outcomes in Patients with Acute Heart Failure
Previous Article in Special Issue
Machine Learning Models in Sepsis Outcome Prediction for ICU Patients: Integrating Routine Laboratory Tests—A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Auto Machine Learning and Convolutional Neural Network in Diabetes Mellitus Research—The Role of Histopathological Images in Designing and Exploring Experimental Models

by
Iulian Tătaru
1,
Simona Moldovanu
2,3,*,
Oana-Maria Dragostin
4,*,
Carmen Lidia Chiţescu
4,
Alexandra-Simona Zamfir
5,
Ionut Dragostin
4,
Liliana Strat
6 and
Carmen Lăcrămioara Zamfir
1
1
Department of Morphofunctional Sciences I, “Grigore T. Popa” University of Medicine and Pharmacy, 700115 Iasi, Romania
2
Department of Computer Science and Information Technology, Faculty of Automation, Computers, Electrical Engineering and Electronics, “Dunarea de Jos” University of Galati, 800146 Galati, Romania
3
The Modelling & Simulation Laboratory, “Dunarea de Jos” University of Galati, 47 Domneasca Str., 800008 Galati, Romania
4
Research Centre in the Medical-Pharmaceutical Field, Faculty of Medicine and Pharmacy, “Dunarea de Jos” University of Galati, 800201 Galati, Romania
5
Department of Medical Sciences III, “Grigore T. Popa” University of Medicine and Pharmacy, 700115 Iasi, Romania
6
Department of Mother & Child, “Grigore T. Popa” University of Medicine and Pharmacy, 700115 Iasi, Romania
*
Authors to whom correspondence should be addressed.
Biomedicines 2025, 13(6), 1494; https://doi.org/10.3390/biomedicines13061494
Submission received: 21 March 2025 / Revised: 7 June 2025 / Accepted: 12 June 2025 / Published: 18 June 2025
(This article belongs to the Special Issue Artificial Intelligence Applications in Cancer and Other Diseases)

Abstract

:
Histopathological images represent a valuable data source for pathologists, who can provide clinicians with essential landmarks for complex pathologies. The development of sophisticated computational models for histopathological images has received significant attention in recent years, but most of them rely on free datasets. Materials and Methods: Motivated by this drawback, the authors created an original histopathological image dataset that resulted from an animal experimental model, acquiring images from normal female rats/rats with experimentally induced diabetes mellitus (DM)/rats who received an antidiabetic therapy with a synthetic compound (AD_SC). Images were acquired from vaginal, uterine, and ovarian samples from both MD and AD_DC specimens. The experiment received the approval of the Medical Ethics Committee of the “Gr. T. Popa” University of Medicine and Pharmacy, Iași, Romania (Approval No. 169/22.03.2022). The novelty of the study consists of the following aspects. The first is the use of a diabetes-induced animal model to evaluate the impact of an antidiabetic therapy with a synthetic compound in female rats, focusing on three distinct organs of the reproductive system (vagina, ovary, and uterus), to provide a more comprehensive understanding of how diabetes affects female reproductive health as a whole. The second comprises image classification with a custom-built convolutional neural network (CB-CNN), the extraction of textural features (contrast, entropy, energy, and homogeneity), and their classification with PyCaret Auto Machine Learning (AutoML). Results: Experimental findings indicate that uterine tissue, both for MD and AD_DC, can be diagnosed with an accuracy of 94.5% and 85.8%, respectively. The Linear Discriminant Analysis (LDA) classifier features indicate a high accuracy of 86.3% when supplied with features extracted from vaginal tissue. Conclusions: Our research underscores the efficacy of classifying with two AI algorithms, CNN and machine learning.

1. Introduction

The growing prevalence of diabetes mellitus represents a medical problem of maximum interest [1]. All the renal, vascular, cardiac, ophthalmic, and nervous diabetes mellitus complications are well known and widely debated. At the same time, however, there are more and more discussions about the relevance of gender in diabetes mellitus, under the conditions in which, especially in women, diabetes generates diverse sexual dysfunctions and reproductive problems [2]. How much female reproductive function can be affected by diabetes and what mechanisms are involved are just two of the questions that arouse interest today. In the same context, the discovery of novel therapeutic options for diabetes mellitus supports the interest in developing this topic. The application of advanced computational techniques, particularly deep learning (DL) and Machine Learning (ML), to improve the classification accuracy of genital organs has gained significant popularity and attention in the last decade. DL and ML are two subsets of artificial intelligence (AI), with excellent applicability in various medical imaging tasks, such as tumor segmentation or categorization, anatomical contour delineation, therapy outcome prediction, feature extraction, and classification. DL models applied in gynecological pathologies can enhance clinicians’ ability to categorize and diagnose different cases with increased reliability and consistency using in parallel the features extracted automatically from medical images of various image types, mainly histopathology images [3].
DL and ML have attracted more and more attention in recent years for their ability to improve diagnostic accuracy and clinical decision-making across a wide range of medical imaging applications.
In new research, the authors looked at how well AI algorithms could classify histopathological images of the uterus [4,5,6,7,8,9,10,11,12,13], the vagina [14,15,16,17], and the ovaries [18,19,20,21,22,23,24]. Many different types of CNNs were used, including pre-trained ones [4,5,7,12,13,16,19,20,21,22,23,24], custom-built CNNs [8,9,10,14], CNNs and ML and DNNs for classifying morphological and geometric features [14,15], and a mix of a CNN and ML with classifier algorithms replacing the Softmax function [13,17,18].
This research emphasizes the growing prevalence and viability of DL and ML in enhancing the classification accuracy of female reproductive organs. The novelty consists of studying three organs from the female reproductive system (the vagina, ovaries, and uterus). The histopathological images were analyzed with texture features (contrast, entropy, energy, and homogeneity) extracted from the co-occurrence matrix [25]; these features represent the input into PyCaret AutoML [26,27]. Using open-source AutoML tools ensures the employment of 15 ML classifiers and the hyperparameter fine-tuning process. Additionally, custom-built CNNs were applied in the classification process. Although the authors [4,5,7,12,13,16,19,20,21,22,23,24] reported the best results in classifying histopathological images using pre-trained CNNs, the classification accuracy was very low on the proposed dataset.
The following parts of this work are organized as follows. Section 2 highlights the published literature, identifies the research gap, and outlines the authors’ contributions. Section 3 outlines the materials and methods employed, encompassing the hardware, software, dataset, and AI algorithms utilized. Section 4 presents the results, assessment, and comparison to state-of-the-art methods. Section 5 closes the study with suggested insights for further research.

2. Literature Review

Considered not only a major global public health problem, but also a specific and very complex pathological condition, diabetes mellitus is far from being just a metabolic disorder whose defining aspects have all been identified, as there are still aspects yet to be explored.
The consequences of diabetes mellitus on female reproductive function have often been underestimated, as its impact is not perceived to be as severe as that on renal, neurological, ophthalmic, and cardiovascular health [28,29]. Diabetes mellitus causes numerous cellular and molecular changes in the components of the female reproductive system. Once it develops, it triggers the onset of a complex and often silent pathology, especially in comparison with other affected systems, but leads to gradual and significant impairment of reproductive function [30].
Glucose homeostasis and insulin secretion require precise and highly refined regulatory mechanisms [29]. Many factors have the potential to affect these mechanisms, generating marked gender differences in the evolution and complications of diabetes. Recent data claim that the long-term effects of diabetes mellitus on women of reproductive age and menopausal and postmenopausal women are of major significance. The disturbances in these regulatory mechanisms underlie female reproductive dysfunctions; however, due to their variability, providing a comprehensive explanation for the majority of these issues remains difficult [31,32].
Diabetic women are at an increased risk of vaginal and recurrent urinary tract infections, infertility, menstrual cycle impairment, sexual dysfunctions, and severe pregnancy-related complications (pregestational or gestational diabetes affecting the maternal and embryonic organism) [33]. Polycystic ovary syndrome (PCOS), due to its connection to insulin resistance, may also generate dysfunctions in female reproductive function. All of these aspects require a very thorough exploration that takes into account the structural complexity of the female genital system and the multiple and overwhelming hormonal interferences.
Changes in diabetes-induced vaginal, uterine, and ovarian morphology may be explained up to a point, but the mechanisms involved require further investigation. Diabetes induces changes in the normal morphology of female reproductive organs, closely correlated with a sharp decrease in their functions. Advanced vaginal fibrosis, follicular degeneration, an increase in the number of atretic ovarian follicles, and uterine mucosa atrophy can reduce or even compromise female reproductive functions.
In vivo experimental models are utilized to gain a deeper understanding of these processes, offering valuable insights when extrapolated to humans [34]. With proper design, these animal models can provide numerous data related to morphological changes in the female reproductive system, the underlying mechanisms that support it, and new targeted therapies [35].

2.1. Research Gap and Contribution

Despite the successful application of DL in the classification of histopathological images, there is an urgent requirement for strong and reliable systems particularly designed to classify the gynecologic tissues in the case of diabetes mellitus. Current research predominantly focuses on the features of individual DL models or pre-trained CNNs, which are time-consuming, may fail to encompass the full complexity of medical images, and may not fulfill the requisite standards of accuracy and dependability for clinical applications.
By integrating two AI algorithms, an optimized CNN model and AutoML, the classification of images into healthy and gynecological pathologies significantly improved the accuracy that was obtained and became complete.
Table 1 presents state-of-the-art studies that employed DL or ML models for the classification of genital pathology. The studies were analyzed in concordance with the type of tissue analyzed, their weakness, and the AI algorithm used. For a reliable comparison, we only used studies where histopathological images were selected.
The common weakness of studies in the scientific literature is that these deal with only one tissue type (uterine [2,3,4,5,6,7,8,9,10,11,12,13], vaginal [14,15,16,17], or ovarian [18,19,20,21,22,23,24]). Only Onishi et al. [13] proposed to analyze vaginal and uterine tissue. The majority of studies used DNN algorithms, and the extraction of features from histopathological images was proposed by Volinsky-Fremond et al. [14]. When the use of morphological and genomic features was proposed, Jeleń et al. [15] used geometric features that fed a neural network (NN) and a support vector machine (SVM). A recent study by Zafar et al. [11] highly correlated 65 features in a classification process with an NN.
The hybrid models that used pre-trained CNNs and MLs were depicted by Rajan et al. [17], where the Softmax function was replaced with extreme gradient boost (XGBoost), SVM, k-nearest neighbor (KNN), InceptionV3, and residual networks (ResNet). Zhou et al. [18] proposed SVM, random forest (RF), XGBoost, and a custom-built CNN. Wang et al. [20] focused on the pre-trained CNN InceptionV3 and Vision Transformer (ViT).

2.2. Novelty and Contribution

By combining the advantages of custom-built architectures and PyCaret AutoML, our study fills a research gap because the time-consuming, low-classification-accuracy use of only an AI algorithm can be avoided. The contributions of this study are as follows:
1.
The acquisition of four tissue types (healthy, uterine, vaginal, and ovarian) and the creation of a histopathological image dataset for both MD and AD_DC specimens;
2.
Building, optimizing, and training a custom-built CNN with four layers;
3.
Extracting second-order features (contrast, entropy, energy, and homogeneity) from the co-occurrence matrix;
4.
Training the end-to-end PyCaret AutoML algorithm with correlation, energy, contrast, and homogeneity textural features;
5.
A performance evaluation of both the CNN and AutoML in terms of accuracy, F1-score, and area under the curve (AUC).

3. Materials and Methods

The experiments described in this article are part of a larger experiment, which was extensively detailed in a separate article [36] (https://doi.org/10.3390/biomedicines13040922).

3.1. Microscope

Samples from all female rats (n = 7/group) from the control, (untreated)/induced, untreated diabetes mellitus, and induced diabetes mellitus treated with a synthetic compound groups were collected, fixed in 4% formaldehyde solution, embedded in paraffin, cut into 2 μm sections, stained (hematoxylin–eosin staining, H&E), and examined with a Nikon Eclipse 50i microscope with Plan 10×, 40×, and 100× oil objectives. Images were captured and analyzed using the Nikon Digital Sight Ds-Fi1 high resolution digital camera (Nikon Instruments Inc., Tokyo, Japan) attached to the ocular tube.

3.2. Hardware and Software

All AI algorithms and feature extractions were run on a PC with Apple Mac Studio (2022), Apple M1 Max, 32 GB, a 1 TB SSD, and a 32-core GPU, Apple, Cupertino, California, USA.
The training and testing of AI algorithms in the software environment utilized Python (3.11) for PyCaret AutoML (3.0.4) and Python (3.12) for the CNN along with the following libraries: scikit-learn (1.5.2), TensorFlow (2.18.0), PyCaret (2.3.10), Keras (3.6.0), and Visualkeras (0.1.4.3).
The second-order features were extracted by using MATLAB 2021a (The MathWorks, Natick, MA, USA) and Image Processing Library.

3.3. Histopathological Image Dataset and Augmentation

The size of the acquired images was 1280 × 960 (width and height) in JPEG format.
To enhance the diversity of the training data and prevent overfitting, we employed three data augmentation strategies:
(i)
Random rotations and affine transformations. The images were manipulated with random rotations of up to 10 degrees, and affine transformations were employed to implement scaling and translations, in order to enhance the model’s robustness to spatial transformation;
(ii)
Random zoom. This creates surrounding pixel values in the image or interpolates pixel values according to the established percent; in this case, 30% was proposed;
(iii)
Random horizontal and vertical flips. These were applied with probabilities of 30% for the horizontal orientation [37].

3.4. PyCaret AutoML

A machine learning algorithm is a method that processes data to identify a specific hypothesis from a set of candidates, optimizing performance in accurately representing the target function based on training experience. Numerous machine learning algorithms have been developed for various purposes; they differ in how they represent the hypothesis set, the approach to hypothesis selection, and the interaction of the model with the training data [38].
All textural features were imported into the PyCaret tool. It is an open-source, low-code machine learning framework in Python that optimizes specific parameters. PyCaret is frequently utilized in classifying features [39,40] due to its provision of a simplified and efficient process for comprehensive machine learning. It automates numerous activities, including data pretreatment, model selection, feature engineering, and hyperparameter tuning, therefore diminishing the necessity for manual intervention. Fivefold cross-validation enabled the optimization of the five highest-performing models selected from the following ML algorithms: KNN, Extra Trees Classifier (ETC), Logistic Regression, Ridge Classifier, RF, Classifier Light, Ada Boost Classifier, SVM, Linear Kernel Gradient, Naive Bayes, Decision Tree Classifier, Quadratic Discriminant Analysis, Dummy Classifier, Gradient Boosting Machine, Linear Discriminant Analysis (LDA), and Boosting Classifier. Subsequently, we created a random partition of the dataset (80% allocated for training and validation and 20% designated for testing) [41].

3.5. Custom-Built CNN

A class of machine learning models known as “deep learning” models use multilayer representations to extract hierarchical features from complex inputs. CNNs, recurrent neural networks (RNNs), generative adversarial networks (GANs), and deep belief networks are examples of common deep learning models that are based on neural network architectures [38].
This endeavor aimed to accurately analyze gynecologic histopathological images through the use of custom-built models. With an ablation process, the hyperparameters of the CNN were established. The image size was between 50 × 50 and 200 × 200, with a batch size of 4, 8, and 16, and epochs of 10 to 50. The optimum values for hyperparameters when a higher accuracy was obtained were an image size of 180 × 180 pixels, a batch size of 4, and 20 epochs. The customized CNN comprised four convolutional blocks, and the parameters of each layer are shown in Figure 1. Conv2D comprises a convolutional layer succeeded by batch normalization to enhance learning stability by normalizing the input to each layer. The ReLU activation function was utilized to incorporate nonlinearity. MaxPooling layers were incorporated following each convolutional block to diminish the spatial dimensions by fifty percent, preserving just the most salient information. Flattening was employed to transform all resultant two-dimensional arrays from pooled feature maps into a singular, extended linear vector. A dense layer was identified by its neurons being interconnected with every neuron in the preceding layer.

3.6. Co-Occurrence Matrix

A Gray-Level Co-occurrence Matrix (GLCM) provided an excellent foundation for image analysis and has proved to be important in many classification tasks. The contrast, entropy, energy, and homogeneity features are popular approaches to investigating image textures. The two most essential GLCM components are direction and distance. The horizontal, right, vertical, and left diagonal directions are the four possible direction angles for a two-dimensional image (0°, 45°, 90°, and 135°) [15]. The distance value can be adjusted to any larger number in order to determine the correlation between distant pixels, or it can be set to 1 for the pixel that is closest to its neighbour. Given that N is the total number of gray levels and is the component at the location in the normalized symmetrical GLCM, the following features are given [42].
The contrast measures the disparity or Intensity difference between a pixel and Its surrounding area over the full image.
C o n t r a s t = i , j = 0 N 1 S i j i j 2
The degree to which the distribution of GLCM elements closely resembles the GLCM diagonal is known as homogeneity.
H o m o g e n e i t y = i , j = 0 N 1 S i j 1 + i j 2
Energy is calculated by summing squared components.
E n e r g y = i , j = 0 N 1 S i j 2
Entropy is a measure of an image’s randomness; it is low for smooth images and high for non-texturally uniform images [30].
E n t o p y = i , j = 0 N 1 S i j l o g S i j
where S i j is the probability of the value pair i and j.

3.7. Performance Evaluation Metrics

In our experiment, we employed an AutoML model with fifteen classifiers and a custom-built CNN. To accurately evaluate the proposed model, we employed a 5-fold cross-validation method. The confusion matrix produced in a classification process is a 2 × 2 matrix that indicates the model’s performance as a classifier for the test set with known actual target labels. The outcomes of the suggested forgery detection are presented in the confusion matrix as true positives (TPs), true negatives (TNs), false positives (FPs), and false negatives (FNs).
The accuracy is the proportion of accurately predicted occurrences to the total number of instances in the dataset, and the F1-score is a measure of predictive performance. These principal metrics are given by Equations (5)–(7) [43].
Accuracy = (TP + TN)/(TP + TN + FP + FN)
F1-score = (2∙TP)/(2∙TP + FP + FN)
AUC = 1 1 / 2 FP / FP + TN + FN / FN + TP

4. Results and Discussion

Histopathological Image Dataset and Augmentation

The complex dataset after an augmentation process contained histopathological images of adult Wistar female rat (280 g weight) vaginal tissue (control (216), alloxan-induced diabetes mellitus (405), and diabetes mellitus and antidiabetic therapy with a synthetic compound (AD_SC) (603), uterine tissue (control (288), alloxan-induced diabetes mellitus (270), and diabetes mellitus and antidiabetic therapy with a synthetic compound (AD_SC) (612)), and ovary tissue (control (333), alloxan-induced diabetes mellitus (666), and diabetes mellitus and antidiabetic therapy with a synthetic compound (AD_SC) (360)). Samples from each class are shown in Figure 2. The augmented images are shown in Figure 3, after applying augmentation.
The data was randomly partitioned into 80% for training and validation and 20% for testing. We employed fivefold cross-validation on the training and validation sets (80%) using AutoML PyCaret for classification of textural features for the control and DM vagina, healthy and DM uterus, and control and DM ovary histopathological image classes. The experiment was repeated for the control and AD_DC vaginal, healthy and AD_DC uterine, and healthy and AD_DC ovarian histopathological image classes. Using the same number of classes, and keeping the same percentages for the training and testing sets, the custom-built CNN was trained. Hyperparameters were carefully optimized during training to guarantee the selection of the most effective model for the studied classes. Table 2 shows the results obtained with the proposed CNN. The performance of the model is expressed in terms of accuracy and F1-score. In addition, the confusion matrices are presented in the last column. Table 3 shows a detailed comparison of the performance of the top machine learning models fed with textural features via fivefold cross-validation on the test set, elucidating their strengths. In addition, the parameters of each ML classifier are shown. Following a comprehensive evaluation of AutoML models facilitated by the PyCaret tool, we discerned four models (ETC, RF, KNN, and LDA).
The noteworthy accuracy of 0.945 achieved by our custom-built model for uterine classes outshines that of the models proposed by Kitaya et al. (92.8%), Song et al. [8] (91.7%), Zhao et al. [9] (95.34%), Li et al. [10] (75%), and Volinsky-Fremond et al. [14] (78.9%). Also, the best accuracy was obtained in comparison with the pre-trained models proposed by Peng et al. [12] (86% AUC), Kodipalli [22] (67%), and Kwatra and Kaur [23] (92.8%).
Additionally, textural features were extracted from histopathological images in the diabetes mellitus and antidiabetic therapy with a synthetic compound class for the uterine, ovarian, and vaginal organs. We verified the results obtained with the PyCaret tool. The results demonstrate high performance for the antidiabetic therapy vagina class with an accuracy of 0.86 and the validation of the model with an F1-score of 0.88. In this case, Linear Discriminant Analysis was selected by AutoML as the best classifier.
Initially, the effectiveness of our approach may fluctuate when utilized on a large dataset, and AutoMLs are fed a large number of features. The extrapolation of our results to alternative datasets, especially with other diabetes classes, necessitates additional examination. Secondly, despite employing stringent augmentation methods to improve model robustness, the small number of rats from which the images were acquired and the existence of noise or artifacts in the input data may have adversely affected the model performance.
The limitations of this study include the high computational power required. In accordance with the hardware architecture, the time consumed exceeded two minutes. This was due to the use of CNNs and the limited number of histopathological images, as the dataset was originally based on a small number of rats. The use of pre-trained CNNs was not feasible due to the very low accuracy of the obtained results.

5. Conclusions and Future Scope

An important advancement in medical pathology is the use of digital or digitalized histopathology images for the identification of genital pathology for DS and AD_DC tissues. Additionally, because ML and CNN approaches may identify numerous unexplored regions, they have opened up new study prospects.
We consider that the high degree of interest in the research area related to diabetes mellitus and its effects on the female reproductive system supports more and more the use of digital or digitalized histopathologic images. Additionally, because ML and CNN approaches may identify numerous unexplored regions, they have opened up new study prospects.
This study introduced a novel model (a custom-built CNN for predicting three types of gynecological tissue (vaginal, uterine, and ovarian)) and, in parallel, classified textural features extracted from vaginal, uterine, and ovarian tissue using the PyCaret AutoML tool. In the first part of the study, the proposed CNN was found to provide high classification accuracy for uterine DS tissue as well as uterine AD_DC tissue. The second part of the study evaluated the classification performance of PyCaret AutoML; in this case, the best selected classifier was LDA for vaginal AD_DC tissue. Future studies should investigate supplementary classification algorithms or alternative preprocessing strategies to increase the accuracy of classification. Moreover, the authors propose to increase the number of images acquired in the laboratory to create a large dataset of histopathological images, classified by clinicians into subsets, as a valuable resource for researchers. This dataset could be used in designing experimental animal models, allowing for more extensive studies on the impact of diabetes mellitus on the female reproductive system.

Author Contributions

Conceptualization, S.M. and C.L.Z.; methodology, I.T. and S.M.; software, S.M.; validation, O.-M.D. and C.L.Z.; formal analysis, A.-S.Z., I.D. and L.S.; investigation, I.T., A.-S.Z. and I.D.; resources, S.M. and L.S.; data curation, C.L.Z. and S.M.; writing—original draft preparation, I.T. and C.L.C.; writing—review and editing, S.M. and C.L.Z.; visualization, O.-M.D. and C.L.C.; supervision, C.L.Z.; project administration, S.M.; funding acquisition, I.T. and C.L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The animal study protocol was approved by the Medical Ethics Committee of the “Grigore T. Popa” University of Medicine and Pharmacy, Iasi, Romania, (approval no. 169/22.03.2022).

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DMdiabetes mellitus
AD-SCantidiabetic therapy with a synthetic compound
CB-CNNcustom-built convolutional neural network
AutoMLPyCaret Auto Machine Learning
LDALinear Discriminant Analysis
DLdeep learning
MLmachine learning
AIartificial intelligence
PCOSpolycystic ovary syndrome
XGBoostextreme gradient boosting
KNNk-nearest neighbor
RESNETresidual networks
RFrandom forest

References

  1. Ciarambino, T.; Crispino, P.; Leto, G.; Mastrolorenzo, E.; Para, O.; Giordano, M. Influence of Gender in Diabetes Mellitus and Its Complications. Int. J. Mol. Sci. 2022, 23, 8850. [Google Scholar] [CrossRef] [PubMed]
  2. Wing, R.R.; Bond, D.S.; Gendrano, I.N., III; Wadden, T.; Bahnson, J.; Lewis, C.E.; Brancati, F.; Schneider, S.; Kitabchi, A.E.; Van Dorsten, B.; et al. Effect of intensive lifestyle intervention on sexual dysfunction in women with type 2 diabetes: Results from an ancillary LookAHEAD study. Diabetes Care 2013, 36, 2937–2944. [Google Scholar] [CrossRef]
  3. Ahsan, M.M.; Luna, S.A.; Siddique, Z. Machine-Learning-Based Disease Diagnosis: A Comprehensive Review. Healthcare 2022, 10, 541. [Google Scholar] [CrossRef]
  4. Patil, V.I.; Patil, S.R. Optimized Transfer Learning With Hybrid Feature Extraction for Uterine Tissue Classification Using Histopathological Images. Microsc. Res. Tech. 2025, 88, 1582–1598. [Google Scholar] [CrossRef]
  5. Sun, H.; Zeng, X.; Xu, T.; Peng, G.; Ma, Y. Computer-aided diagnosis in histopathological images of the endometrium using a convolutional neural network and attention mechanisms. IEEE J. Biomed. Health Inform. 2019, 24, 1664–1676. [Google Scholar] [CrossRef] [PubMed]
  6. Kitaya, K.; Yasuo, T.; Yamaguchi, T.; Morita, Y.; Hamazaki, A.; Murayama, S.; Mihara, T.; Mihara, M. Construction of Deep Learning-based Convolutional Neural Network Model for Automatic Detection of Fluid Hysteroscopic Endometrial Micropolyps in Infertile Women with Chronic Endometritis. Eur. J. Obstet. Gynecol. Reprod. Biol. 2024, 297, 249–253. [Google Scholar] [CrossRef] [PubMed]
  7. Asadpour, V.; Puttock, E.J.; Getahun, D.; Fassett, M.J.; Xie, F. Automated placental abruption identification using semantic segmentation, quantitative features, SVM, ensemble and multi-path CNN. Heliyon 2023, 9, e13577. [Google Scholar] [CrossRef]
  8. Song, J.; Im, S.; Lee, S.H.; Jang, H.J. Deep Learning-Based Classification of Uterine Cervical and Endometrial Cancer Subtypes from Whole-Slide Histopathology Images. Diagnostics 2022, 12, 2623. [Google Scholar] [CrossRef]
  9. Zhao, F.; Dong, D.; Du, H.; Guo, Y.; Su, X.; Wang, Z.; Xie, X.; Wang, M.; Zhang, H.; Cao, X.; et al. Diagnosis of Endometrium Hyperplasia and Screening of Endometrial Intraepithelial Neoplasia in Histopathological Images Using a Global-to-Local Multi-Scale Convolutional Neural Network. Comput. Methods Programs Biomed. 2022, 221, 106906. [Google Scholar] [CrossRef]
  10. Li, T.; Liao, R.; Chan, C.; Greenblatt, E.M. Deep learning analysis of endometrial histology as a promising tool to predict the chance of pregnancy after frozen embryo transfers. J. Assist. Reprod. Genet. 2023, 40, 901–910. [Google Scholar] [CrossRef]
  11. Zafar, M.M.; Javaid, N.; Shaheen, I.; Alrajeh, N.; Aslam, S. Enhancing clinical decision support with explainable deep learning framework for C-Section forecasting. Computing 2025, 107, 20. [Google Scholar] [CrossRef]
  12. Peng, S.; Huang, H.; Cheng, M.; Yang, Y.; Li, F. Efficiently recognition of vaginal micro-ecological environment based on Convolutional Neural Network. In Proceedings of the 2020 IEEE International Conference on E-Health Networking, Application & Services (HEALTHCOM), Shenzhen, China, 1–2 March 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar]
  13. Onishi, S.; Egami, R.; Nakamura, Y.; Nagashima, Y.; Nishihara, K.; Matsuo, S.; Murai, A.; Hayashi, S.; Uesumi, Y.; Kato, A.; et al. Digital workflows for pathological assessment of rat estrous cycle stage using images of uterine horn and vaginal tissue. J. Pathol. Inform. 2022, 13, 100120. [Google Scholar] [CrossRef] [PubMed]
  14. Volinsky-Fremond, S.; Horeweg, N.; Andani, S.; Barkey Wolf, J.; Lafarge, M.W.; de Kroon, C.D.; Ørtoft, G.; Høgdall, E.; Dijkstra, J.; Jobsen, J.J.; et al. Prediction of recurrence risk in endometrial cancer with multimodal deep learning. Nat. Med. 2024, 30, 1962–1973. [Google Scholar] [CrossRef]
  15. Jeleń, U.; Stankiewicz-Antosz, I.; Chosia, M.; Jeleń, M. Cervical Cancer Diagnosis with Feature Selection and Deep Learning. Appl. Sci. 2025, 15, 1458. [Google Scholar] [CrossRef]
  16. Singh, J.; Kaliyar, R.K.; Kumari, R.; Sharma, N.; Manjunath, S.; Dharani Prasad, P. Advancements in Early Detection of Cervical Cancer using Machine Learning and Deep Learning Models for Cervicography Analysis. In Proceedings of the International Conference on Emerging Innovations and Advanced Computing (INNOCOMP), Sonipat, India, 25–26 May 2024. [Google Scholar]
  17. Rajan, B.K.; Harshan, M.H.; Swaminathan, R. Augmenting interpretation of vaginoscopy observations in cycling bitches with deep learning model. BMC Vet. Res. 2024, 20, 401. [Google Scholar] [CrossRef]
  18. Zhou, R.; Zhao, B.; Ding, H.; Fu, Y.; Li, H.; Wei, Y.; Xie, J.; Chen, C.; Yin, F.; Huang, D. Survival Prediction of Ovarian Serous Carcinoma Based on Machine Learning Combined with Pathological Images and Clinical Information. AIP Adv. 2024, 14, 045324. [Google Scholar] [CrossRef]
  19. El-Latif, E.I.A.; El-Dosuky, M.; Darwish, A.; Hassanien, A.E. A deep learning approach for ovarian cancer detection and classification based on fuzzy deep learning. Sci. Rep. 2024, 14, 26463. [Google Scholar] [CrossRef]
  20. Wang, C.W.; Lee, Y.C.; Chang, C.C.; Lin, Y.J.; Liou, Y.A.; Hsu, P.C.; Chang, C.C.; Sai, A.K.O.; Wang, C.H.; Chao, T.K. A weakly supervised deep learning method for guiding ovarian cancer treatment and identifying an effective biomarker. Cancers 2022, 14, 1651. [Google Scholar] [CrossRef] [PubMed]
  21. Behera, S.K.; Das, A.; Sethy, P.K. Deep fine-KNN classification of ovarian cancer subtypes using efficientNet-B0 extracted features: A comprehensive analysis. J. Cancer Res. Clin. Oncol. 2024, 150, 361. [Google Scholar] [CrossRef]
  22. Kodipalli, A.; Guha, S.; Dasar, S.; Ismail, T. An inception-ResNet deep learning approach to classify tumours in the ovary as benign and malignant. Expert Syst. 2022, e13215. [Google Scholar] [CrossRef]
  23. Kwatra, C.V.; Kaur, H. A comprehensive investigation into the use of machine learning to forecast ovarian cancer. In Proceedings of the 14th International Conference on Computing Communication and Networking Technologies (ICCCNT), Delhi, India, 6–8 July 2023; pp. 1–8. [Google Scholar]
  24. Radhakrishnan, M.; Sampathila, N.; Muralikrishna, H.; Swathi, K.S. Advancing ovarian cancer diagnosis through deep learning and explainable AI: A multiclassification approach. IEEE Access 2024, 12, 116968–116986. [Google Scholar] [CrossRef]
  25. Lăzărescu, A.M.D.; Moldovanu, S.; Moraru, L. Iris-Based Biometric Identification Using a Combination of the Right-Left Iris Statistical Features. J. Phys. Conf. Ser. 2024, 2701, 012006. [Google Scholar] [CrossRef]
  26. Tabacaru, G.; Moldovanu, S.; Răducan, E.; Barbu, M. A Robust Machine Learning Model for Diabetic Retinopathy Classification. J. Imaging 2024, 10, 8. [Google Scholar] [CrossRef]
  27. Evrimler, S.; Gedik, M.A.; Serel, T.A.; Ertunc, O.; Ozturk, S.A.; Soyupek, S. Bladder Urothelial Carcinoma: Machine Learning-based Computed Tomography Radiomics for Prediction of Histological Variant. Acad. Radiol. 2022, 29, 1682–1689. [Google Scholar] [CrossRef]
  28. Andlib, N.; Sajad, M.; Thakur, S.C. Association of diabetes mellitus with risk of reproductive impairment in females: A comprehensive review. Acta Histochem. 2024, 126, 152173. [Google Scholar] [CrossRef]
  29. Tariq, S.; Nurulain, S.M.; Rashed, H.; Lotfy, M.; Emerald, S.B.; Koturan, S.; Tekes, K.; Adeghate, E. Diabetes-induced changes in the morphology and nociceptinergic innervation of the rat uterus. J. Mol. Histol. 2016, 47, 21–33. [Google Scholar] [CrossRef]
  30. Montero-Pena, I.; Morales-De Pando, J.M.; Diaz-Gomez, A.; Gonzalez-Medina, G.; Ribelles-Garcia, A.; Perez-Arana, G.; Prada-Oliveira, J.A. Characterization of the histological changes in ovaries of Goto-Kakizaki diabetic rats. Eur. J. Anat. 2024, 28, 77–81. [Google Scholar] [CrossRef]
  31. Ali, E.M.T.; Abdalllah, H.I.; El-Sayed, S.M. Histomorphological, VEGF and TGF- β immunoexpression changes in the diabetic rats ovary and the potential amelioration following treatment with metformin and insulin. J. Mol. Histol. 2020, 51, 287–305. [Google Scholar] [CrossRef] [PubMed]
  32. Escobar-Morreale, H.F.; Roldan-Martin, M.B. Type 1 diabetes and Polycystic Ovary Sindrome: Systematic review and Meta-analysis. Diabetes Care 2016, 39, 639–648. [Google Scholar] [CrossRef]
  33. Rocha, D.S.; Dentz, M.V.; Model, J.F.A.; Vogt, E.L.; Ohlweiler, R.; Lima, M.V.; de Souza, S.K.; Kucharski, L.C. Female Wistar rats present particular glucose flux when submitted to classic protocols of experimental diabetes. Biomed. J. 2023, 46, 100539. [Google Scholar] [CrossRef]
  34. Kottaisamy, C.P.D.; Raj, D.S.; Prasanth Kumar, V.; Sankaran, U. Experimental animal models for diabetes and its related complications—A review. Lab. Anim. Res. 2021, 37, 23. [Google Scholar] [CrossRef] [PubMed]
  35. Yagihashi, S. Contribution of animal models to diabetes research: Its history, significance, and translation to humans. J. Diabetes Investig. 2023, 14, 1015–1037. [Google Scholar] [CrossRef]
  36. Tătaru, I.; Gardikiotis, I.; Dragostin, O.-M.; Confederat, L.; Gîrd, C.; Zamfir, A.-S.; Morariu, I.D.; Chiţescu, C.L.; Dinu (Iacob), A.; Costea Popescu, L.; et al. Hormonal status and oxidative stress monitoring in an in vivo 2 model of experimentally induced diabetes in female rats. Biomedicines 2025, 13, 922. [Google Scholar]
  37. Xu, M.; Yoon, S.; Fuentes, A.; Park, D.S. A comprehensive survey of image augmentation techniques for deep learning. Pattern Recognit. 2023, 137, 10934. [Google Scholar] [CrossRef]
  38. Ihme, M.; Chung, W.T.; Mishra, A.A. Combustion machine learning: Principles, progress and prospects. Prog. Energy Combust. Sci. 2022, 91, 101010. [Google Scholar] [CrossRef]
  39. Narayanan, A.N.; Das, S.S.; Mirnalinee, T. Evaluation of AutoML Frameworks for Computational ADMET Screening in Drug Discovery & Development. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine, BIBM, Istanbul, Turkiye, 5–8 December 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 4929–4931, ISBN 9798350337488. [Google Scholar]
  40. Chu, C.S.; Simpson, J.D.; O’Neill, P.M.; Berry, N.G. Machine learning–Predicting Ames mutagenicity of small molecules. J. Mol. Graph. Model. 2021, 109, 108011. [Google Scholar] [CrossRef]
  41. Das, D.; Naskar, R. Image splicing detection using low-dimensional feature vector of texture features and Haralick features based on Gray Level Co-occurrence Matrix. Signal Process. Image Commun. 2024, 125, 117134. [Google Scholar] [CrossRef]
  42. Moraru, L.; Moldovanu, S.; Culea-Florescu, A.-L.; Bibicu, D.; Dey, N.; Ashour, A.S.; Sherratt, R.S. Texture spectrum coupled with entropy and homogeneity image features for myocardium muscle characterization. Curr. Bioinform. 2019, 14, 295–304. [Google Scholar] [CrossRef]
  43. Luo, X.; Wang, W.; Xu, Y.; Lai, Z.; Jin, X.; Zhang, B.; Zhang, D. A deep convolutional neural network for diabetic retinopathy detection via mining local and long-range dependence. CAAI Trans. Intell. Technol. 2023, 1, 153–166. [Google Scholar] [CrossRef]
Figure 1. Custom-built CNN architecture.
Figure 1. Custom-built CNN architecture.
Biomedicines 13 01494 g001
Figure 2. Samples of histopathological images. (A). Vagina (a1) Control, (a2) DM, and (a3) DM + synthetic compound; (B). Ovary (b1) Control, (b2) DM, and (b3) DM + synthetic compound; (C) Uterus (c1) Control, (c2) DM, and (c3) DM + synthetic compound.
Figure 2. Samples of histopathological images. (A). Vagina (a1) Control, (a2) DM, and (a3) DM + synthetic compound; (B). Ovary (b1) Control, (b2) DM, and (b3) DM + synthetic compound; (C) Uterus (c1) Control, (c2) DM, and (c3) DM + synthetic compound.
Biomedicines 13 01494 g002aBiomedicines 13 01494 g002b
Figure 3. Augmented histopathological images.
Figure 3. Augmented histopathological images.
Biomedicines 13 01494 g003
Table 1. Comparison of DL models for gynecological tissue.
Table 1. Comparison of DL models for gynecological tissue.
References/YearWeaknessAI AlgorithmsResults
Accuracy (%)
Patil and Patil [4]/2025-focuses on enhancing a CNN type
-time-consuming
TL-CNN89.59%
Sun et al. [5]/2019-low accuracy
-a limited ability to model
CNN84.50%
Kitaya et al. [6]/2024-custom-built CNNCNN92.8%
Asadpour et al. [7]/2024-focuses on using a pre-trained CNN
-low accuracy
ResNet-5082.88%
Song et al. [8]/2022-only a custom-built DNNDNN91.7%
Zhao et al. [9]/2022-global-to-local multi-scale CNN (G2LNet) CNN95.34%
Li et al. [10]/2023-only a custom-built DNN
-low accuracy
DNN75%
Zafar et al. [11]/2025-ML uterine tissue highly correlated 65 features.Gated Highway Multi-layer-perceptron (GHiM)86.6%
Peng et al. [12]/2020-only a pre-trained CNNResNet V286% (AUC)
Onishi et al. [13]/2022-low accuracy
-without fine-tuning
10 conventional ML 84%
Volinsky-Fremond et al. [14]/2024-low accuracyMultimodal DNN 78.9%
Jeleń et al. [15]/2025-without fine-tuningNN, SVM93.3%
Jagendra et al. [16]/2024-only a neural network and a pre-trained CNNVGG 16, Recurrent neural network (RNN)94.65% 87.65%
Rajan et al. [17]/2024-time-consumingInception v3, ResNet, XGBoost, SVM, KNN 90.37%
Zhou et al. [18]/2024-time-consumingSVM, RFXGBoost, CNN94.6%
El-Latif et al. [19]/2024-only a pre-trained CNNResNet-598.99%
Wang et al. [20]/2022-low accuracyInceptionV3, ViT76%
Behera et al. [21]/2024-only a pre-trained CNN
-low accuracy
EfficientNet-B078%
Kodipalli [22]/2022-only a pre-trained CNN
-low accuracy
ResNet v2 67%
Kwatra and Kaur [23]/2023-pre-trained
-time-consuming
MobileNetV3
ResNet50
96.3%
92.08%
Radhakrishnan et al. [24]/2024-time-consumingMobileNetV2, VGG19, ResNet18, ResNeXt, Xception, EfficientNet97.96%
Table 2. Comparison of the best performances of the CNN algorithm.
Table 2. Comparison of the best performances of the CNN algorithm.
Classes Test Set AccuracyF1-ScoreAUCConfusion Matrices
[[TP FP][FN TN]]
Ovarian DC1330.8420.8570.842[[63 9][12 49]]
Uterine DC550.9450.9500.943[[29 2][1 23]]
Vaginas DC620.750.750.739[[26 8][8 20]]
Ovarian AD_DC440.7710.7330.736[[22 5][6 11]]
Uterine AD_DC920.8580.8530.858[[38 7][6 41]]
Vaginas AD_DC820.7710.8530.786[[37 6][12 27]]
Table 3. Comparison of the best performances of PyCaret AutoML.
Table 3. Comparison of the best performances of PyCaret AutoML.
ClassesAccuracyF1-ScoreThe Selected Classifier
Ovarian DC0.7850.837ExtraTreesClassifier(criterion = ‘gini’, estimators = 100, random_state = 123)
Uterine DC0.6150.573KNeighborsClassifier(algorithm = ‘auto’, leaf_size = 30, metric = ‘minkowski’, n_neighbors = 5)
Vaginas DC0.7750.788LinearDiscriminantAnalysis(solver = ‘svd’, tol = 0.0001)
Ovarian AD_DC0.7510.822RandomForestClassifier(criterion = ‘gini’, n_estimators = 100)
Uterine AD_DC0.70.53LinearDiscriminantAnalysis(solver = ‘svd’, tol = 0.0001)
Vaginas AD_DC0.860.88LinearDiscriminantAnalysis (solver = ‘svd’, tol = 0.0001)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tătaru, I.; Moldovanu, S.; Dragostin, O.-M.; Chiţescu, C.L.; Zamfir, A.-S.; Dragostin, I.; Strat, L.; Zamfir, C.L. Auto Machine Learning and Convolutional Neural Network in Diabetes Mellitus Research—The Role of Histopathological Images in Designing and Exploring Experimental Models. Biomedicines 2025, 13, 1494. https://doi.org/10.3390/biomedicines13061494

AMA Style

Tătaru I, Moldovanu S, Dragostin O-M, Chiţescu CL, Zamfir A-S, Dragostin I, Strat L, Zamfir CL. Auto Machine Learning and Convolutional Neural Network in Diabetes Mellitus Research—The Role of Histopathological Images in Designing and Exploring Experimental Models. Biomedicines. 2025; 13(6):1494. https://doi.org/10.3390/biomedicines13061494

Chicago/Turabian Style

Tătaru, Iulian, Simona Moldovanu, Oana-Maria Dragostin, Carmen Lidia Chiţescu, Alexandra-Simona Zamfir, Ionut Dragostin, Liliana Strat, and Carmen Lăcrămioara Zamfir. 2025. "Auto Machine Learning and Convolutional Neural Network in Diabetes Mellitus Research—The Role of Histopathological Images in Designing and Exploring Experimental Models" Biomedicines 13, no. 6: 1494. https://doi.org/10.3390/biomedicines13061494

APA Style

Tătaru, I., Moldovanu, S., Dragostin, O.-M., Chiţescu, C. L., Zamfir, A.-S., Dragostin, I., Strat, L., & Zamfir, C. L. (2025). Auto Machine Learning and Convolutional Neural Network in Diabetes Mellitus Research—The Role of Histopathological Images in Designing and Exploring Experimental Models. Biomedicines, 13(6), 1494. https://doi.org/10.3390/biomedicines13061494

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop