Next Article in Journal
Atlas-Assisted Bone Age Estimation from Hand–Wrist Radiographs Using Multimodal Large Language Models: A Comparative Study
Next Article in Special Issue
Prognostic Value of Blood Urea Nitrogen for Acute Kidney Injury and Mortality in Vasculitis: A Large Cohort Study Using Multivariate Joint Model and Machine Learning
Previous Article in Journal
Molecular Landscape of Resected Thymomas: Insights from Mutational Profiling
Previous Article in Special Issue
Machine Learning Model with Fourier-Transform Infrared Spectroscopy (FTIR) as a Proof-of-Concept Tool for Predicting Group A Streptococcus (GAS) emm-Type in the Pediatric Population
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Addressing Class Imbalance in Fetal Health Classification: Rigorous Benchmarking of Multi-Class Resampling Methods on Cardiotocography Data

by
Zainab Subhi Mahmood Hawrami
1,
Mehmet Ali Cengiz
2,* and
Emre Dünder
3
1
Ministry of Higher Education and Scientific Research—KRG, Kirkuk Main Road, Erbil 44001, Kurdistan Region, Iraq
2
Department of Mathematics and Statistics, College of Science, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia
3
Department of Statistics, Faculty of Science, Ondokuz Mayıs University, Samsun 55139, Turkey
*
Author to whom correspondence should be addressed.
Diagnostics 2026, 16(3), 485; https://doi.org/10.3390/diagnostics16030485
Submission received: 31 December 2025 / Revised: 22 January 2026 / Accepted: 29 January 2026 / Published: 5 February 2026
(This article belongs to the Special Issue Artificial Intelligence in Biomedical Diagnostics and Analysis 2025)

Abstract

Background/Objectives: Fetal health is essential in prenatal care, influencing both maternal and fetal outcomes. Cardiotocography (CTG) monitors uterine contractions and fetal heart rate, yet manual interpretation exhibits significant inter-examiner variability. Machine learning offers automated alternatives; however, class imbalance in CTG datasets where pathological cases constitute less than 10% leads to poor detection of minority classes. This study aims to provide the first systematic benchmark comparing five resampling strategies across seven classifier families for multi-class CTG classification, evaluated using imbalance-aware metrics rather than overall accuracy alone. Methods: Seven machine learning models were employed: Naïve Bayes (NB), Random Forest (RF), Linear Discriminant Analysis (LDA), k-Nearest Neighbors (KNN), Linear Support Vector Machine (SVM), Multinomial Logistic Regression (MLR), and Multi-Layer Perceptron (MLP). To address class imbalance, we evaluated the original unbalanced dataset (base) and five resampling methods: SMOTE, BSMOTE, ADASYN, NearMiss, and SCUT. Performance was evaluated on a held-out test set using Balanced Accuracy (BACC), Macro-F1, the Macro-Matthews Correlation Coefficient (Macro-MCC), and Macro-Averaged ROC-AUC. We also report per-class ROC curves. Results: Among all models, RF proved most reliable. Training on the original distribution (base) yielded the highest BACC (0.9118), whereas RF combined with BSMOTE provided the strongest class-balanced performance (Macro-MCC = 0.8533, Macro-F1 = 0.9073) with a near-perfect ROC-AUC (approximately 0.986–0.989). Overall, resampling effects proved model dependent. While some classifiers achieved optimal performance on the natural class distribution, oversampling techniques, particularly SMOTE and BSMOTE, demonstrated significant improvements in minority class discrimination and class-balanced metrics across multiple model families. Notably, certain models benefited substantially from resampling, exhibiting enhanced Macro-F1, BACC, and minority class recall without sacrificing overall accuracy. Conclusions: These findings establish robust, model-agnostic baselines for CTG-based fetal health screening. They highlight that strategic oversampling can translate improved minority class discrimination into clinically meaningful performance gains, supporting deployment in cost-sensitive and threshold-aware clinical settings.

1. Introduction

Reducing fetal mortality and maintaining continuous surveillance of fetal health status are essential for ensuring the well-being of both the mother and the fetus [1]. Therefore, fetal heart rate (FHR) monitoring through cardiotocography (CTG) is a widely adopted technology for assessing fetal well-being during pregnancy and labor [2]. From a clinical perspective, CTG monitoring plays a significant role in determining obstetric outcomes. During pregnancy, women undergo substantial physiological and morphological changes that may influence fetal development and movement. Recognizing these changes is essential for healthcare providers to distinguish between normal physiological adaptations and potential pathological conditions that could compromise maternal and fetal health. Given these considerations, effective fetal monitoring is crucial in reducing adverse perinatal outcomes. Studies suggest that increased fetal movements may have a protective effect against cesarean delivery [3]. Healthcare professionals recommend routine assessments to evaluate fetal health, with CTG being widely acknowledged as a standard method for monitoring both fetal and maternal conditions during pregnancy [4]. Despite its widespread use, CTG interpretation frequently depends on the clinician’s expertise, which introduces variability in assessments and increases the likelihood of misinterpretation. Conventional CTG analysis may generate false alarms or fail to detect genuine signs of fetal distress, leading to either unnecessary interventions or insufficient monitoring.
The consequences of such misinterpretation extend beyond immediate clinical decisions. The burden of preventable perinatal morbidity extends beyond immediate clinical outcomes. Hypoxic-ischemic encephalopathy (HIE), a severe consequence of perinatal asphyxia, has an incidence of approximately 1.5 per 1000 live births in high-income countries. In low- and middle-income countries, estimates range from approximately 2.3 to 26 per 1000 live births, depending on region, diagnostic definitions, and access to care [5]. These statistics underscore the critical importance of effective fetal surveillance strategies, particularly in detecting and preventing hypoxic injury during labor and delivery.
On a global scale, the worldwide prevalence of adverse perinatal outcomes highlights the necessity of enhancing fetal surveillance systems. According to the United Nations Interagency Group for Child Mortality Estimation (UN IGME) report, approximately 2 million stillbirths occur annually, as of 2023, with 46% happening during labor. The majority of intrapartum deaths (98%) occur in low- and middle-income countries, where high rates of home births and unattended deliveries remain significant challenges. Improving access to care, particularly in these settings, is critical to reducing intrapartum-related deaths and disabilities.
Central to effective fetal surveillance is the accurate interpretation of CTG recordings. CTG analysis involves evaluating five key features: baseline FHR (typically 110–160 beats per minute), baseline variability (fluctuations in the baseline rate), accelerations (transient increases in FHR), decelerations (transient decreases), and uterine contraction patterns. The International Federation of Gynecology and Obstetrics (FIGO) updated guidelines in 2015 established a three-tier classification system. This system categorizes CTG traces as normal, suspicious, or pathological based on combinations of these features [6]. One of the major challenges in intrapartum fetal monitoring is that, despite the international classification systems for CTG patterns, there is a very high false-positive rate of around 60% [7]. The diagnostic accuracy of admission CTG differs across studies. Positive predictive values have been reported between 19% and 88%, while negative predictive values range from 88.6% to 100% [8,9,10,11,12].
Adding to these diagnostic challenges, continuous CTG monitoring produces substantial volumes of data, requiring considerable time for healthcare providers to analyze and interpret accurately [13]. As a result, computerized CTG analysis has been seen as a promising solution to overcome these issues [14]. Moreover, automated early warning systems that integrate computerized cardiotocography with perinatal health parameter databases can provide objective alerts for patients at risk, potentially improving outcomes [15]. To address these limitations, advancements in computerized fetal heart rate monitoring, application of big data and artificial intelligence, and innovations in home and remote monitoring are being explored [16].
To address these limitations in manual CTG interpretation, researchers have increasingly turned to machine learning approaches. The integration of machine learning into CTG analysis has gained momentum over the past decade. Studies on fetal health classification focus on the analysis of cardiotocographic data used in the evaluation of fetal health. Mehbodniya et al. [1] investigated the prediction of fetal health status from CTG data using various ML algorithms, including SVM, RF, MLP, and KNN. In this study, the authors classified fetal health status as either normal, suspect, or pathological. The RF algorithm and XGBoost exhibited the best performance, demonstrating higher accuracy, sensitivity, and F1-score values. Subsequently, Ocak [17] applied genetic algorithm-based feature selection with SVM classifiers, achieving 99.3% and 100% accuracy for normal and pathological CTG classification, respectively. The proposed scheme outperformed ANN and ANFIS-based methods. Hoodbhoy et al. [18] compared ten different machine learning algorithms on CTG data. The study employed SMOTE to address class imbalance, and the classification model developed using the XGBoost technique had the highest prediction accuracy for an adverse fetal outcome.
While these studies demonstrate the potential of machine learning for CTG-based fetal health classification, a critical challenge that affects model reliability remains largely underexplored: the need for high-quality and balanced datasets. However, real-world medical datasets are frequently imbalanced due to naturally skewed outcome distributions. Many existing studies have overlooked this issue despite its critical influence on model performance and clinical validity. Class imbalance biases models toward majority classes, reducing their capacity to identify minority patterns [19]. This asymmetry in clinical consequences highlights the importance of appropriate evaluation metrics. Standard accuracy measures can be misleading under severe class distribution skews, often favoring the majority class while masking poor performance on minority classes [20]. Models trained on balanced data typically achieve more stable convergence, enhanced interpretability, and improved detection of minority outcomes. When natural class balance is not achievable, techniques such as resampling, class weighting, or data augmentation help mitigate bias and enhance both fairness and clinical utility.
To address these challenges, this study investigates multi-class fetal health classification from CTG data under conditions of class imbalance. We evaluate seven machine learning algorithms: Naïve Bayes (NB), Random Forest (RF), Linear Discriminant Analysis (LDA), k-Nearest Neighbors (KNN), Linear Support Vector Machine (SVM), Multinomial Logistic Regression (MLR), and Multi-Layer Perceptron (MLP).
The selection of these seven algorithms was motivated by their representation of fundamentally different learning paradigms: Naïve Bayes represents probabilistic classifiers that assume feature independence, while Random Forest exemplifies ensemble methods that aggregate multiple decision trees to reduce variance. Linear Discriminant Analysis and Support Vector Machines represent linear approaches that seek optimal separating boundaries, whereas k-Nearest Neighbors relies on local instance similarity without explicit model construction. Multinomial Logistic Regression extends classical regression to multi-class settings, and Multi-Layer Perceptron captures non-linear relationships through layered neural architectures. As noted by Fernández et al. [21], different classifier families exhibit varying sensitivity to class imbalance, making algorithmic diversity essential for comprehensive evaluation.
To mitigate the effects of class imbalance, each algorithm was tested across five resampling strategies: SMOTE (Synthetic Minority Oversampling Technique), Borderline-SMOTE (BSMOTE), ADASYN (Adaptive Synthetic Sampling), NearMiss, and SCUT. These five resampling techniques were selected to represent distinct balancing strategies. SMOTE and its variants (Borderline-SMOTE, ADASYN) generate synthetic minority samples through interpolation, with each variant addressing specific limitations of the original algorithm. NearMiss applies undersampling by removing majority class instances based on their distance to minority samples, while SCUT combines both oversampling and undersampling in a hybrid framework. Comparing these approaches allows us to determine whether synthetic data generation or majority class reduction is more effective for CTG classification.
We also included the raw, unbalanced dataset as a baseline. Model performance is assessed on a held-out test set using four complementary metrics: Balanced Accuracy (BACC), Macro-F1-Score, Macro-Matthews Correlation Coefficient (Macro-MCC), and Macro-Averaged ROC-AUC. Per-class ROC curves are also presented to illustrate the discriminative performance for each fetal health category.
The primary contributions of this study are threefold: (1) systematic comparison of five resampling strategies across seven classifiers, addressing the fact that most studies evaluate only one or two methods; (2) rigorous evaluation using four imbalance-aware metrics rather than overall accuracy alone; and (3) practical guidelines for selecting classifier–resampling combinations suitable for clinical deployment. To the best of our knowledge, this is the first benchmarking study to systematically compare multiple resampling strategies across diverse classifier families specifically for imbalanced multi-class fetal health classification using class-balanced evaluation metrics.
The remainder of this paper is organized as follows: Section 2 presents related works, summarizing previous studies on fetal health classification using CTG data and identifying the current gaps motivating this research. Section 3 details the proposed methodology, including the dataset description, workflow design, algorithmic structure, and imbalance handling strategies. Section 3.1 outlines the machine learning algorithms and resampling techniques employed in model development. Section 3.2 describes the model evaluation framework, including the performance metrics adapted for multi-class and imbalanced data. Section 4 presents the results and discussion that interpret the experimental results, including ROC analyses. Section 5 concludes with key findings and outlines directions for future research.

2. Related Work

For more than three decades, CTG’s predictive capacity has remained controversial despite its widespread use in fetal risk assessment. Reported sensitivities range from 2% to 100%, while specificities vary between 37% and 100% [6,22]. This wide variability is driven largely by subjective visual interpretation. This motivates the adoption of ML to standardize analysis and improve discrimination among normal, suspect, and pathological fetal states. Using the canonical UCI CTG dataset (2126 records, 21 features, 3 class labels), multiple studies have demonstrated that classical ML methods can achieve high performance. Methods such as RF, SVM, Decision Tree (DT), and KNN have exceeded 90% accuracy in various evaluations [23,24].
Early studies confirmed this promise but also exposed limitations. Sahin et al. [25] compared eight algorithms including Logistic Regression, RF, ANN, KNN, DT, SVM, RBFN, and CART using ten-fold cross-validation. They found RF to be the most reliable for binary classification (normal vs. pathological). Spilka et al. [26] modeled intrapartum FHR trajectories with sparse SVMs, achieving approximately 70% accuracy. This underscored the challenge of noisy, time-varying signals. Complementary studies reported additional findings. Ramla et al. [27] achieved approximately 88.9% accuracy using DT techniques. Madiraju et al. [28] showed that MLPs can learn non-linear relations within FHR variability, accelerations, and decelerations. Collectively, these works established feasibility while highlighting the need to manage class imbalance and improve generalization.
Subsequent research incorporated feature selection, optimization, and cost-sensitive methods to strengthen performance. Piri et al. [29] benchmarked an association-based classifier against DT, Linear Regression, SVM, KNN, GNB, RF, and XGBoost. They observed modest gains after feature selection (83% to 84% accuracy). The authors noted that insufficient data augmentation can promote overfitting and hinder generalization. Vani et al. [30] introduced a weighted DNN and evaluated its performance using sensitivity, specificity, F1-score, and G-mean. The weighted DNN achieved a G-mean of 91% for positive-class detection, outperforming the SVM baseline. However, limited hyperparameter tuning likely constrained overall performance. Building on this work, Piri et al. [31] employed an evolutionary multi-objective genetic algorithm (MOGA-CD) to identify influential predictors in the CTG dataset. Using the selected features, DT, SVM, GNB, RF, and XGBoost achieved the highest classification accuracies among the evaluated models. Related contributions emphasized interpretability and error reduction. Zhou et al. [32] highlighted the value of DTs for identifying high-risk cases. Li et al. [33] showed that ML can reduce subjective error in CTG interpretation. Pradhan et al. [34] reported that RF achieved approximately 99% accuracy when compared with Logistic Regression, KNN, and Gradient-Boosting Machines.
As pipelines matured, explicit imbalance handling and ensemble methods became decisive. Rahmayanti et al. [35] compared ANN, LSTM, XGBoost, SVM, KNN, LightGBM, and RF across three scenarios. These scenarios varied outlier control, VIF-based collinearity filtering, and upsampling with SMOTE. LightGBM was uniquely consistent, achieving near-perfect performance (accuracy, F1-score, and AUC approximately 0.99) across settings. Kaliappan et al. [36] proposed an ensemble voting scheme combining KNN, GNB, and AdaBoost with Monte Carlo cross-validation. They reported approximately 99% accuracy, high recall (approximately 98%), and an F1-score of approximately 97%. Regmi and Shah [37] evaluated dimensionality-reduction pipelines (PCA, LDA) alongside attention-based and tabular learners. Their TabNet model achieved 94.36% accuracy on the fetal health dataset. Within classical pipelines, the LDA-based combination outperformed the RF with PCA variant. Their best pipeline for prenatal abnormality classification reached 91.13% accuracy. Two studies by Salini et al. [38] applied LR, RF, DT, SVM, KNN, GBM, and voting models. They repeatedly found RF to be strongest, with approximately 93% accuracy. Addressing imbalance from a different angle, Zeng et al. [39] proposed an ensemble cost-sensitive SVM (ESVM) using time–frequency features to distinguish normal versus abnormal fetuses.
Very recent benchmarks reinforce these patterns while extending to meta-learning and broader deep learning comparisons. Nazlı et al. [40] evaluated a large panel of algorithms: CatBoost, DT, ExtraTrees, GB, KNN, LightGBM, RF, SVM, ANN, and DNN with SMOTE to correct imbalance. They found LightGBM best (accuracy = 90.73%, balanced accuracy = 91.34%), confirming its robustness on three-class CTG. Ahmed et al. [13] combined seven ML and five DL models under SMOTE. A stacking meta-classifier surpassed all individual learners (accuracy = 98.9%, precision = 99.0%, recall = 98.6%, F1 = 99.3%, AUC = 99.8%). This suggests practical suitability for real-time CTG decision support. Bhukya et al. [41] compared wide-ranging ML methods (KNN, DT, SVM, RF, SGD, GB, AdaBoost, XGBoost) and DL methods (CNN, LSTM, BiLSTM). They found RF to be highest (approximately 99.39%), with XGBoost and SVM close behind. Among DL methods, BiLSTM and LSTM outperformed CNN but were slightly less stable on the suspect class.
Most prior studies report overall accuracy, which can be misleading for imbalanced datasets by favoring majority class performance. While some studies achieve higher accuracy values, these metrics may not reflect true performance on minority classes such as pathological cases. In contrast, our study employs class imbalance-robust metrics (BACC, Macro-F1, Macro-MCC, and Macro-Averaged ROC-AUC) that provide fairer evaluation across all fetal health categories. Unlike previous studies that addressed class imbalance through isolated techniques or focused on binary classification scenarios, we comprehensively benchmark five resampling strategies (SMOTE, BSMOTE, ADASYN, NearMiss, SCUT) across seven diverse model families (NB, RF, LDA, KNN, Linear SVM, MLR, and MLP). Our methodological rigor distinguishes this work through several key practices. A detailed methodology is presented in Section 3.

3. Methodology

3.1. Dataset Overview

CTG provides valuable information on FHR and UC, which are crucial to assessing the well-being of the fetus. Historically, CTG traces have been printed and interpreted manually, a process that is time-consuming and susceptible to inter-observer variability. Automating classification can provide timely decision support, facilitating earlier diagnosis and more efficient use of clinical resources.
To enable such automation, a standardized classification framework is essential. The CTG classification system categorizes fetal heart rate recordings into three distinct patterns, each carrying specific clinical implications [42,43]. These categories are as follows: normal, pathological, and suspect.
In a normal tracing, the baseline fetal heart rate ranges between 110 and 160 beats per minute, with variability maintained at 5 to 25 beats per minute and no repetitive decelerations present. These parameters indicate that fetal heart rate and uterine contraction patterns fall within healthy physiological boundaries, with no immediate concerns regarding fetal well-being. Clinically, such findings suggest adequate fetal oxygenation and typically require only routine monitoring without any intervention.
A pathological pattern, by contrast, is characterized by a baseline heart rate falling below 100 beats per minute, along with variability that may be reduced, elevated, or sinusoidal in nature. This category also includes recurrent late or prolonged decelerations lasting beyond 30 min, or beyond 20 min if variability is already reduced, as well as any single deceleration extending past 5 min. These abnormalities point to a high probability of fetal hypoxia or acidosis and demand a prompt clinical response. Management may involve conservative approaches such as maternal repositioning and oxygen supplementation, or, in more urgent situations, expedited delivery.
The suspect category occupies the middle ground between these two extremes. It applies when one or more normal criteria are not met yet no clearly pathological features are observed. This pattern reflects irregularities in the fetal heart rate that, while not immediately alarming, warrant closer attention and may signal developing problems. Clinically, suspect tracings call for heightened vigilance, more frequent monitoring, and preparedness to intervene should the pattern worsen.
Given these clinical consequences, accurate identification of pathological cases is particularly critical. To develop and evaluate automated classification models based on this framework, a reliable dataset with expert-validated labels is required. This study used the publicly available Cardiotocography (CTG) dataset from the University of California Irvine Machine Learning Repository (UCI ML Repository) [44]. The fetal CTG records were generated by SisPorto 2.0 (Speculum, Lisbon, Portugal), a software system for automated CTG analysis [45]. SisPorto 2.0 computes groups of 194 indicators commonly used in fetal monitoring, including fetal heart rate (FHR) parameters, uterine contractions (UC), fetal movements, short- and long-term variability measures, and histogram-based features aligning with standard clinical practice. These variables form the feature set used for model development in this work.
The dataset is a widely used benchmark in fetal health classification because of its structured feature set, expert annotations, and clear class labels. It comprises 2126 third-trimester cases, each with 21 attributes describing fetal status. Three specialist obstetricians independently evaluated each CTG and assigned a consensus label, which serves as the gold standard. Fetal status is categorized into three classes: normal (1655 records; 77.8%), suspect (295; 13.9%), and pathological (176; 8.3%), as illustrated in Figure 1. The variables used to quantify FHR and UC, along with histogram and variability features, are summarized in Table 1 [46].
As is evident from this distribution, the dataset exhibits considerable class imbalance, with pathological cases representing only 8.3% of all records. In this study, we address the class imbalance problem inherent in CTG-based fetal health data by evaluating multiple resampling strategies and their impact on machine learning model performance for multi-class fetal health classification [40].

3.2. Workflow of the Model

To address the challenges of multi-class and imbalanced CTG data, we designed a structured machine learning pipeline. This pipeline integrates stratified data partitioning, cross-validation, preprocessing, class balancing, classifier training, and performance evaluation. The workflow was implemented to ensure robustness and reproducibility and prevent data leakage. We applied seven different classifiers and evaluated them using both threshold-based and ranking-based performance measures. The following steps outline the detailed workflow, from data preparation to classifier training and final evaluation. Figure 2 illustrates the complete process, highlighting the adaptability and effectiveness of our approach for fetal health classification.
Workflow Steps
  • Data acquisition and labeling: The CTG dataset is obtained with three outcome classes: normal, suspect, and pathologic.
  • Train–test split: The data is partitioned using a 70/30 stratified split. The test set is set aside as an independent validation cohort for unbiased final evaluation.
  • Cross-validation for hyperparameter tuning: Within the training set, 5-fold cross-validation is performed. During each fold, four parts are used for training and one part serves as the validation set. This process is used exclusively for selecting the best hyperparameters. A fixed random seed (seed = 123) is used to ensure identical cross-validation folds across all models, enabling a fair performance comparison.
  • Class imbalance handling: Five different resampling methods are applied: SMOTE, BSMOTE, ADASYN, NearMiss, and SCUT. Each method is applied exclusively to the training data to address class imbalance and prevent data leakage. The original unbalanced training set is also retained as a baseline.
  • Balanced training data: Resampling produces balanced training sets with corrected class proportions. This reduces majority class bias during model training.
  • Feature preprocessing: Both training and test sets are standardized. The mean and standard deviation are computed only on the training data and then these same parameters are applied to the test data.
  • Model training and hyperparameter tuning: Seven machine learning algorithms are trained: NB, RF, LDA, KNN, SVM, MLR, and MLP. Hyperparameters are tuned via grid search within the cross-validation process. Probability outputs are generated for ROC analysis.
  • Prediction generation: Each trained model is used to produce predictions and probabilities on the held-out validation fold, aggregating results across folds.
  • Final evaluation: The retrained models are evaluated on the untouched test set. Performance is measured using BACC, Macro-F1, Macro-MCC, and Macro-Averaged ROC-AUC. Per-class ROC curves are generated using a one-vs-all approach, and individual class AUC values are averaged to compute the Macro-AUC.

3.3. Algorithms

Machine learning (ML) is a branch of artificial intelligence that extracts patterns from raw data through algorithmic modeling. Recent advances have accelerated its use in healthcare, enabling high-performing models for classifying fetal health. These developments provide a foundation for future work, allowing researchers to build on prior successes and refine predictive tools. ML shows substantial promise for streamlining fetal health classification and supporting clinical decision-making. By understanding what drives model predictions, we can improve model design and ultimately enhance patient outcomes [13,47]. ML is now widely applied across perinatal research. Prior studies have used ML to estimate fetal weight [48], assess the probability of fetal hypoxia [49], and predict fetal growth and gestational age [50]. In this study, we focus specifically on classifying fetal health from cardiotocography (CTG) data using machine learning approaches [35].
We evaluate seven different algorithms, each tuned over relevant hyperparameters to ensure a fair comparison. Models are trained with 5-fold cross-validation and standardized features (centering and scaling). To test robustness under class imbalance, we assess performance on multiple resampled datasets. These include the raw data as well as SMOTE, BSMOTE, ADASYN, NearMiss, and SCUT variants. The following sections describe the algorithms, their key characteristics, and the hyperparameter settings used in this work.

3.3.1. NB

This is a probabilistic classification method based on Bayes’ theorem. It estimates the probability of an event using prior information about related conditions [51]. The model assumes conditional independence among features within each class, even though some interdependence may exist in practice [52,53]. Despite its simplicity, NB is computationally efficient. It often performs well on text classification tasks, particularly when combined with TF-IDF features [54]. However, the independence assumption can limit its performance on more complex datasets [55]. In this study, we tuned NB using three key hyperparameters: First, usekernel enables or disables kernel density estimation for continuous predictors. Second, fL applies Laplace smoothing to address zero-frequency issues. Third, adjust controls the kernel bandwidth to adjust the smoothness of density estimation.

3.3.2. RF

This is a supervised ensemble method that extends bootstrap aggregating (bagging). It constructs multiple decision trees and aggregates their predictions to improve accuracy and reduce overfitting. Each tree is trained on a bootstrap sample of the data. At each split, a random subset of features is selected to increase diversity among trees [56]. For classification tasks, the final output is determined by majority voting across all trees. As the number of trees grows, the generalization error stabilizes. This error depends on both the strength of individual trees and their correlation [57]. In this study, RF was tuned by varying the number of variables considered at each split (mtry). We tested several proportions of the total predictors: p , 3 p 4 , p 2 , p 4 , p 10 , where p represents the number of features. The forest size was fixed at 1000 trees (ntree = 1000).

3.3.3. KNN

This is one of the simplest and earliest classification algorithms [58]. It classifies a new observation based on the labels of its nearest neighbors in the training set. The method assumes that similar samples are likely to belong to the same class. The parameter k specifies the number of neighbors considered in the voting process. Different values of k can produce different classification results for the same input. KNN is a non-parametric supervised learning method that does not require distributional assumptions. This makes it straightforward to implement and widely applicable [59]. In this study, KNN was tuned by varying the number of neighbors (k). We tested k values of 1, 3, 5, 7, 9, and 11.

3.3.4. Linear SVMs

Support Vector Machines classify data by finding a hyperplane that maximizes the margin between classes in a multi-dimensional feature space. Each observation is mapped to an n-dimensional space, and the algorithm identifies the hyperplane that best separates the classes. In this study, we used a linear SVM variant. The performance of the SVM largely depends on the regularization parameter, which controls the trade-off between margin maximization and classification errors. We tuned the cost parameter (C) at values of 0.1, 1, 10, and 100, with probability modeling enabled to generate class probabilities [53,60,61,62,63].

3.3.5. LDA

Linear Discriminant Analysis is a supervised machine learning algorithm used for classification and dimensionality reduction. Introduced by R.A. Fisher in 1936, it has become a foundational method in statistical pattern recognition [64,65]. LDA identifies linear combinations of features that maximize class separation. It does this by minimizing within-class variance while maximizing between-class variance. This projects high-dimensional data into a lower-dimensional subspace that ensures maximum class separability [66,67]. By representing samples through these linear projections, LDA constructs base vectors that capture class differences, enabling both dimensionality reduction and effective classification [68]. In this study, LDA was implemented without additional hyperparameter tuning as it does not require a parameter grid (grid = NULL).

3.3.6. MLR

Multinomial Logistic Regression is an extension of binary logistic regression designed for situations where the dependent variable has more than two nominal, unordered categories. It models the probability of each outcome category relative to a baseline using maximum-likelihood estimation [69,70]. This method is widely used for predicting outcomes with multiple classes, such as health status, job roles, or product preferences. Similar to binary logistic regression, MLR offers a flexible approach for modeling categorical outcomes [71,72]. In this study, MLR was tuned using a regularization parameter (decay) to prevent overfitting. We tested values of 0 and 1 × 10 10 . The model was trained with a maximum weight capacity of 100,000 and tracing disabled to improve computational efficiency.

3.3.7. MLP

Multi-Layer Perceptron is a feedforward artificial neural network widely used in supervised learning tasks such as classification, regression, and function approximation [13,73]. An MLP consists of multiple layers of interconnected neurons. The outputs of one layer serve as inputs to the next. Each neuron applies an activation function to its weighted inputs and transmits the result forward. This enables the network to capture complex patterns in the data [74,75]. The model is typically trained using the backpropagation algorithm. This algorithm iteratively updates connection weights by minimizing prediction error. Known for its flexibility and predictive accuracy, MLP has been successfully applied across diverse domains [76]. In this study, the MLP was tuned by varying two key hyperparameters: First, the number of hidden units (size) was tested at 3, 5, 7, 9, and 11. Second, weight decay (decay) was tested at values of 0, 1 × 10 5 , 1 × 10 4 , 1 × 10 3 , 1 × 10 2 , and 1 × 10 1 . The maximum number of iterations was set to 300.

3.4. Class Imbalance and Resampling Methods

Class imbalance significantly affects both model performance and clinical utility. It introduces bias toward the majority class and impairs the detection of rare but clinically critical minority cases [77]. When one class contains far more observations than others, overall model performance degrades [78]. In typical imbalanced datasets, most samples belong to the dominant class while only a small fraction represent the minority [79]. Models trained on such data tend to favor the majority class and often fail to recognize minority patterns [80]. As a result, minority class instances, often the most critical to identify, are frequently misclassified.
Imbalanced datasets pose four recurring challenges: bias, class overlap, high-dimensional feature spaces, and limited sample sizes [81]. These issues are especially pronounced in healthcare applications, such as diabetes diagnosis and skin lesion classification, as well as other safety-critical domains [82]. In medical diagnostics, the consequences can be severe. Models biased toward the majority class may miss rare diseases, leading to delayed or missed treatment and compromised patient care [83,84]. Addressing class imbalance is therefore fundamental to building reliable, fair, and clinically useful machine learning systems.
This concern extends directly to cardiotocography (CTG) analysis, where machine learning methods face severe class imbalance in clinically relevant outcomes. Standard algorithms trained on imbalanced data often achieve high overall accuracy yet fail to detect minority cases, a pattern observed across medicine and engineering [85].
In CTG analysis specifically, prior studies showed mixed results for resampling techniques. Almadi et al. [86] reported that applying SMOTE to a CTG dataset reduced accuracy despite strong baseline performance. Khan et al. [87] used random oversampling to mitigate imbalance. Nazli et al. [40] found that SMOTE improved Balanced Accuracy and reduced error rates across multiple models. Piri and Mohapatra [88] showed that seven resampling techniques substantially improved performance for RF, SVM, and DT on the UCI CTG dataset. These findings highlight that, while resampling can be beneficial, its effectiveness depends on both the dataset and the model, requiring careful design and validation.
Resampling-based approaches, such as Synthetic Minority Oversampling and random oversampling, offer practical ways to modify the training distribution and achieve more balanced class representation [89,90,91]. Their effectiveness depends on several factors, including imbalance ratio, dataset size, dimensionality, and class overlap. Selecting an appropriate strategy is crucial, as different techniques perform better under different conditions [92]. Moreover, hybrid methods that combine complementary strengths may further improve minority class recognition [93].
In this study, we applied five resampling techniques: SMOTE, BSMOTE, ADASYN, NearMiss, and SCUT. These methods were used to generate balanced training sets from the imbalanced CTG dataset. They aimed to improve minority class representation and enhance classifier robustness. Unless otherwise noted, neighborhood-based algorithms used k = 5, and oversampling targeted class parity (oversampling ratio = 1). This means minority classes were expanded to approximately match the majority class size.

3.4.1. SMOTE

This is an oversampling method that increases minority class representation to approach parity with the majority class [94]. Rather than simply duplicating existing observations, SMOTE synthesizes new minority samples through a more sophisticated process. It selects a minority instance, identifies its k-nearest minority neighbors, and generates new points through linear interpolation along the segments connecting them [95,96]. This approach enriches the minority region of the feature space, reduces the overfitting risk associated with simple duplication, and can improve downstream classifier performance [97].

3.4.2. BSMOTE

Borderline-SMOTE extends SMOTE by focusing on the most informative samples. Instead of oversampling all minority instances, BSMOTE generates synthetic samples only for minority instances located near the decision boundary between classes, where misclassification is most likely to occur [98]. By concentrating on these borderline cases, BSMOTE produces fewer but more strategically placed synthetic instances. This focused approach improves class separability and reduces overfitting compared to standard SMOTE [99,100].

3.4.3. ADASYN

This method addresses class imbalance by estimating the local distribution of the minority class and applying a weighted distribution to minority instances according to their learning difficulty [101]. Using the nearest neighbor technique, it generates synthetic instances in proportion to each instance’s difficulty, thereby balancing the dataset and enhancing classifier performance. The method offers two main advantages: First, it reduces bias from imbalanced class distributions. Second, it shifts the decision boundary toward hard-to-learn samples, improving minority class classification [102]. ADASYN assigns different weights to minority samples and automatically determines how many synthetic samples each should produce to achieve class balance [103].

3.4.4. NearMiss

This is a family of undersampling methods that address class imbalance by removing the majority class samples closest to minority instances. The algorithm first computes distances between majority and minority observations. It then focuses on majority points that lie near the minority set, thereby concentrating training data around the decision boundary [104,105]. In practice, the method works as follows: For each minority sample, it identifies its nearest majority neighbors. It retains the majority points most informative for class separation and discards the remaining majority samples to achieve a more balanced dataset. This targeted reduction of the majority class improves class separability and can enhance model generalization to unseen data [106,107,108].

3.4.5. SCUT

This method combines undersampling and oversampling to address class imbalance in multi-class datasets. The dataset is first divided into n subsets ( D 1 , ,   D n ), one for each class. The overall mean count m, representing the average number of instances across classes, is calculated. For classes with fewer instances than the mean, oversampling is applied until the class size reaches m. The SMOTE algorithm is used to determine the appropriate sampling percentage. For classes with more instances than the mean, undersampling is carried out to reduce their size to m. In this case, the Expectation Maximization (EM) clustering technique [109] is employed to identify clusters within each class. Instances are then randomly selected from these clusters so that the total equals m. Instead of fixing the number of samples per cluster, the selection is distributed flexibly to achieve uniform representation. The selected instances are merged to form the final balanced class. Classes that already contain exactly m instances remain unchanged [110].

3.5. Model Evaluation

Each classifier was evaluated on the held-out test set using the confusion matrix and a set of performance metrics specifically designed for imbalanced, multi-class data. The confusion matrix provides a structured summary of classification outcomes. It captures the counts of correct and incorrect assignments through four key components: true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN). From these basic components, we can define the total number of actual positives as P = TP + FN and the total number of actual negatives as N = TN + FP [111].
Classification metrics are derived directly from the confusion matrix and do not depend on probability estimates. Instead, they provide straightforward and interpretable evaluations of a model’s ability to correctly categorize instances. In this study, we focused on four key metrics: BACC, Macro-F1-Score, Macro-MCC, and Macro-Averaged ROC-AUC. These metrics were specifically chosen to address the challenges of imbalanced multi-class classification.
  • BACC: This metric mitigates the effect of class imbalance by averaging sensitivity ( T P / P ) and specificity ( T N / N ).
    BACC = 1 2 T P P + T N N
  • Macro-F1-Score: The F1-score is the harmonic mean of precision and recall, providing a balanced measure that accounts for both false positives and false negatives. In multi-class settings, the Macro-F1 is computed by calculating the F1-score separately for each class and then averaging the results. For each class i, precision ( P i ), recall ( R i ), and the corresponding F1-score ( F 1 i ) These are computed as follows:
    P i = T P i T P i + F P i , R i = T P i T P i + F N i , F 1 i = 2 P i R i P i + R i
    Finally, the Macro-F1-Score is obtained by taking the arithmetic mean across all k classes:
    Macro - F 1   =   1 k i = 1 k F 1 i
    This macro-averaging approach treats all classes equally, regardless of their size. This makes it particularly suitable for imbalanced datasets where minority class performance is critical. The Macro-F1 is widely applied across binary, multi-class, and multi-label classification problems, particularly in domains where class imbalance is prevalent [112,113].
  • Macro-MCC: The Matthews Correlation Coefficient (MCC) evaluates the quality of binary classifications by incorporating all four entries of the confusion matrix. It ranges from + 1 (perfect agreement) to 1 (complete disagreement), with 0 representing random prediction [114]. For multi-class problems, we employ a macro-averaging approach. MCC is computed separately for each class using a one-vs-rest strategy, where each class is treated as positive while all others are grouped as negative. The per-class MCC values are then averaged to obtain the Macro-Averaged MCC:
    Macro - MCC   =   1 k i = 1 k T P i T N i     F P i F N i ( T P i + F P i ) ( T P i + F N i ) ( T N i + F P i ) ( T N i + F N i )
    where k represents the number of classes, and T P i , T N i , F P i , and F N i denote the true positives, true negatives, false positives, and false negatives for class i, respectively. This macro-averaging approach ensures that each class contributes equally to the evaluation, making the metric robust to class imbalance [115,116].
  • Macro-Averaged ROC-AUC: Receiver Operating Characteristic (ROC) curves provide a visual and quantitative tool for evaluating classifier performance across different decision thresholds. Originally developed for medical decision-making, ROC analysis has become widely adopted in machine learning and data mining research. The area under the ROC curve (AUC) offers a single scalar measure of overall classification performance. AUC values range from 0 to 1, where 1 indicates perfect discrimination and 0.5 corresponds to random guessing. In practice, any meaningful classifier should achieve an AUC above 0.5 [117,118].
    For the multi-class problems, we employed a one-vs-rest strategy to compute the ROC curves and AUC values for each class independently [119]. In this approach, each class is treated sequentially as the positive class, while all remaining classes are combined to form the negative class. For our three-class CTG dataset (normal, suspect, pathological), this procedure generated three separate ROC curves. Each curve plots sensitivity (true-positive rate) against 1-specificity (false-positive rate) across all possible classification thresholds. The Macro-Averaged ROC-AUC was then calculated as the arithmetic mean of the individual class AUC values:
    Macro - AUC   =   1 k i = 1 k AUC i
    where k represents the number of classes and A U C i denotes the area under the ROC curve for class i. By averaging across classes rather than aggregating predictions, this macro-averaging approach ensures that each class contributes equally to the final metric, regardless of its frequency in the dataset. This property makes the Macro-AUC particularly well-suited for imbalanced classification tasks [120].
We selected these metrics because the CTG dataset is highly imbalanced. Conventional measures such as standard accuracy can obscure poor performance on minority classes. This equal weighting is particularly important in medical applications, where correct classification of rare cases such as pathological fetal conditions is of significant clinical importance. By using these imbalance-aware metrics alongside Balanced Accuracy, we ensured a comprehensive and fair assessment of model performance across all fetal health categories.

4. Results and Discussion

This section provides a comprehensive analysis of the performance of various machine learning models under different resampling methods. The analysis first focuses on overall performance across classifier–resampling combinations (Table 2), then examines class-wise outcomes with clinical relevance (Table 3). The findings are subsequently positioned within the recent CTG literature using Table A1 (Appendix A), followed by an evaluation of performance through ROC analyses (Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9).
Table 2 summarizes the comparative performance of seven machine learning algorithms evaluated under six configurations: the original unbalanced dataset (base) and five resampling methods (SMOTE, BSMOTE, ADASYN, NearMiss, and SCUT). Performance was assessed using four imbalance-aware metrics: BACC, Macro-MCC, Macro-F1, and Macro-Averaged ROC-AUC. The results are analyzed from three complementary perspectives: algorithm-specific performance, resampling method effectiveness, and metric-dependent interpretation.
According to the results in Table 2 the RF classifier consistently emerges as the most reliable model for the CTG dataset across all resampling strategies. The highest BACC (0.9118) is obtained by RF on the base data. Meanwhile, RF combined with BSMOTE yields slightly superior class-balanced performance metrics (Macro-MCC = 0.8533, Macro-F1 = 0.9073) compared with RF-base. This confirms its robustness and stability. The MLP model also performs strongly without resampling (BACC = 0.8636). SMOTE slightly increases its Macro-MCC and Macro-F1 (0.7862 and 0.8564) but lowers BACC. For KNN, the best Macro-MCC and Macro-F1 (0.7820 and 0.8513) are achieved with SMOTE, though the base model retains the highest BACC (0.8473). This indicates that synthetic neighbors improve minority class recall but may slightly blur decision boundaries. For linear or margin-based models (SVM, LDA, NB, MLR), the base datasets generally produce the best BACC values. In contrast, SMOTE or BSMOTE yield modest gains in Macro-MCC and Macro-F1. These trends are consistent with the expected effects of oversampling: shifting class priors and smoothing boundaries.
From a resampling perspective, the base configuration consistently achieved the highest BACC scores across all classifiers. This contrasts with conventional expectations regarding imbalance correction. Among oversampling methods, BSMOTE outperformed SMOTE and ADASYN for ensemble-based models by focusing on borderline minority instances. SMOTE provided moderate Macro-F1 improvements for distance-based and neural network models. ADASYN rarely outperformed alternatives and often yielded the lowest BACC scores due to heavy oversampling in difficult regions. NearMiss’s undersampling consistently degraded performance, underscoring the risk of discarding informative majority class samples. SCUT performed competitively with KNN but did not dominate overall.
From a metric-dependent perspective, evaluation criterion choice significantly influenced model ranking. BACC favored the base configuration, but may overestimate generalization when test data retains original class imbalance. Macro-MCC provided a more conservative estimate, with RF-BSMOTE achieving the highest value. Macro-F1 exhibited similar patterns to BACC but demonstrated greater sensitivity to minority class performance. The strong alignment between Macro-MCC and Macro-F1 trends suggests both metrics reliably capture class-balanced performance. In contrast, BACC may be more susceptible to threshold and prior-shift effects.
Collectively, these results confirm that resampling effectiveness is inherently model dependent. It must be assessed using multiple imbalance-aware metrics rather than relying on overall accuracy alone. RF trained on the base dataset emerged as the optimal model for multi-class fetal health classification from CTG. It balances high BACC with strong class-balanced metrics. BSMOTE offers a competitive alternative for scenarios prioritizing enhanced minority class detection, particularly when paired with ensemble methods. This experimental design serves as an ablation analysis, where comparing each resampling condition against the base (no resampling) baseline within each classifier isolates the specific contribution of resampling to model performance.
To complement these performance metrics, Table 3 presents per-class sensitivity, specificity, and precision values, providing a granular assessment of classifier behavior across fetal health categories. These metrics are particularly relevant for evaluating pathological class detection, where missed cases carry severe clinical consequences. For the normal class, all classifiers achieved high sensitivity (≥0.802), with RF-base reaching 0.974. Specificity values were more variable, ranging from 0.699 (LDA-base) to 0.985 (LDA-ADASYN). Precision for the normal class was consistently high (≥0.923), indicating reliable identification of healthy cases.
For the suspect class, which represents an intermediate clinical category, sensitivity remained challenging across all models. The highest suspect sensitivity was achieved by LDA-SMOTE and LDA-BSMOTE (0.919), while RF-base exhibited lower sensitivity (0.779) despite its overall superior performance. This trade-off reflects the inherent difficulty of distinguishing the suspect class from both normal and pathological categories. Specificity for the suspect class ranged from 0.826 (LDA-ADASYN) to 0.973 (RF-base), indicating variable false-positive rates. Precision values for the suspect class were generally lower than for other classes, with NearMiss and SCUT variants yielding values below 0.52 for several classifiers.
For the pathological class, RF consistently demonstrated the highest sensitivity values, with RF-BSMOTE achieving 0.936 and RF-NearMiss/SCUT reaching 0.957. These results indicate that RF-based models correctly identified over 93% of pathological cases regardless of resampling strategy. In contrast, NB exhibited the lowest pathological sensitivity (0.468 with ADASYN), highlighting its limitations for minority class detection. Regarding specificity, all classifiers maintained values above 0.94 for the pathological class, confirming minimal false-positive rates. Precision for pathological detection was highest with RF-BSMOTE (0.957) and RF-base (0.956), demonstrating that positive predictions for this critical class were highly reliable.
These per-class results complement the aggregate metrics in Table 2 by revealing class-specific trade-offs. Notably, oversampling methods improved pathological sensitivity for most classifiers but occasionally reduced suspect precision due to synthetic sample overlap near class boundaries. The RF classifier maintained the most balanced performance across all three classes, supporting its selection as the preferred model for clinical deployment.
To position our findings within the existing literature, a comparative summary of recent CTG classification studies is provided in Table A1 (Appendix A). The comparison includes study year, classification methods, number of classes, performance metrics, and imbalance handling strategies. Most prior studies reported only overall accuracy, which can be misleading for imbalanced datasets where the majority class dominates performance metrics. While several studies achieved accuracy values exceeding 95%, direct comparison remains difficult due to differences in evaluation metrics, classification schemes, and validation strategies. Among studies employing SMOTE, Ahmed et al. reported a high accuracy (98.9%) and F1-score (99.3%), yet relied primarily on aggregate metrics without per-class evaluation. In contrast, Nazlı et al. reported BACC (91.34%) using SMOTE, which aligns closely with our BACC result for RF (0.9118). This consistency across independent studies using balanced metrics strengthens confidence in the reported performance levels for three-class CTG classification.
To complement Table 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9 illustrate the ROC curves for each classifier under all resampling conditions. Each panel presents the macro-averaged AUC along with class-specific curves for the normal, suspect, and pathologic categories. This provides a comprehensive view of model discrimination performance across different resampling strategies. The algorithm-specific findings are summarized as follows:
  • NB (Figure 3): SMOTE achieves the highest performance (AUC = 0.942) compared to the base model (0.926), whereas BSMOTE and NearMiss reduce it to approximately 0.911–0.913. Despite these AUC improvements, BACC decreases under both SMOTE (0.8128 to 0.7671) and SCUT (0.8128 to 0.7013), demonstrating that enhanced ranking ability does not guarantee improved threshold-based classification accuracy.
  • RF (Figure 4): This achieves near-perfect AUCs (approximately 0.986–0.989) across all resampling methods, confirming its robustness and superior predictive calibration. The base, SMOTE, and BSMOTE models achieve identical top performance (AUC = 0.989), indicating highly stable and accurate class separation. Minor declines under ADASYN, NearMiss, and SCUT suggest that resampling offers no additional benefit.
  • LDA (Figure 5): This maintains high AUCs across methods (approximately 0.951–0.956), with a minor reduction under ADASYN (0.934). The base configuration remains optimal.
  • KNN (Figure 6): Overall AUC improves from the base model (0.870) under SMOTE (0.899) and SCUT (0.905), while NearMiss achieves the highest value (0.949). However, BACC consistently decreases across all configurations despite these AUC gains. This divergence occurs because improving ranking ability (AUC) does not guarantee better performance at a fixed decision threshold; resampling shifts class priors and alters optimal threshold positions, resulting in improved discrimination without corresponding gains in classification accuracy.
  • SVM (Figure 7): This records excellent performance with base and BSMOTE models (approximately 0.971–0.972), while ADASYN and NearMiss slightly lower AUCs to approximately 0.954–0.958.
  • MLR (Figure 8): This shows consistently strong performance (AUC approximately 0.964–0.970), with the base model performing best (0.970).
  • MLP (Figure 9): The base model exhibits the highest AUC (0.981), while oversampling slightly decreases performance (SMOTE = 0.979, BSMOTE and ADASYN= 0.974).

5. Conclusions

This study evaluated seven machine learning algorithms across six resampling configurations for multi-class fetal health classification from CTG data under severe class imbalance. Using four imbalance-robust metrics (BACC, Macro-MCC, Macro-F1, and Macro-AUC), we demonstrated that RF consistently emerges as the most reliable classifier. RF trained on the original unbalanced dataset achieved the highest BACC (0.9118), while RF combined with BSMOTE delivered superior class-balanced performance (Macro-MCC = 0.8533, Macro-F1 = 0.9073) with near-perfect AUC values (0.986–0.989). MLP also demonstrated strong performance without resampling, whereas linear and probabilistic models generally performed best on the base configuration. These patterns remained consistent across both threshold-based metrics (Table 2) and ROC curve analyses (Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9).
Resampling effects proved to be model-dependent, with notable differences in how algorithms responded to class balance manipulation. While RF, MLP, and linear models achieved optimal BACC on the natural class distribution, oversampling techniques, particularly SMOTE and BSMOTE, demonstrated significant improvements in minority class discrimination and class-balanced metrics across multiple model families. Specifically, KNN benefited substantially from SMOTE, exhibiting enhanced Macro-F1 (from 0.8346 to 0.8513) and Macro-MCC (from 0.7574 to 0.7820) without sacrificing overall accuracy. Similarly, MLP with SMOTE showed improved class-balanced metrics (Macro-MCC = 0.7862, Macro-F1 = 0.8564) and enhanced minority class recall. These findings demonstrate that strategic oversampling can translate improved class discrimination into clinically meaningful performance gains for specific model architectures, particularly distance-based and neural network methods.
The superiority of the base configuration for BACC can be explained by three factors: First, synthetic oversampling may introduce noise or unrealistic feature combinations that do not reflect the true underlying data distribution. Second, the test set retains the original class imbalance (normal: 77.8%, suspect: 13.9%, pathological: 8.3%), creating a distributional mismatch with oversampled training data that affects generalization. Third, models trained on balanced data may optimize decision boundaries that are suboptimal for the naturally imbalanced distribution encountered in clinical practice. Despite this, BSMOTE consistently improved class-balanced metrics across ensemble methods, confirming its value for applications where enhanced minority class detection is critical.
Importantly, improvements in AUC did not always lead to higher BACC; certain model–resampling combinations exhibited improved AUC despite decreased BACC, highlighting the distinction between ranking-based and threshold-based performance metrics. This pattern was observed in NB with SMOTE and SCUT as well as KNN, across all resampling methods. This divergence confirms that resampling can enhance ranking ability while reducing threshold-based accuracy due to class prior shifting effects.
Per-class results further highlighted clinically relevant differences. For the pathological class, RF achieved the highest sensitivity (0.957 with SCUT) and precision (0.957 with BSMOTE), ensuring reliable detection of high-risk cases. The suspect class remained the most challenging category, with lower sensitivity values across all classifiers due to its intermediate clinical characteristics. The normal class exhibited consistently high sensitivity and precision across all models. These findings emphasize that aggregate metrics alone may obscure clinically important class-specific variations, reinforcing the need for per-class evaluation in imbalanced medical classification tasks.
From a clinical perspective, these findings suggest that RF trained on the original class distribution offers the most reliable approach for routine fetal health monitoring, where pathological cases naturally occur at low frequency. The model’s consistently high performance across all metrics, combined with its robustness to resampling variations, makes it particularly suitable for deployment in clinical decision support systems. For scenarios prioritizing enhanced detection of rare pathological cases, such as high-risk obstetric units or screening programs, RF combined with BSMOTE, or alternatively KNN with SMOTE, provides compelling options that balance overall accuracy with improved minority class sensitivity. Methodologically, this study demonstrates that evaluating classifier performance under class imbalance requires multiple complementary metrics rather than relying on overall accuracy alone. The choice of resampling method must be tailored to the specific algorithm architecture and clinical priorities.
While these results are promising for clinical decision support, several limitations warrant consideration: First, this study utilized a single publicly available dataset from one institution, which may limit generalizability to different clinical settings, populations, or CTG acquisition devices. However, the UCI CTG dataset is widely used in fetal health research, allowing comparison with existing studies. Furthermore, testing five resampling methods across seven classifiers provides methodological guidance applicable to other imbalanced datasets. Nevertheless, the optimal resampling strategy may not directly transfer, and external validation is recommended before clinical deployment. Second, this study relies on preprocessed CTG features rather than raw signal traces, which may discard temporal patterns beneficial for classification. Moreover, the dataset lacks clinical variables such as maternal age, gestational age, and medical history, limiting comprehensive clinical modeling. Third, while we employed stratified train–test splitting and cross-validation to prevent data leakage, the relatively small sample size of minority classes (295 suspect, 176 pathologic) may have affected the stability of performance estimates. Additionally, the specific impact of each methodological element cannot be fully quantified in isolation. As a benchmarking study, our design compares model–resampling combinations; decomposing individual contributions would require factorial experimental designs.
Future research should extend this work in several directions: First, these methods should be evaluated on additional CTG datasets from different institutions to assess generalizability across diverse clinical populations and data distributions. Second, deep learning methods could be applied to raw CTG signals integrated with clinical metadata to capture temporal patterns and enable comprehensive clinical modeling. Third, data augmentation techniques or larger datasets with more balanced class distributions could improve the stability of performance estimates for minority classes. Fourth, future studies should employ repeated stratified cross-validation to report mean ± standard deviation and apply statistical significance tests (e.g., Friedman test, McNemar’s test) for more robust performance comparisons. Fifth, additional machine learning methods such as advanced ensemble algorithms (XGBoost, CatBoost, LightGBM) and deep learning architectures could be evaluated to determine whether they outperform the classical methods tested here. Sixth, other resampling techniques such as SMOTE variants (Borderline-SMOTE2, SVM-SMOTE) and hybrid methods combining oversampling and undersampling warrant investigation. Seventh, detailed misclassification analysis should be conducted to identify feature patterns associated with incorrectly classified pathological cases, providing valuable insights for clinical decision-making. Eighth, feature importance analysis and model explanation techniques (e.g., SHAP, permutation importance) should be incorporated to identify which CTG attributes most strongly influence classification decisions. Finally, additional imbalance-aware metrics such as G-mean, Cohen’s Kappa, and class-specific precision–recall curves could provide complementary perspectives on classifier performance.

Author Contributions

Z.S.M.H. contributed to the preparation of the original draft and was responsible for writing, reviewing, and editing the manuscript, as well as conducting the formal analysis and validating the results. E.D. contributed to the study methodology, wrote the code, performed formal analyses, validated the results, and participated in reviewing and editing the manuscript. M.A.C. provided supervision, oversaw project administration, and secured funding for the study. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-DDRSP2601).

Institutional Review Board Statement

Ethical approval was waived because this study is based entirely on the analysis of a publicly available dataset. The authors did not collect any primary data from human participants, nor did they have any direct contact with patients or their medical records.

Informed Consent Statement

Informed consent was waived due to the retrospective nature of the study.

Data Availability Statement

The original data presented in the study is openly available in the UCI ML Repository at https://archive.ics.uci.edu/dataset/193/cardiotocography (accessed on 20 September 2024).

Acknowledgments

The authors confirm that ClaudeAI (Opus 4.5, Anthropic) and ChatGPT 5.2 (OpenAI) was used exclusively to assist with English language editing, clarification of phrasing, and organizing LaTeX formatting. The tool was not involved in creating original scientific content, analyzing data, or drawing conclusions. All research findings and interpretations are entirely the work of the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Detailed performance comparisons of the proposed method with existing CTG classification studies are presented in Table A1.
Table A1. Performance comparison of the proposed method with existing CTG classification studies.
Table A1. Performance comparison of the proposed method with existing CTG classification studies.
StudyYearMethodsClassesPerformanceImbalance Handling
Sahin et al. [25]2015ANN, SVM, k-NN, RF, CART, LR, C4.5, RBFNBinaryAccuracy = 99.2%; F1-score = 99.2%; AUC = 99.9%No
Piri et al. [29]2019DT, LR, SVM, KNN, GNB, RF, XGBoost3-classAccuracy = 84%No
Piri et al. [31]2020MOGA-CD + DT, SVM, GNB, RF, XGBoost3-classAccuracy = 94%No
Pradhan et al. [34]2021RF, LR, KNN, GBM3-classAccuracy = 93%No
Rahmayanti et al. [35]2022ANN, LSTM, XGBoost, SVM, KNN, LightGBM, RF3-classAccuracy = 89–99%; F1-score = 98%; AUC = 99%SMOTE
Kaliappan et al. [36]2023DT, RF, SVM, KNN, GNB, AdaBoost, Gradient Boosting, Voting Classifier, Neural Networks3-classAccuracy = 99%; F1-score ≈ 97%; MCC = 96%Random Oversampling
Salini et al. [38]2024RF, LR, DT, SVC, KNN, Voting Classifier3-classAccuracy = 93%No
Nazlı et al. [40]2025CatBoost, DT, ExtraTrees, GB, KNN, LightGBM, RF, SVM, ANN, DNN3-classBACC = 91.34%SMOTE
Ahmed et al. [13]2025ML (RF, SVM, KNN, DT, LR, XGBoost, ET); DL (ANN, CNN, RNN, LSTM, GRU) + Stacking Meta-Classifier3-classAccuracy = 98.9%; F1-score = 99.3%; AUC = 99.8%SMOTE
Bhukya et al. [41]2025ML: KNN, DT, SVM, RF, SGD, GB, AdaBoost, XGBoost; DL: CNN, LSTM, BiLSTM3-classAccuracy ≈ 99.39%; F1-score = 99%No
Proposed work2026NB, RF, LDA, KNN, Linear SVM, MLR, MLP3-classBACC = 91.18% (RF + Base), 90.92% (RF + BSMOTE); Macro-F1 = 90.73%; Macro-MCC = 85.33%; Macro-ROC-AUC = 98.9% (RF + BSMOTE)Base, SMOTE, BSMOTE, ADASYN, NearMiss, SCUT

References

  1. Mehbodniya, A.; Lazar, A.J.P.; Webber, J.; Sharma, D.K.; Jayagopalan, S.; Singh, K.K.P.; Rajan, R.; Pandya, S.; Sengan, S. Fetal health classification from cardiotocographic data using machine learning. Expert Syst. 2022, 39, e12899. [Google Scholar] [CrossRef]
  2. Mendis, L.; Palaniswami, M.; Brownfoot, F.; Keenan, E. Computerised cardiotocography analysis for the automated detection of fetal compromise during labour: A review. Bioengineering 2023, 10, 1007. [Google Scholar] [CrossRef] [PubMed]
  3. Hantoushzadeh, S.; Gargari, O.K.; Jamali, M.; Farrokh, F.; Eshraghi, N.; Asadi, F.; Mirzamoradi, M.; Razavi, S.J.; Ghaemi, M.; Aski, S.K.; et al. The association between increased fetal movements in the third trimester and perinatal outcomes: A systematic review and meta-analysis. BMC Pregnancy Childbirth 2024, 24, 365. [Google Scholar] [CrossRef] [PubMed]
  4. Sundar, C.; Chitradevi, M.; Geetharamani, G. An analysis on the performance of K-means clustering algorithm for cardiotocogram data clustering. Int. J. Comput. Sci. Appl. 2012, 2, 11–20. [Google Scholar] [CrossRef]
  5. Namusoke, H.; Nannyonga, M.M.; Ssebunya, R.; Nakibuuka, V.K.; Mworozi, E. Incidence and short-term outcomes of neonates with hypoxic ischemic encephalopathy in a peri-urban teaching hospital, Uganda: A prospective cohort study. Matern. Health Neonatol. Perinatol. 2018, 4, 6. [Google Scholar] [CrossRef]
  6. Ayres-de-Campos, D.; Costa-Santos, C.; Bernardes, J. Prediction of neonatal state by computer analysis of fetal heart rate tracings: The antepartum arm of the SisPorto® multicentre validation study. Eur. J. Obstet. Gynecol. Reprod. Biol. 2005, 118, 52–60. [Google Scholar] [CrossRef]
  7. Pinas, A.; Chandraharan, E. Continuous cardiotocography during labour: Analysis, classification and management. Best Pract. Res. Clin. Obstet. Gynaecol. 2016, 30, 33–47. [Google Scholar] [CrossRef]
  8. Chittacharoen, A.; Chaitum, A.; Suthutvoravut, S.; Herabutya, Y. Fetal acoustic stimulation for early intrapartum assessment of fetal well-being. Int. J. Gynecol. Obstet. 2000, 69, 275–277. [Google Scholar] [CrossRef]
  9. Goonewardene, M.; Hanwellage, K. Fetal acoustic stimulation test for early intrapartum fetal monitoring. Ceylon Med. J. 2011, 56, 14–18. [Google Scholar] [CrossRef]
  10. Rahman, H.; Renjhen, P.; Dutta, S. Reliability of admission cardiotocography for intrapartum monitoring in low resource setting. Niger. Med. J. 2012, 53, 145–149. [Google Scholar] [CrossRef]
  11. David, B.; Saraswathi, K. Role of admission CTG as a screening test to predict fetal outcome and mode of delivery. Res. J. Pharm. Biol. Chem. Sci. 2014, 5, 295–299. [Google Scholar]
  12. Housseine, N.; Punt, M.C.; Browne, J.L.; Meguid, T.; Klipstein-Grobusch, K.; Kwast, B.E.; Franx, A.; Grobbee, D.E.; Rijken, M.J. Strategies for intrapartum foetal surveillance in low- and middle-income countries: A systematic review. PLoS ONE 2018, 13, e0206295. [Google Scholar] [CrossRef] [PubMed]
  13. Ahmed, S.S.; Mahmoud, N.M. Early detection of fetal health status based on cardiotocography using artificial intelligence. Neural Comput. Appl. 2025, 37, 16753–16779. [Google Scholar] [CrossRef]
  14. Ignatov, P.N.; Lutomski, J.E. Quantitative cardiotocography to improve fetal assessment during labor: A preliminary randomized controlled trial. Eur. J. Obstet. Gynecol. Reprod. Biol. 2016, 205, 91–97. [Google Scholar] [CrossRef]
  15. Eenkhoorn, C.; van den Wildenberg, S.; Goos, T.G.; Dankelman, J.; Franx, A.; Eggink, A.J. A systematic catalog of studies on fetal heart rate pattern and neonatal outcome variables. J. Perinat. Med. 2025, 53, 94–109. [Google Scholar] [CrossRef]
  16. Lovers, A.; Daumer, M.; Frasch, M.G.; Ugwumadu, A.; Warrick, P.; Vullings, R.; Pini, N.; Tolladay, J.; Petersen, O.B.; Lederer, C.; et al. Advancements in fetal heart rate monitoring: A report on opportunities and strategic initiatives for better intrapartum care. BJOG Int. J. Obstet. Gynaecol. 2025, 132, 853–866. [Google Scholar] [CrossRef]
  17. Ocak, H. A medical decision support system based on support vector machines and the genetic algorithm for the evaluation of fetal well-being. J. Med. Syst. 2013, 37, 9913. [Google Scholar] [CrossRef]
  18. Hoodbhoy, Z.; Noman, M.; Shafique, A.; Nasim, A.; Chowdhury, D.; Hasan, B. Use of machine learning algorithms for prediction of fetal risk using cardiotocographic data. Int. J. Appl. Basic Med. Res. 2019, 9, 226–230. [Google Scholar] [CrossRef]
  19. Mooijman, P.; Catal, C.; Tekinerdogan, B.; Lommen, A.; Blokland, M. The effects of data balancing approaches: A case study. Appl. Soft Comput. 2023, 132, 109853. [Google Scholar] [CrossRef]
  20. He, H.; Garcia, E.A. Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar] [CrossRef]
  21. Fernández, A.; García, S.; Galar, M.; Prati, R.C.; Krawczyk, B.; Herrera, F. Learning from Imbalanced Data Sets; Springer: Cham, Switzerland, 2018. [Google Scholar]
  22. Devoe, L.D.; Castillo, R.A.; Sherline, D.M. The nonstress test as a diagnostic test: A critical reappraisal. Am. J. Obstet. Gynecol. 1985, 152, 1047–1053. [Google Scholar] [CrossRef]
  23. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  24. Cömert, Z.; Kocamaz, A. Comparison of machine learning techniques for fetal heart rate classification. Acta Phys. Pol. A 2017, 132, 451–454. [Google Scholar] [CrossRef]
  25. Sahin, H.; Subasi, A. Classification of the cardiotocogram data for anticipation of fetal risks using machine learning techniques. Appl. Soft Comput. 2015, 33, 231–238. [Google Scholar] [CrossRef]
  26. Spilka, J.; Frecon, J.; Leonarduzzi, R.; Pustelnik, N.; Abry, P.; Doret, M. Intrapartum fetal heart rate classification from trajectory in sparse SVM feature space. In Proceedings of the IEEE Engineering in Medicine and Biology Conference (EMBC), Milan, Italy, 25–29 August 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 2335–2338. [Google Scholar] [CrossRef]
  27. Ramla, M.; Sangeetha, S.; Nickolas, S. Fetal health state monitoring using decision tree classifier from cardiotocography measurements. In Proceedings of the International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 14–15 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1799–1803. [Google Scholar] [CrossRef]
  28. Madiraju, R.; Upadhyay, U.; C, M.; Bharati, R. Fetal health analysis based on CTG. In Proceedings of the 11th International Conference on Communication and Signal Processing (ICCSP), Melmaruvathur, India, 5–7 June 2025; IEEE: Piscataway, NJ, USA, 2025; pp. 1706–1711. [Google Scholar] [CrossRef]
  29. Piri, J.; Mohapatra, P. Exploring fetal health status using an association-based classification approach. In Proceedings of the International Conference on Information Technology (ICIT), Bhubaneswar, India, 19–21 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 166–171. [Google Scholar] [CrossRef]
  30. Vani, R. Weighted deep neural network based clinical decision support system for the determination of fetal health. Int. J. Recent Technol. Eng. 2019, 8, 8564–8569. [Google Scholar] [CrossRef]
  31. Piri, J.; Mohapatra, P.; Dey, R. Fetal health status classification using MOGA-CD based feature selection approach. In Proceedings of the IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT), Bangalore, India, 2–4 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar] [CrossRef]
  32. Zhou, D.; Wang, J.; Xu, X. Multi-channel signal analysis for fetal heart rate monitoring using cardiotocography. IEEE Trans. Instrum. Meas. 2020, 69, 1013–1023. [Google Scholar]
  33. Li, J.; Liu, X. Fetal health classification based on machine learning. In Proceedings of the IEEE 2nd International Conference on Big Data, Artificial Intelligence and Internet of Things Engineering (ICBAIE), Nanchang, China, 26–28 March 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 899–902. [Google Scholar] [CrossRef]
  34. Pradhan, A.K.; Rout, J.K.; Maharana, A.B.; Balabantaray, B.K.; Ray, N.K. A machine learning approach for the prediction of fetal health using CTG. In Proceedings of the International Conference on Optical and Intelligent Technologies (OCIT), Hubli, India, 25–27 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 239–244. [Google Scholar] [CrossRef]
  35. Rahmayanti, N.; Pradani, H.; Pahlawan, M.; Vinarti, R. Comparison of machine learning algorithms to classify fetal health using cardiotocogram data. Procedia Comput. Sci. 2022, 197, 162–171. [Google Scholar] [CrossRef]
  36. Kaliappan, J.; Bagepalli, A.R.; Almal, S.; Mishra, R.; Hu, Y.-C.; Srinivasan, K. Impact of cross-validation on machine learning models for early detection of intrauterine fetal demise. Diagnostics 2023, 13, 1692. [Google Scholar] [CrossRef]
  37. Regmi, B.; Shah, C. Classification methods based on machine learning for the analysis of fetal health data. arXiv 2023, arXiv:2311.10962. [Google Scholar] [CrossRef]
  38. Salini, Y.; Mohanty, S.N.; Ramesh, J.V.N.; Yang, M.; Mukkoti, M.V.C. Cardiotocography data analysis for fetal health classification using machine learning models. IEEE Access 2024, 12, 26005–26022. [Google Scholar] [CrossRef]
  39. Zeng, R.; Lu, Y.; Long, S.; Wang, C.; Bai, J. Cardiotocography signal abnormality classification using time-frequency features and ensemble cost-sensitive SVM classifier. Comput. Biol. Med. 2021, 130, 104218. [Google Scholar] [CrossRef] [PubMed]
  40. Nazli, I.; Korbeko, E.; Dogru, S.; Kugu, E.; Sahingoz, O.K. Early detection of fetal health conditions using machine learning for classifying imbalanced cardiotocographic data. Diagnostics 2025, 15, 1250. [Google Scholar] [CrossRef] [PubMed]
  41. Bhukya, R.K.; Kayande, D.D.; Jain, A.; Agrawal, S.; Chitravanshi, P.; Walia, S.V. Enhancing fetal health classification: A study on fetal cardiotocograms. Procedia Comput. Sci. 2025, 260, 217–225. [Google Scholar] [CrossRef]
  42. Ayres-de-Campos, D.; Spong, C.Y.; Chandraharan, E. FIGO consensus guidelines on intrapartum fetal monitoring: Cardiotocography. Int. J. Gynecol. Obstet. 2015, 131, 13–24. [Google Scholar] [CrossRef]
  43. Wong, W.K.; Juwono, F.H.; Apriono, C.; Fitri, I.R. Fetal health prediction from cardiotocography recordings using Kolmogorov–Arnold networks. IEEE Open J. Eng. Med. Biol. 2025, 6, 345–351. [Google Scholar] [CrossRef]
  44. Campos, D.; Bernardes, J. Cardiotocography. UCI Mach. Learn. Repos. 2000. [Google Scholar] [CrossRef]
  45. Ayres-de-Campos, D.; Bernardes, J.; Garrido, A.; Marques-de-Sa, J.; Pereira-Leite, L. SisPorto 2.0: A program for automated analysis of cardiotocograms. J. Matern.-Fetal Med. 2000, 9, 311–318. [Google Scholar] [CrossRef]
  46. Ilham, A.; Kindarto, A.; Fathurohman, A.; Khikmah, L.; Ramadhani, R.D.; Jawad, S.A.; Liana, D.A.; Amylia, A.; Oleiwi, A.K.; Mutiar, A.; et al. CFCM-SMOTE: A robust fetal health classification to improve precision modelling in multi-class scenarios. Int. J. Comput. Digit. Syst. 2024, 16, 471–486. [Google Scholar] [CrossRef]
  47. Yin, Y.; Bingi, Y. Using machine learning to classify human fetal health and analyze feature importance. BioMedInformatics 2023, 3, 280–298. [Google Scholar] [CrossRef]
  48. Solt, I.; Caspi, O.; Beloosesky, R.; Weiner, Z.; Avdor, E. Machine learning approach to fetal weight estimation. Am. J. Obstet. Gynecol. 2019, 220, S666–S667. [Google Scholar] [CrossRef]
  49. Alsaggaf, W.; Cömert, Z.; Nour, M.; Polat, K.; Brdesee, H.; Toğaçar, M. Predicting fetal hypoxia using common spatial pattern and machine learning from cardiotocography signals. Appl. Acoust. 2020, 167, 107429. [Google Scholar] [CrossRef]
  50. Ananth, C.V.; Brandt, J.S. Fetal growth and gestational age prediction by machine learning. Lancet Digit. Health 2020, 2, e336–e337. [Google Scholar] [CrossRef] [PubMed]
  51. Lindley, D.V. Fiducial distributions and Bayes’ theorem. J. R. Stat. Soc. Ser. B (Methodol.) 1958, 20, 102–107. [Google Scholar] [CrossRef]
  52. Rish, I. An empirical study of the naive Bayes classifier. In Proceedings of the IJCAI Workshop on Empirical Methods in Artificial Intelligence, Seattle, WA, USA, 4 August 2001; IJCAI: Seattle, WA, USA, 2001; pp. 41–46. [Google Scholar]
  53. Uddin, S.; Khan, A.; Hossain, M.E.; Moni, M.A. Comparing different supervised machine learning algorithms for disease prediction. BMC Med. Inform. Decis. Mak. 2019, 19, 281. [Google Scholar] [CrossRef] [PubMed]
  54. Gupta, A.; Batla, A.; Kumar, C.; Jain, G. Comparative analysis of machine learning models for fake news classification. J. Xi’An Shiyou Univ. Nat. Sci. Ed. 2023, 19, 1250–1254. [Google Scholar]
  55. Zhang, L. Features extraction based on Naive Bayes algorithm and TF-IDF for news classification. PLoS ONE 2025, 20, e0327347. [Google Scholar] [CrossRef]
  56. Montesinos-López, O.A.; Montesinos-López, A.; Mosqueda-Gonzalez, B.A.; Montesinos-López, J.C.; Crossa, J.; Ramirez, N.L.; Valladares-Anguiano, F.A. A zero altered Poisson random forest model for genomic-enabled prediction. G3 Genes Genomes Genet. 2021, 11, jkaa057. [Google Scholar] [CrossRef]
  57. Tyralis, H.; Papacharalampous, G.; Langousis, A. A brief review of random forests for water scientists and practitioners and their recent history in water resources. Water 2019, 11, 910. [Google Scholar] [CrossRef]
  58. Nhu, V.H.; Shirzadi, A.; Shahabi, H.; Singh, S.K.; Al-Ansari, N.; Clague, J.J.; Ahmad, B.B. Shallow landslide susceptibility mapping: A comparison between logistic model tree, logistic regression, naive Bayes tree, artificial neural networks, and support vector machine algorithms. Int. J. Environ. Res. Public Health 2020, 17, 2749. [Google Scholar] [CrossRef]
  59. Singh, M.S.; Thongam, K.; Choudhary, P.; Bhagat, P.K. An integrated machine learning approach for congestive heart failure prediction. Diagnostics 2024, 14, 736. [Google Scholar] [CrossRef]
  60. Joachims, T. Making large-scale SVM learning practical. In Advances in Kernel Methods: Support Vector Learning; MIT Press: Cambridge, MA, USA, 1999; pp. 169–184. Available online: http://svmlight.joachims.org/ (accessed on 22 October 2025).
  61. Hussain, M.; Wajid, S.K.; Elzaart, A.; Berbar, M. A comparison of SVM kernel functions for breast cancer detection. In Proceedings of the International Conference on Computer Graphics, Imaging and Visualization (CGIV), Singapore, 17–19 August 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 145–150. [Google Scholar] [CrossRef]
  62. Pal, S.; Peng, Y.; Aselisewine, W.; Barui, S. A support vector machine-based cure rate model for interval censored data. Stat. Methods Med. Res. 2023, 32, 2405–2422. [Google Scholar] [CrossRef]
  63. Iftikhar, H.; Khan, M.; Khan, Z.; Khan, F.; Alshanbari, H.M.; Ahmad, Z. A comparative analysis of machine learning models: A case study in predicting chronic kidney disease. Sustainability 2023, 15, 2754. [Google Scholar] [CrossRef]
  64. Dogantekin, E.; Dogantekin, A.; Avci, D.; Avci, L. An intelligent diagnosis system for diabetes on linear discriminant analysis and adaptive network based fuzzy inference system: LDA-ANFIS. Digit. Signal Process. 2010, 20, 1248–1255. [Google Scholar] [CrossRef]
  65. Çalişir, D.; Doğantekin, E. An automatic diabetes diagnosis system based on LDA–Wavelet support vector machine classifier. Expert Syst. Appl. 2011, 38, 8311–8315. [Google Scholar] [CrossRef]
  66. Tharwat, A.; Gaber, T.; Ibrahim, A.; Hassanien, A.E. Linear discriminant analysis: A detailed tutorial. AI Commun. 2017, 30, 169–190. [Google Scholar] [CrossRef]
  67. Egwom, O.J.; Hassan, M.; Tanimu, J.J.; Hamada, M.; Ogar, O.M. An LDA–SVM machine learning model for breast cancer classification. BioMedInformatics 2022, 2, 345–358. [Google Scholar] [CrossRef]
  68. Choubey, D.K.; Kumar, M.; Shukla, V.; Tripathi, S.; Dhandhania, V.K. Comparative analysis of classification methods with PCA and LDA for diabetes. Curr. Diabetes Rev. 2020, 16, 833–850. [Google Scholar] [CrossRef]
  69. Friedman, J.H.; Hastie, T.; Tibshirani, R. Regularization paths for generalized linear models via coordinate descent. J. Stat. Softw. 2010, 33, 1–22. [Google Scholar] [CrossRef]
  70. Madhu, B.; Ashok, N.C.; Balasubramanian, S. Multinomial logistic regression predicted probability map to visualize the influence of socio-economic factors on breast cancer occurrence in Southern Karnataka. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2014, 40, 193–196. [Google Scholar] [CrossRef]
  71. Ueki, K.; Hino, H.; Kuwatani, T. Geochemical discrimination and characteristics of magmatic tectonic settings: A machine-learning-based approach. Geochem. Geophys. Geosystems 2018, 19, 1327–1347. [Google Scholar] [CrossRef]
  72. Itano, K.; Ueki, K.; Iizuka, T.; Kuwatani, T. Geochemical discrimination of monazite source rock based on machine learning techniques and multinomial logistic regression analysis. Geosciences 2020, 10, 63. [Google Scholar] [CrossRef]
  73. Jain, A.K.; Mao, J.; Mohiuddin, K.M. Artificial neural networks: A tutorial. Computer 1996, 29, 31–44. [Google Scholar] [CrossRef]
  74. Abdelbasset, W.K.; Elkholi, S.M.; Opulencia, M.J.C.; Diana, T.; Su, C.-H.; Alashwal, M.; Nguyen, H.C. Development of multiple machine-learning computational techniques for optimization of heterogeneous catalytic biodiesel production from waste vegetable oil. Arab. J. Chem. 2022, 15, 103843. [Google Scholar] [CrossRef]
  75. Ekinci, G.; Ozturk, H.K. Forecasting wind farm production in the short, medium, and long terms using various machine learning algorithms. Energies 2025, 18, 1125. [Google Scholar] [CrossRef]
  76. Sumayli, A. Development of advanced machine learning models for optimization of methyl ester biofuel production from papaya oil: Gaussian process regression (GPR), multilayer perceptron (MLP), and K-nearest neighbor (KNN) regression models. Arab. J. Chem. 2023, 16, 104833. [Google Scholar] [CrossRef]
  77. Suguna, R.; Prakash, J.S.; Pai, H.A.; Mahesh, T.R.; Kumar, V.V.; Yimer, T.E. Mitigating class imbalance in churn prediction with ensemble methods and SMOTE. Sci. Rep. 2025, 15, 16256. [Google Scholar] [CrossRef]
  78. Haykin, S. Neural Networks: A Comprehensive Foundation; Prentice Hall PTR: Englewood Cliffs, NJ, USA, 1994. [Google Scholar]
  79. Chen, S.M. Data Science and Big Data: An Environment of Computational Intelligence; Springer: Cham, Switzerland, 2017. [Google Scholar]
  80. Leevy, J.L.; Khoshgoftaar, T.M.; Bauder, R.A.; Seliya, N. A survey on addressing high-class imbalance in big data. J. Big Data 2018, 5, 42. [Google Scholar] [CrossRef]
  81. Sowjanya, A.M.; Mrudula, O. Effective treatment of imbalanced datasets in health care using modified SMOTE coupled with stacked deep learning algorithms. Appl. Nanosci. 2023, 13, 1829–1840. [Google Scholar] [CrossRef]
  82. Chowdhury, M.M.; Ayon, R.S.; Hossain, M.S. An investigation of machine learning algorithms and data augmentation techniques for diabetes diagnosis using class imbalanced BRFSS dataset. Healthc. Anal. 2024, 5, 100297. [Google Scholar] [CrossRef]
  83. Suresh, T.; Brijet, Z.; Subha, T.D. Imbalanced medical disease dataset classification using enhanced generative adversarial network. Comput. Methods Biomech. Biomed. Eng. 2023, 26, 1702–1718. [Google Scholar] [CrossRef]
  84. Altalhan, M.; Algarni, A.; Alouane, M.T.H. Imbalanced data problem in machine learning: A review. IEEE Access 2025, 13, 13686–13699. [Google Scholar] [CrossRef]
  85. Chakravarthy, A.D.; Bonthu, S.; Chen, Z.; Zhu, Q. Predictive models with resampling: A comparative study of machine learning algorithms and their performances on handling imbalanced datasets. In Proceedings of the IEEE International Conference on Machine Learning and Applications (ICMLA), Boca Raton, FL, USA, 16–19 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1492–1495. [Google Scholar] [CrossRef]
  86. Almadi, M.; Alotaibi, F.; Almudawah, R.; Ali, A.; Nasser, Y.; Nasser, N. Data-driven machine learning models for enhanced fetal health classification and monitoring. In Proceedings of the International Conference on Computing, Data Management and Analytics (CDMA), Riyadh, Saudi Arabia, 16–17 February 2025; pp. 189–192. [Google Scholar] [CrossRef]
  87. Khan, M.; Ahmad, A.; Sarfraz, M. L1 regularization based fetal health analysis using ML techniques. In Proceedings of the International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kamand, India, 24–28 June 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–6. [Google Scholar] [CrossRef]
  88. Piri, J.; Mohapatra, P. Imbalanced cardiotocography data classification using re-sampling techniques. In Machine Intelligence and Data Analytics for Sustainable Systems; Springer: Singapore, 2021; pp. 681–692. [Google Scholar] [CrossRef]
  89. Cicak, S.; Avci, U. Handling imbalanced data in predictive maintenance: A resampling-based approach. In Proceedings of the International Conference on Human-Oriented Robotics and Applications (HORA), Istanbul, Turkey, 8–10 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar] [CrossRef]
  90. Brownlee, J. Random Oversampling and Undersampling for Imbalanced Classification. 2021. Available online: https://machinelearningmastery.com/random-oversampling-and-undersampling-for-imbalanced-classification/ (accessed on 28 October 2025).
  91. Paula, B.; Torgo, L.; Ribeiro, R. A survey of predictive modelling under imbalanced distributions. arXiv 2015, arXiv:1505.01658. [Google Scholar] [CrossRef]
  92. Kraiem, M.S.; Sánchez-Hernández, F.; Moreno-García, M.N. Selecting the suitable resampling strategy for imbalanced data classification regarding dataset properties: An approach based on association models. Appl. Sci. 2021, 11, 8546. [Google Scholar] [CrossRef]
  93. Yang, Y.; Khorshidi, H.A.; Aickelin, U. A review on over-sampling techniques in classification of multi-class imbalanced datasets: Insights for medical problems. Front. Digit. Health 2024, 6, 1430245. [Google Scholar] [CrossRef] [PubMed]
  94. Kaur, P.; Gosain, A. Comparing the behavior of oversampling and undersampling approach of class imbalance learning by combining class imbalance problem with noise. In ICT Based Innovations; Springer: Singapore, 2017. [Google Scholar] [CrossRef]
  95. Ichihashi, H.; Honda, K.; Notsu, A.; Miyamoto, E. FCM classifier for high-dimensional data. In Proceedings of the IEEE World Congress on Computational Intelligence (FUZZ-IEEE), Hong Kong, China, 1–6 June 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 200–206. [Google Scholar] [CrossRef][Green Version]
  96. Pezoulas, V.C.; Zaridis, D.I.; Mylona, E.; Androutsos, C.; Apostolidis, K.; Tachos, N.S.; Fotiadis, D.I. Synthetic data generation methods in healthcare: A review on open-source tools and methods. Comput. Struct. Biotechnol. J. 2024, 23, 2892–2910. [Google Scholar] [CrossRef] [PubMed]
  97. Turlapati, V.P.K.; Prusty, M.R. Outlier-SMOTE: A refined oversampling technique for improved detection of COVID-19. Intell.-Based Med. 2020, 3, 100023. [Google Scholar] [CrossRef]
  98. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A new over-sampling method in imbalanced data sets learning. In Intelligent Computing; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar] [CrossRef]
  99. Zhuang, D.; Zhang, B.; Yang, Q.; Yan, J.; Chen, Z.; Chen, Y. Efficient text classification by weighted proximal SVM. In Proceedings of the IEEE International Conference on Data Mining (ICDM), Houston, TX, USA, 27–30 November 2005; IEEE: Piscataway, NJ, USA, 2005; p. 8. [Google Scholar] [CrossRef]
  100. Majzoub, H.A.; Elgedawy, I. AB-SMOTE: An affinitive borderline SMOTE approach for imbalanced data binary classification. Int. J. Mach. Learn. Comput. 2020, 10, 31–37. [Google Scholar] [CrossRef]
  101. Göcs, L.; Johanyák, Z.C. Feature selection with weighted ensemble ranking for improved classification performance on the CSE-CIC-IDS2018 dataset. Computers 2023, 12, 147. [Google Scholar] [CrossRef]
  102. Shoohi, L.M.; Saud, J.H. Adaptation proposed methods for handling imbalanced datasets based on over-sampling technique. Al-Mustansiriyah J. Sci. 2020, 31, 25–29. [Google Scholar] [CrossRef]
  103. Chen, Z.; Zhou, L.; Yu, W. ADASYN-Random forest based intrusion detection model. In Proceedings of the 2021 4th International Conference on Signal Processing and Machine Learning, Beijing, China, 18–20 August 2021; ACM: New York, NY, USA, 2021; pp. 152–159. [Google Scholar] [CrossRef]
  104. Mani, I.; Zhang, J. kNN Approach to Unbalanced Data Distributions: A Case Study involving Information Extraction. In Proceedings of the Workshop on Learning from Imbalanced Datasets (ICML), Washington, DC, USA, 21 August 2003; ICML: Washington, DC, USA, 2003; pp. 1–7. [Google Scholar]
  105. Wickramasinghe, I.; Kalutarage, H. Naive Bayes: Applications, variations and vulnerabilities: A review of literature with code snippets for implementation. Soft Comput. 2021, 25, 2277–2293. [Google Scholar] [CrossRef]
  106. Blanquero, R.; Carrizosa, E.; Ramírez-Cobo, P.; Sillero-Denamiel, M.R. Constrained Naïve Bayes with application to unbalanced data classification. Cent. Eur. J. Oper. Res. 2022, 30, 1403–1425. [Google Scholar] [CrossRef]
  107. Gunawan, W.; Devianto, Y.; Sari, A.P. Imbalanced data NearMiss for comparison of SVM and Naive Bayes algorithms. Comput. Eng. Appl. J. 2024, 13, 34–43. [Google Scholar] [CrossRef]
  108. Jafarnejad, A.; Rezasoltani, A.; Khani, A.M. Comparative analysis of machine learning algorithms in predicting jumps in stock closing price: Case study of Iran Khodro using NearMiss and SMOTE approaches. Iran. J. Financ. 2025, 9, 27–54. [Google Scholar]
  109. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B (Methodol.) 1977, 39, 1–22. [Google Scholar] [CrossRef]
  110. Agrawal, A.; Viktor, H.L.; Paquet, E. SCUT: Multi-class imbalanced data classification using SMOTE and cluster-based undersampling. In Proceedings of the International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management (IC3K), Lisbon, Portugal, 12–14 November 2015; IEEE: Piscataway, NJ, USA, 2015; Volume 1, pp. 226–234. [Google Scholar]
  111. Burduk, R. Classification performance metric for imbalance data based on recall and selectivity normalized in class labels. arXiv 2020, arXiv:2006.13319. [Google Scholar] [CrossRef]
  112. Lipton, Z.C.; Elkan, C.; Narayanaswamy, B. Optimal thresholding of classifiers to maximize F1 measure. In Machine Learning and Knowledge Discovery in Databases; Springer: Berlin/Heidelberg, Germany, 2014; pp. 225–239. [Google Scholar] [CrossRef]
  113. Opitz, J.; Burst, S. Macro F1 and micro F1. arXiv 2019, arXiv:1911.03347. [Google Scholar] [CrossRef]
  114. Matthews, B.W. Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochim. Biophys. Acta—Protein Struct. 1975, 405, 442–451. [Google Scholar] [CrossRef]
  115. Ahuja, S.; Panigrahi, B.K.; Dey, N.; Taneja, A.; Gandhi, T.K. McS-Net: Multi-class Siamese network for severity of COVID-19 infection classification from lung CT scan slices. Appl. Soft Comput. 2022, 131, 109683. [Google Scholar] [CrossRef]
  116. Tamura, J.; Itaya, Y.; Hayashi, K.; Yamamoto, K. Statistical inference of the Matthews correlation coefficient for multiclass classification. arXiv 2025, arXiv:2503.06450. [Google Scholar] [CrossRef]
  117. Bradley, A.P. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997, 30, 1145–1159. [Google Scholar] [CrossRef]
  118. Hanley, J.A.; McNeil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef]
  119. Hand, D.J.; Till, R.J. A simple generalisation of the area under the ROC curve for multiple class classification problems. Mach. Learn. 2001, 45, 171–186. [Google Scholar] [CrossRef]
  120. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
Figure 1. Percentage distribution of fetal health classes (normal, suspect, pathological).
Figure 1. Percentage distribution of fetal health classes (normal, suspect, pathological).
Diagnostics 16 00485 g001
Figure 2. Workflow of the proposed CTG-based fetal health classification model.
Figure 2. Workflow of the proposed CTG-based fetal health classification model.
Diagnostics 16 00485 g002
Figure 3. ROC curves for the NB algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Figure 3. ROC curves for the NB algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Diagnostics 16 00485 g003
Figure 4. ROC curves for RF algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Figure 4. ROC curves for RF algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Diagnostics 16 00485 g004
Figure 5. ROC curves for the LDA algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Figure 5. ROC curves for the LDA algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Diagnostics 16 00485 g005
Figure 6. ROC curves for the KNN algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Figure 6. ROC curves for the KNN algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Diagnostics 16 00485 g006
Figure 7. ROC curves for SVM algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Figure 7. ROC curves for SVM algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Diagnostics 16 00485 g007
Figure 8. ROC curves for MLR algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Figure 8. ROC curves for MLR algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Diagnostics 16 00485 g008
Figure 9. ROC curves for MLP algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Figure 9. ROC curves for MLP algorithm under resampling methods on the CTG test set (dashed line: AUC = 0.5).
Diagnostics 16 00485 g009
Table 1. Description of cardiotocogram (CTG) attributes used in the model.
Table 1. Description of cardiotocogram (CTG) attributes used in the model.
SymbolVariable Description
ClassFetal state class: normal; suspect; pathologic
LBFetal heart beats per minute
ACAccelerations per second
FMFetal movements per second
UCNumber of uterine contractions per second
DLNumber of light decelerations per second
DSNumber of severe decelerations per second
DPNumber of prolonged decelerations per second
ASTVPercentage of time with abnormal short-term variability
MSTVMean value of short-term variability
ALTVPercentage of time with abnormal long-term variability
MLTVMean value of long-term variability
WidthWidth of FHR histogram
MaxMaximum of FHR histogram
MinMinimum of FHR histogram
NmaxNumber of histogram peaks
NzerosNumber of histogram zeros
ModeHistogram mode
MeanHistogram mean
MedianHistogram median
VarianceHistogram variance
TendencyHistogram tendency: 1 = left asymmetric; 0 = symmetric; 1 = right asymmetric
Table 2. Comparison of classifier performance across resampling techniques on the CTG dataset.
Table 2. Comparison of classifier performance across resampling techniques on the CTG dataset.
AlgorithmResampling MethodBACCMacro-MCCMacro-F1
NBbase0.81280.64740.7550
SMOTE0.76710.65110.7616
BSMOTE0.79140.63400.7474
ADASYN0.77870.60610.7256
NearMiss0.70830.62100.7288
SCUT0.70130.61820.7276
RFbase0.91180.84770.9003
SMOTE0.88970.84280.8973
BSMOTE0.90920.85330.9073
ADASYN0.88150.84370.8957
NearMiss0.78800.76160.8357
SCUT0.86570.82830.8898
LDAbase0.78740.65400.7626
SMOTE0.70410.64150.7352
BSMOTE0.72090.65740.7490
ADASYN0.64900.58800.6916
NearMiss0.70050.63150.7324
SCUT0.70180.63590.7313
KNNbase0.84730.75740.8346
SMOTE0.84490.78200.8513
BSMOTE0.83750.77130.8413
ADASYN0.83120.76580.8401
NearMiss0.74790.68200.7850
SCUT0.82450.76790.8445
SVMbase0.81900.69460.7835
SMOTE0.75750.70420.7855
BSMOTE0.77910.71500.7932
ADASYN0.70130.65560.7429
NearMiss0.71930.66540.7566
SCUT0.75460.69030.7793
MLRbase0.77420.66830.7699
SMOTE0.74370.69240.7770
BSMOTE0.75330.69080.7801
ADASYN0.69140.63930.7340
NearMiss0.70730.63550.7443
SCUT0.73820.67530.7649
MLPbase0.86360.78490.8560
SMOTE0.83260.78620.8564
BSMOTE0.82300.76100.8391
ADASYN0.80400.75090.8306
NearMiss0.75270.67790.7797
SCUT0.78520.72190.8126
Note: A detailed comparison with existing CTG classification studies is provided in Table A1 (Appendix A).
Table 3. Per-class sensitivity (recall), specificity, and precision (PPV) across classifiers and resampling strategies on the CTG dataset.
Table 3. Per-class sensitivity (recall), specificity, and precision (PPV) across classifiers and resampling strategies on the CTG dataset.
AlgorithmResampling MethodSensitivity/RecallSpecificityPrecision/PPV
Normal Suspect Pathological Normal Suspect Pathological Normal Suspect Pathological
NBbase0.9350.6740.5750.7220.9240.9970.9270.5800.931
SMOTE0.9010.7790.6380.7970.9080.9870.9440.5680.790
BSMOTE0.9210.7210.5530.7290.9190.9930.9280.5790.867
ADASYN0.8970.8020.4680.7220.9040.9930.9250.5660.846
NearMiss0.8340.8840.6600.9020.8510.9750.9700.4810.674
SCUT0.8280.8840.6810.9020.8510.9710.9700.4810.653
RFbase0.9740.7790.9150.8570.9730.9970.9630.8170.956
SMOTE0.9560.8490.9150.9020.9580.9950.9740.7600.935
BSMOTE0.9660.8140.9360.8720.9690.9970.9660.8050.957
ADASYN0.9530.8720.9150.9250.9550.9930.9800.7500.915
NearMiss0.9030.8490.9570.9320.9310.9710.9810.6580.726
SCUT0.9290.8840.9570.9320.9350.9950.9810.6790.938
LDAbase0.9530.5700.7020.6990.9530.9850.9230.6530.786
SMOTE0.8120.9190.7450.9770.8320.9700.9930.4590.660
BSMOTE0.8440.9190.7020.9550.8590.9730.9860.5030.674
ADASYN0.8020.7330.8510.9850.8260.9460.9950.3960.556
NearMiss0.8140.8720.7660.9550.8330.9700.9860.4490.667
SCUT0.8220.8950.7230.9700.8370.9700.9910.4610.654
KNNbase0.9620.6980.8090.7970.9640.9880.9470.7500.844
SMOTE0.9470.8260.8090.8720.9510.9880.9660.7250.844
BSMOTE0.9510.8020.7870.8650.9530.9870.9640.7260.822
ADASYN0.9430.8020.8090.8650.9480.9870.9640.7040.826
NearMiss0.8460.8720.8510.8950.8750.9780.9680.5210.755
SCUT0.9210.8490.8510.8950.9290.9880.9710.6520.851
SVMbase0.9510.6980.6380.7820.9370.9930.9430.6320.882
SMOTE0.8790.8610.7870.9620.8800.9800.9890.5290.755
BSMOTE0.8990.8720.7230.9400.8950.9850.9830.5640.791
ADASYN0.8750.7560.8090.9700.8820.9590.9910.5000.613
NearMiss0.8460.8720.7870.9620.8620.9700.9880.4970.673
SCUT0.8730.8370.7870.9400.8750.9810.9820.5110.771
MLRbase0.9370.6630.7020.7820.9350.9830.9420.6130.767
SMOTE0.8770.8490.7870.9550.8840.9750.9870.5330.712
BSMOTE0.8930.8370.7450.9100.9020.9760.9740.5710.714
ADASYN0.8670.7440.8090.9470.8820.9560.9840.4960.594
NearMiss0.8480.8020.7870.8950.8710.9680.9680.4930.661
SCUT0.8750.8370.7450.9400.8790.9760.9820.5180.714
MLPbase0.9620.7330.8510.8200.9660.9900.9530.7680.870
SMOTE0.9350.8260.8940.8950.9460.9850.9710.7030.824
BSMOTE0.9290.8020.8510.8800.9330.9880.9670.6510.851
ADASYN0.9150.8140.8720.9020.9220.9850.9730.6200.820
NearMiss0.8730.8260.7870.8800.8900.9800.9650.5380.755
SCUT0.9070.7910.8510.8720.9200.9810.9640.6070.784
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hawrami, Z.S.M.; Cengiz, M.A.; Dünder, E. Addressing Class Imbalance in Fetal Health Classification: Rigorous Benchmarking of Multi-Class Resampling Methods on Cardiotocography Data. Diagnostics 2026, 16, 485. https://doi.org/10.3390/diagnostics16030485

AMA Style

Hawrami ZSM, Cengiz MA, Dünder E. Addressing Class Imbalance in Fetal Health Classification: Rigorous Benchmarking of Multi-Class Resampling Methods on Cardiotocography Data. Diagnostics. 2026; 16(3):485. https://doi.org/10.3390/diagnostics16030485

Chicago/Turabian Style

Hawrami, Zainab Subhi Mahmood, Mehmet Ali Cengiz, and Emre Dünder. 2026. "Addressing Class Imbalance in Fetal Health Classification: Rigorous Benchmarking of Multi-Class Resampling Methods on Cardiotocography Data" Diagnostics 16, no. 3: 485. https://doi.org/10.3390/diagnostics16030485

APA Style

Hawrami, Z. S. M., Cengiz, M. A., & Dünder, E. (2026). Addressing Class Imbalance in Fetal Health Classification: Rigorous Benchmarking of Multi-Class Resampling Methods on Cardiotocography Data. Diagnostics, 16(3), 485. https://doi.org/10.3390/diagnostics16030485

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop