Next Article in Journal
Non-Invasive Capnography Versus Pulse Oximetry for Early Detection of Respiratory Depression During Pediatric Procedural Sedation: A Prospective Observational Study
Previous Article in Journal
Clinical and Echocardiographic Factors Influencing Patent Ductus Arteriosus Treatment in Preterm Neonates
 
 
children-logo
Article Menu
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Diagnostic Accuracy of a Machine Learning-Derived Appendicitis Score in Children: A Multicenter Validation Study

by
Emrah Aydın
1,*,
Taha Eren Sarnıç
2,
İnan Utku Türkmen
2,
Narmina Khanmammadova
1,
Ufuk Ateş
3,
Mustafa Onur Öztan
4,
Tamer Sekmenli
5,
Necip Fazıl Aras
6,
Tülin Öztaş
7,
Ali Yalçınkaya
8,
Murat Özbek
8,
Deniz Gökçe
8,
Hatice Sonay Yalçın Cömert
9,
Osman Uzunlu
10,
Aliye Kandırıcı
11,
Nazile Ertürk
12,
Alev Süzen
12,
Fatih Akova
13,
Mehmet Paşaoğlu
13,
Egemen Eroğlu
14,
Gülnur Göllü Bahadır
3,
Ahmet Murat Çakmak
3,
Salim Bilici
7,
Ramazan Karabulut
8,
Mustafa İmamoğlu
9,
Haluk Sarıhan
9 and
Süleyman Cüneyt Karakuş
12
add Show full author list remove Hide full author list
1
Department of Pediatric Surgery, Tekirdağ Namık Kemal University School of Medicine, Tekirdağ 59030, Turkey
2
Center for Applied Data Science, TED University, Ankara 06420, Turkey
3
Department of Pediatric Surgery, Ankara University School of Medicine, Ankara 06100, Turkey
4
Department of Pediatric Surgery, İzmir Katip Çelebi University School of Medicine, İzmir 35620, Turkey
5
Department of Pediatric Surgery, Selçuk University School of Medicine, Konya 42130, Turkey
6
Department of Pediatric Surgery, Yozgat State Hospital, Yozgat 66100, Turkey
7
Department of Pediatric Surgery, Diyarbakır Gazi Yaşargil Training and Research Hospital, Diyarbakır 21090, Turkey
8
Department of Pediatric Surgery, Gazi University School of Medicine, Ankara 06500, Turkey
9
Department of Pediatric Surgery, Karadeniz Teknik University School of Medicine, Trabzon 61080, Turkey
10
Department of Pediatric Surgery, Pamukkale University School of Medicine, Denizli 20160, Turkey
11
Department of Pediatric Surgery, Okmeydanı Prof. Dr. Cemil Taşçıoğlu State Hospital, İstanbul 34384, Turkey
12
Department of Pediatric Surgery, Muğla Sıtkı Koçman University School of Medicine, Muğla 48000, Turkey
13
Department of Pediatric Surgery, Biruni University School of Medicine, İstanbul 34010, Turkey
14
Department of Pediatric Surgery, Amerikan Hospital, İstanbul 34365, Turkey
*
Author to whom correspondence should be addressed.
Children 2025, 12(7), 937; https://doi.org/10.3390/children12070937
Submission received: 19 May 2025 / Revised: 15 June 2025 / Accepted: 11 July 2025 / Published: 16 July 2025
(This article belongs to the Section Global Pediatric Health)

Abstract

Background: Accurate diagnosis of acute appendicitis in children remains challenging due to variable presentations and limitations of existing clinical scoring systems. While machine learning (ML) offers a promising approach to enhance diagnostic precision, most prior studies have been limited by small sample sizes, single-center data, or a lack of external validation. Methods: This prospective, multicenter study included 8586 pediatric patients to develop a machine learning-based diagnostic model using routinely available clinical and hematological parameters. A separate, prospectively collected external validation cohort of 3000 patients was used to assess model performance. The Random Forest algorithm was selected based on its superior performance during model comparison. Diagnostic accuracy, sensitivity, specificity, Area Under Curve (AUC), and calibration metrics were evaluated and compared with traditional scoring systems such as Pediatric Appendicitis Score (PAS), Alvarado, and Appendicitis Inflammatory Response Score (AIRS). Results: The ML model outperformed traditional clinical scores in both development and validation cohorts. In the external validation set, the Random Forest model achieved an AUC of 0.996, accuracy of 0.992, sensitivity of 0.998, and specificity of 0.993. Feature-importance analysis identified white blood cell count, red blood cell count, and mean platelet volume as key predictors. Conclusions: This large, prospectively validated study demonstrates that a machine learning-based scoring system using commonly accessible data can significantly improve the diagnosis of pediatric appendicitis. The model offers high accuracy and clinical interpretability and has the potential to reduce diagnostic delays and unnecessary imaging.

1. Introduction

Acute appendicitis is one of the most common surgical emergencies in childhood, accounting for a significant proportion of abdominal pain cases presented to pediatric emergency departments worldwide. The annual incidence of appendicitis is estimated at approximately 233 per 100,000 individuals, with a lifetime risk ranging from 6.7% to 8.6% [1]. In children, especially those under the age of five, the diagnosis is particularly challenging due to atypical presentations and communication barriers. Despite being a well-recognized entity, timely and accurate diagnosis of appendicitis remains a critical concern, as delays can result in complications such as perforation, abscess formation, and peritonitis [1,2]. Traditional diagnostic tools include a combination of clinical examination, laboratory investigations, and imaging modalities such as ultrasonography or computed tomography (CT). However, reliance on imaging is not without drawbacks—exposure to radiation, operator dependency, and cost being notable limitations [3,4]. Clinical scoring systems such as the Alvarado Score, Pediatric Appendicitis Score (PAS), Appendicitis Inflammatory Response Score (AIRS), and Shera Score have been widely used to aid decision-making but often lack sufficient sensitivity or specificity when applied across varied pediatric populations [1,5,6]. The reported negative appendectomy rates remain as high as 20% in some series, and up to 30–40% of children still present with perforated appendicitis at the time of diagnosis [1,7]. These statistics highlight the persistent need for improved, objective, and reproducible diagnostic methods that can be integrated into real-world clinical workflows.
Recent advances in artificial intelligence (AI), particularly in machine learning (ML), offer promising solutions to address diagnostic uncertainty in acute care settings. Machine learning algorithms excel at detecting complex patterns in large, multidimensional datasets—capabilities that are increasingly leveraged in medical diagnostics, prognosis prediction, image interpretation, and clinical decision support [8]. In the context of emergency medicine, ML has demonstrated utility in diverse domains, including the triage of chest pain, early sepsis detection, and risk stratification in trauma. Applied to appendicitis, ML models have been developed to classify patients based on clinical and biochemical features with encouraging performance metrics, often surpassing traditional scoring systems in sensitivity and specificity [2,9,10]. However, most of these models are limited by small sample sizes, homogenous data sources, or a lack of external validation. Furthermore, there is a growing demand for interpretable, transparent, and clinically deployable ML solutions, especially in pediatric populations where diagnostic thresholds and expressions differ significantly from adults. The integration of AI-based tools into routine practice necessitates rigorous validation across multiple centers and demographic settings to ensure their generalizability and safety.
Complete blood count (CBC) is one of the most ordered, rapid, and cost-effective laboratory tests in emergency settings. Several parameters within the CBC—such as white blood cell count (WBC), neutrophil-to-lymphocyte ratio (NLR), and mean platelet volume (MPV)—have been studied as potential biomarkers for diagnosing appendicitis and its complications. Studies have shown that elevated WBC and NLR are associated with both uncomplicated and perforated appendicitis, while increased MPV has been suggested as a predictor of perforation severity [11,12]. For example, a study evaluating the predictive role of MPV and NLR in appendicitis found that both markers were significantly higher in patients with perforated appendicitis compared to those with phlegmonous or localized peritonitis. Similarly, other research has demonstrated that NLR values above specific cut-offs can effectively differentiate between complicated and uncomplicated appendicitis. Beyond appendicitis, CBC-derived parameters have also been utilized in the diagnostic stratification of diseases such as COVID-19, myocardial infarction, and various malignancies, emphasizing their versatility and clinical value [13,14,15]. Nevertheless, the interpretation of CBC components is often influenced by age, comorbid conditions, and inter-individual variability, underscoring the need for models that can synthesize these variables into reliable diagnostic tools.
Despite the availability of multiple scoring systems and laboratory markers, diagnostic errors and delays in pediatric appendicitis persist. Many existing tools have limited generalizability, are not tailored to pediatric populations, or depend heavily on imaging studies, which may not always be available or advisable. Moreover, machine learning studies in this domain often suffer from methodological weaknesses, including overfitting, lack of calibration analysis, and absence of head-to-head comparisons with established clinical tools [10,16,17]. Importantly, there remains a gap in the literature for validated, externally tested ML-based scoring systems that utilize routinely available clinical and laboratory data to support rapid, non-invasive diagnosis in diverse pediatric populations. In this study, we aimed to address these limitations by externally validating a machine learning-derived scoring system for pediatric appendicitis using a large, multicenter dataset. We hypothesized that by prospectively addressing the limitations of our previously developed and published model and enhancing its predictive performance using a larger and more diverse dataset, the refined machine learning algorithm would demonstrate superior diagnostic accuracy compared to conventional scoring systems and serve as a clinically reliable decision support tool in the evaluation of suspected pediatric appendicitis [18].

2. Materials and Methods

2.1. Study Design and Setting

From 1 January to 31 December 2019, after receiving Institutional Review Board approval (2018.322.IRB2.048) and in accordance with the Declaration of Helsinki, a prospective, multicenter, national cohort study was conducted in 13 tertiary pediatric hospitals across Türkiye. The study aimed to improve and externally validate a previously published machine learning-based scoring system that was originally developed using retrospective data from five centers. The current study adhered to STROBE criteria and was designed to prospectively capture clinical, laboratory and radiological data from pediatric patients presenting with abdominal pain to emergency departments. Data collection was performed in real-time by attending physicians who supervised case recruitment, ensuring the accuracy and integrity of the records. Participation in the study was voluntary and uncompensated. Patients with known comorbidities or who had received antibiotics within the preceding week were excluded from enrollment. All clinical management decisions were made independently by treating physicians without influence from the study team. Final diagnoses were established through histopathological examination of surgical specimens, which served as the gold standard for both confirming appendicitis and classifying its severity. Cases were categorized as non-complicated appendicitis if inflammation was limited, and as complicated appendicitis if gangrenous or perforated features were identified on histopathology.

2.2. Participants

As part of the study protocol, a total of 8586 pediatric patients who presented to emergency departments with abdominal pain were enrolled for model development. Data were collected prospectively at the time of presentation using a standardized digital data entry system designed for uniform and structured input across centers. The dataset included 63 variables comprising demographics, clinical history, physical examination findings, CBC parameters, imaging results, intraoperative and pathology data, and clinical scoring systems such as Alvarado, PAS, AIR, RIPASA, and Lintula. Observations with missing essential variables or biologically implausible outliers (e.g., extreme red cell distribution width values) were excluded, resulting in a cleaned dataset of 8586 observations used to develop the machine learning-based appendicitis prediction model.
Following model development, a separate group of 3036 pediatric patients from the same participating centers was used for external validation. These patients were also evaluated prospectively, independent of the training dataset. Inclusion criteria for both phases were as follows: (1) age between 3 and 18 years, (2) clinical suspicion of acute appendicitis, and (3) availability of complete preoperative clinical and CBC data. Patients were excluded if they had known comorbidities, had taken antibiotics within the previous week, or lacked definitive histopathological results. Clinical management decisions were made independently by healthcare providers. Final diagnoses were determined by histopathological examination, which also served as the reference standard for classification into non-complicated or complicated appendicitis.

2.3. Outcome Definition

The primary outcome was the accurate diagnosis of acute appendicitis, confirmed by histopathological examination of the surgical specimen. Patients were categorized as having acute appendicitis or not based on histological findings. A secondary classification stratified cases into uncomplicated (e.g., phlegmonous appendicitis) and complicated appendicitis (e.g., perforation, gangrene, abscess formation). Negative appendectomy was defined as the removal of a macroscopically normal appendix in the absence of histological signs of inflammation. These definitions were applied consistently across all centers.

2.4. Data Collection and Variables

All clinical and laboratory variables were extracted from electronic health records using a standardized data abstraction protocol. Demographic variables included age and sex. Clinical presentation data encompassed symptoms such as migratory right lower quadrant (RLQ) pain, nausea/vomiting, anorexia, fever, and RLQ tenderness. Laboratory parameters included CBC-derived indices such as white blood cell (WBC) count, neutrophil count, lymphocyte count, platelet count, mean platelet volume (MPV), red cell distribution width (RDW), and hematocrit. From these, derived ratios such as the neutrophil-to-lymphocyte ratio (NLR) and platelet-to-lymphocyte ratio (PLR) were calculated. Laboratory data were collected at the time of presentation, prior to any surgical intervention or antibiotic administration. Variables with more than 15% missing data were excluded, and the remaining missing values were imputed using the median imputation strategy. All numerical variables were standardized to z-scores to facilitate model performance across different centers.

2.5. Model Development and Validation

The original scoring system was developed using supervised machine learning techniques on a cohort of 8586 pediatric patients. A total of 71 different models were generated and compared using performance metrics such as accuracy, sensitivity, specificity, and AUC. The final model selected was a random forest classifier, which demonstrated the highest diagnostic accuracy and generalizability. For the current validation, the model was applied without retraining to an independent multicenter dataset of 3036 pediatric patients, to assess its external validity. Model inputs consisted of the same 12 standardized clinical and laboratory variables used in the original derivation cohort. The performance of the model was evaluated in terms of its ability to distinguish between appendicitis and non-appendicitis cases, as well as between complicated and uncomplicated appendicitis.

2.6. Comparison with Existing Scores

To contextualize the performance of the machine learning-based score, the same validation dataset was analyzed using three commonly used clinical scoring systems: the Alvarado score, the PAS, and the AIRS. Scores were calculated using available clinical and laboratory data, and their diagnostic accuracy was evaluated using ROC analysis. Each score’s performance was compared to that of the machine learning model using AUC, sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV). Comparative analysis was performed to assess whether the new model provided incremental value over traditional scoring systems.

2.7. Statistical Analysis

All statistical analyses were performed using Python 2.7.16. The normality of continuous variables was assessed using the Kolmogorov–Smirnov test. Continuous variables were presented as means and standard deviations, while categorical variables were presented as frequencies and percentages. Model performance was evaluated using metrics including sensitivity, specificity, PPV, NPV, accuracy, and AUC. ROC curves were plotted. Feature importance was analyzed via recursive feature elimination and decision tree visualization. To prevent overfitting, five-fold cross-validation was applied. A p-value of <0.05 was considered statistically significant for all hypothesis tests.

3. Results

3.1. Patient Characteristic

The study cohort comprised 8586 pediatric patients for model development, with a higher rate of male patients (n = 6816, 79.39%). The mean age was 7.5 ± 5.37 years. Despite the dominance of the male gender, no statistically significant correlation was observed between the patient’s diagnosis and gender (p = 0.54). There were 3954 (46.05%) patients diagnosed with appendicitis, of which 3658 (92.51%) were non-complicated cases.
The most common presenting symptoms were abdominal pain, vomiting, nausea, and fever. Among all patients, 4462 (51.96%) had an X-ray scan, and 5071 (59.06%) underwent ultrasonography (USG). Appendicolith was found in 3 patients (0.06%) via X-ray, free air in 1 (0.02%), and air-fluid levels in 36 (0.8%). Among patients who underwent USG, 686 (13.53%) were diagnosed with appendicitis.
The external validation dataset included 3036 patients, with 69.50% male and a mean age of 11.29 ± 3.99 years. The demographic features between the appendicitis and non-appendicitis groups were similar. Among these patients, 94.37% were diagnosed with appendicitis, with 42.40% being non-complicated and 51.97% complicated.

3.2. Model Performance

During model development, white blood cell variables showed no statistically significant difference between appendicitis and non-appendicitis groups (p > 0.05), although higher median values were observed in the appendicitis group (Table 1). Patients’ visualizations of the parameters (e.g., diagonal density plots, scatter plots) revealed overlapping distributions between groups, limiting clear separability. However, trends such as higher MPV and lower PLT in appendicitis patients, as well as extreme high/low MCV values and concentrated high MCHC in the appendicitis group, were observed.
While several models were trained and evaluated, the Random Forest (RF) model demonstrated the highest performance in predicting the presence of appendicitis. In the external validation dataset, the RF model again achieved outstanding results with an AUC of 0.996, accuracy of 0.992, sensitivity of 0.998, and specificity of 0.993 for appendicitis diagnosis (Table 2).
For classification of appendicitis as complicated vs. non-complicated, the RF model maintained high performance with an AUC of 0.995, accuracy of 0.992, sensitivity of 0.993, and specificity of 0.991 (Table 3).

3.3. Comparison with Traditional Scoring Systems

After identifying the most predictive variables, the dataset was analyzed using multiple existing scoring systems including PAS, AIRS, and Alvarado (Table 4). While these traditional systems showed high specificity, their sensitivity and accuracy were considerably lower. Among them, PAS had the highest success rate. In contrast, Random Forest achieved over 95% success across all performance metrics in both development and validation datasets Table 5).

3.4. Calibration and Feature Importance

Although classical statistical significance was limited for many individual variables, several feature trends were observed in model-based and visualization analyses. Density plots indicated weak individual separability, but the machine learning model was able to leverage complex feature interactions. Visual analysis and ranking of predictors in the RF model highlighted the importance of WBC, MPV, PLT, MCHC, and MCV variables for appendicitis prediction. Detailed rankings are provided in the Supplementary Materials.

4. Discussion

In this multicenter, prospective cohort study, we developed and validated a machine learning-based scoring system to improve the diagnostic accuracy of acute appendicitis in pediatric patients. Using a dataset of 8586 patients from 13 tertiary pediatric centers in Türkiye, we trained several machine learning models and identified the RF algorithm as the most effective in distinguishing appendicitis cases. External validation on a separate cohort of 3036 patients confirmed the model’s robust performance, achieving an AUC of 0.996, accuracy of 0.992, sensitivity of 0.998, and a specificity of 0.993.
The results clearly demonstrate the model’s capacity to outperform traditional scoring systems such as PAS, AIRS, and Alvarado. While these clinical tools have been widely used, our findings revealed that they suffered from relatively lower sensitivity and overall accuracy in both model development and validation phases [19,20]. In contrast, the RF model maintained high performance across all evaluation metrics. This reinforces the potential of machine learning-based models to enhance diagnostic precision in conditions where traditional tools may lack generalizability or are insufficiently sensitive.
Feature analysis revealed that while many classical hematological parameters such as white blood cell count and neutrophil percentage showed overlapping distributions between appendicitis and non-appendicitis groups, the machine learning model successfully integrated subtle variations across multiple features. Notably, MPV, PLT, MCHC, and MCV contributed significantly to the model’s predictions. These findings align with prior literature suggesting the diagnostic value of inflammatory and red blood cell indices in appendicitis, although their standalone clinical utility remains limited due to inter-individual variability.
In recent years, several studies have implemented ML techniques to support the diagnosis of acute appendicitis, particularly in pediatric populations [6,19,21,22]. A 2023 systematic review by Lam et al. emphasized that although ML models such as random forest, SVM, and deep learning architectures have demonstrated promising results with AUC values commonly exceeding 0.90, most studies suffer from key limitations such as small sample size, lack of external validation, and inconsistent outcome definitions [10]. For instance, Tamyalew et al. developed an X-ray-based ML model for large bowel obstruction classification using deep convolutional neural networks but lacked a structured clinical input dataset, which limits its comparability to symptom- and CBC-based approaches [23]. Similarly, Yu et al. demonstrated limited standalone performance for CRP, WBC, and procalcitonin in appendicitis, supporting the integration of such features into multidimensional ML models [24].
Notably, recent individual studies have explored ML applications in pediatric appendicitis with varying input features (CBC components, symptoms, ultrasound findings) [20,25]. However, most of these models were evaluated on retrospective data with fewer than 1000 patients and often used single-center records [21]. In contrast, the present study includes over 8500 patients for model development and 3000 for prospective validation across 13 tertiary centers, making it one of the largest and most rigorously validated appendicitis prediction models published to date. Moreover, while previously reported AUCs typically range between 0.85 and 0.94, the random forest model in this study achieved an AUC of 0.996, outperforming even the most recent comparable models in both diagnosis and severity classification. Unlike earlier works that focused narrowly on appendicitis vs. non-appendicitis, the present model also reliably differentiates between complicated and non-complicated appendicitis—a clinical distinction of increasing importance given the growing interest in non-operative management.
The external validation results are particularly noteworthy. Despite being applied to a distinct patient cohort from the same centers and timeframe, the model preserved its diagnostic strength, confirming its consistency and real-world applicability. This is crucial in pediatric populations where atypical symptom presentations and communication limitations frequently lead to diagnostic delays or unnecessary imaging and surgery.
A further strength of the study lies in its scale and design. With over 11,000 patients evaluated across development and validation phases, this is one of the largest prospective pediatric appendicitis modeling efforts to date. Additionally, all data were collected prospectively using a standardized digital infrastructure, minimizing documentation bias and maximizing uniformity. Histopathology was used as the definitive diagnostic standard, further strengthening the reliability of outcomes.
Nevertheless, several limitations must be acknowledged. First, although the external validation set was prospectively collected, it originated from the same institutional network within the same year, which may limit generalizability to other healthcare systems. Second, we did not assess the model’s impact on clinical decision-making, resource utilization, or patient outcomes in real time. These would require prospective interventional studies. Third, while model performance was excellent, the black-box nature of machine learning models, including Random Forest, may limit clinician interpretability, even with feature-importance analysis.
In conclusion, this study demonstrates that a robust, prospectively validated machine learning-based model can significantly improve diagnostic accuracy for pediatric appendicitis. The RF model outperformed existing clinical scores and maintained consistent performance in an external cohort. With further validation and implementation studies, such tools may enhance early diagnosis, reduce reliance on imaging, and improve outcomes in pediatric emergency care.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/children12070937/s1.

Author Contributions

E.A. and İ.U.T. contributed to the study conception and design. Material preparation and data collection were performed by all authors. E.A., İ.U.T., and T.E.S. were responsible for the data analysis. E.A., N.K., İ.U.T., and T.E.S. were responsible for drafting the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of the coordinating center (protocol code 2018.322.IRB2.048; date of approval: 20 December 2018). Ethical approval was also obtained from the local ethics committees of all participating institutions.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent for publication was not required as no identifiable patient data were included.

Data Availability Statement

The data supporting the findings of this study are not publicly available due to privacy and ethical restrictions. However, the algorithm may be made available by the corresponding author upon reasonable request for research purposes.

Acknowledgments

The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AUCArea Under the Curve
AIRSAppendicitis Inflammatory Response Score
CARTClassification and Regression Tree
CBCComplete Blood Count
CTComputed Tomography
LRLogistic Regression
KNNk-th Nearest Neighbor
MLMachine Learning
MPVMean Platelet Volume
MCVMean Corpuscular Volume
MCHCMean Corpuscular Hemoglobin Concentration
NLRNeutrophil-to-Lymphocyte Ratio
PASPediatric Appendicitis Score
PLTPlatelet Count
PLRPlatelet-to-Lymphocyte Ratio
RFRandom Forest
RIPASARaja Isteri Pengiran Anak Saleha Appendicitis Score
ROCReceiver Operating Characteristic
SVMSupport Vector Machine
USGUltrasonography
WBCWhite Blood Cell Count

References

  1. Andersson, R.E. The Magic of an Appendicitis Score. World J. Surg. 2015, 39, 110–111. [Google Scholar] [CrossRef] [PubMed]
  2. Hsieh, C.H.; Lu, R.H.; Lee, N.H.; Chiu, W.T.; Hsu, M.H.; Li, Y.C. Novel Solutions for an Old Disease: Diagnosis of Acute Appendicitis with Random Forest, Support Vector Machines, and Artificial Neural Networks. Surgery 2011, 149, 87–93. [Google Scholar] [CrossRef] [PubMed]
  3. Doria, A.S.; Moineddin, R.; Kellenberger, C.J.; Epelman, M.; Beyene, J.; Schuh, S.; Babyn, P.S.; Dick, P.T. US or CT for Diagnosis of Appendicitis in Children and Adults? A Meta-Analysis. Radiology 2006, 241, 83–94. [Google Scholar] [CrossRef] [PubMed]
  4. Park, J.J.; Kim, K.A.; Nam, Y.; Choi, M.H.; Choi, S.Y.; Rhie, J. Convolutional-Neural-Network-Based Diagnosis of Appendicitis via CT Scans in Patients with Acute Abdominal Pain Presenting in the Emergency Department. Sci. Rep. 2020, 10, 9556. [Google Scholar] [CrossRef] [PubMed]
  5. Samuel, M. Pediatric Appendicitis Score. J. Pediatr. Surg. 2002, 37, 877–881. [Google Scholar] [CrossRef] [PubMed]
  6. Rey, R.; Gualtieri, R.; La Scala, G.; Posfay Barbe, K.M. Artificial Intelligence in the Diagnosis and Management of Appendicitis in Pediatric Departments: A Systematic Review. Eur. J. Pediatr. Surg. 2024, 34, 385–391. [Google Scholar] [CrossRef] [PubMed]
  7. Omari, A.H.; Khammash, M.R.; Qasaimeh, G.R.; Shammari, A.K.; Yaseen, M.K.B.; Hammori, S.K. Acute Appendicitis in the Elderly: Risk Factors for Perforation. World J. Emerg. Surg. 2014, 9, 6. [Google Scholar] [CrossRef] [PubMed]
  8. Doupe, P.; Faghmous, J.; Basu, S. Machine Learning for Health Services Researchers. Value Health 2019, 22, 808–815. [Google Scholar] [CrossRef] [PubMed]
  9. Byun, J.; Park, S.; Hwang, S.M. Diagnostic Algorithm Based on Machine Learning to Predict Complicated Appendicitis in Children Using CT, Laboratory, and Clinical Features. Diagnostics 2023, 13, 923. [Google Scholar] [CrossRef] [PubMed]
  10. Lam, A.; Squires, E.; Tan, S.; Swen, N.J.; Barilla, A.; Kovoor, J.; Gupta, A.; Bacchi, S.; Khurana, S. Artificial Intelligence for Predicting Acute Appendicitis: A Systematic Review. ANZ J. Surg. 2023, 93, 2070–2078. [Google Scholar] [CrossRef] [PubMed]
  11. Kaya, A.; Karaman, K.; Aziret, M.; Ercan, M.; Köse, E.; Kahraman, Y.S.; Karacaer, C. The Role of Hematological Parameters in Distinguishing Acute Appendicitis from Lymphoid Hyperplasia. Ulus. Travma Acil Cerrahi Derg. 2022, 28, 434–439. [Google Scholar] [CrossRef] [PubMed]
  12. Acharya, A.; Markar, S.R.; Ni, M.; Hanna, G.B. Biomarkers of Acute Appendicitis: Systematic Review and Cost–benefit Trade-off Analysis. Surg. Endosc. 2017, 31, 1022–1031. [Google Scholar] [CrossRef] [PubMed]
  13. Yardımcı, S. Neutrophil-Lymphocyte Ratio and Mean Platelet Volume Can Be a Predictor for the Severity of Acute Appendicitis. Turkish J. Trauma. Emerg. Surg. 2015, 22, 163–168. [Google Scholar] [CrossRef] [PubMed]
  14. Hajibandeh, S.; Hajibandeh, S.; Hobbs, N.; Mansour, M. Neutrophil-to-Lymphocyte Ratio Predicts Acute Appendicitis and Distinguishes between Complicated and Uncomplicated Appendicitis: A Systematic Review and Meta-Analysis. Am. J. Surg. 2020, 219, 154–163. [Google Scholar] [CrossRef] [PubMed]
  15. Demirkol, M.E.; Kaya, M.; Kocadağ, D.; Özsarı, E. Prognostic Value of Complete Blood Count Parameters in COVID-19 Patients. Northwest. Med. J. 2022, 2, 94–102. [Google Scholar] [CrossRef]
  16. Demir, Ş.; Mert, M.; Yasin, Y.K.; Kahya, M.O.; Demirtaş, O. Importance of Pediatric Appendicitis Scoring System and Ultrasonography in the Diagnosis of Acute Appendicitis in Children. Forbes J. Med. 2023, 4, 259–264. [Google Scholar] [CrossRef]
  17. Maleki, F.; Ovens, K.; Gupta, R.; Reinhold, C.; Spatz, A.; Forghani, R. Generalizability of Machine Learning Models: Quantitative Evaluation of Three Methodological Pitfalls. Radiol. Artif. Intell. 2023, 5, e220028. [Google Scholar] [CrossRef] [PubMed]
  18. Aydin, E.; Türkmen, İ.U.; Namli, G.; Öztürk, Ç.; Esen, A.B.; Eray, Y.N.; Eroğlu, E.; Akova, F. A Novel and Simple Machine Learning Algorithm for Preoperative Diagnosis of Acute Appendicitis in Children. Pediatr. Surg. Int. 2020, 36, 735–742. [Google Scholar] [CrossRef] [PubMed]
  19. Males, I.; Boban, Z.; Kumric, M.; Vrdoljak, J.; Berkovic, K.; Pogorelic, Z.; Bozic, J. Applying an Explainable Machine Learning Model Might Reduce the Number of Negative Appendectomies in Pediatric Patients with a High Probability of Acute Appendicitis. Sci. Rep. 2024, 14, 12772. [Google Scholar] [CrossRef] [PubMed]
  20. Yazici, H.; Ugurlu, O.; Aygul, Y.; Ugur, M.A.; Sen, Y.K.; Yildirim, M. Predicting Severity of Acute Appendicitis with Machine Learning Methods: A Simple and Promising Approach for Clinicians. BMC Emerg. Med. 2024, 24, 101. [Google Scholar] [CrossRef] [PubMed]
  21. Navaei, M.; Doogchi, Z.; Gholami, F.; Tavakoli, M.K. Leveraging Machine Learning for Pediatric Appendicitis Diagnosis: A Retrospective Study Integrating Clinical, Laboratory, and Imaging Data. Heal. Sci. Reports 2025, 8, e70756. [Google Scholar] [CrossRef] [PubMed]
  22. Erman, A.; Ferreira, J.; Ashour, W.A.; Guadagno, E.; St-Louis, E.; Emil, S.; Cheung, J.; Poenaru, D. Machine-Learning-Assisted Preoperative Prediction of Pediatric Appendicitis Severity. J. Pediatr. Surg. 2025, 60, 162151. [Google Scholar] [CrossRef] [PubMed]
  23. Tamyalew, Y.; Salau, A.O.; Ayalew, A.M. Detection and Classification of Large Bowel Obstruction from X-Ray Images Using Machine Learning Algorithms. Int. J. Imaging Syst. Technol. 2023, 33, 158–174. [Google Scholar] [CrossRef]
  24. Yu, C.W.; Juan, L.I.; Wu, M.H.; Shen, C.J.; Wu, J.Y.; Lee, C.C. Systematic Review and Meta-Analysis of the Diagnostic Accuracy of Procalcitonin, C-Reactive Protein and White Blood Cell Count for Suspected Acute Appendicitis. Br. J. Surg. 2013, 100, 322–329. [Google Scholar] [CrossRef] [PubMed]
  25. Shera, A.H.; Nizami, F.A.; Malik, A.A.; Naikoo, Z.A.; Wani, M.A. Clinical Scoring System for Diagnosis of Acute Appendicitis in Children. Indian J. Pediatr. 2011, 78, 287–290. [Google Scholar] [CrossRef] [PubMed]
Table 1. Comparison of complete blood count parameters between groups.
Table 1. Comparison of complete blood count parameters between groups.
AppendicitisNon-Appendicitisp
Red Blood Cell VariablesHgb12.6 ± 1.4012.6 ± 1.410.9087
Htc37.5 ± 3.9237.6 ± 3.940.3276
RDW13.1 ± 1.8513.2 ± 1.570.0117
MCV83.8 ± 8.9678.5 ± 3.540.0000
MCHC33.6 ± 1.3733.1 ± 1.530.0000
White Blood Cell VariablesWBC12,315 ± 546012,050 ± 54000.1107
Lymphocyte 2710 ± 20372690 ± 20900.8771
Neutrophil7220 ± 57326895 ± 56090.2133
NLR2.47 ± 6.792.34 ± 6.770.4547
Thrombosis VariablesPlatelet295,000 ± 93,305310,000 ± 92,6940.0000
MPV7.9 ± 1.466.84 ± 1.440.0000
PDW18.9 ± 16.018.7 ± 5.020.0000
Hgb: Hemoglobin, Htc: Hematocrit, RDW: Red Cell Distribution Width, MCV: Mean Corpuscular Volume, MCHC: Mean Corpuscular Hemoglobin Concentration, WBC: White Blood Cell Count, Lymphocyte: Lymphocyte Count, Neutrophil: Neutrophil Count, NLR: Neutrophil-to-Lymphocyte Ratio, Platelet: Platelet Count, MPV: Mean Platelet Volume, PDW: Platelet Distribution Width.
Table 2. The performance parameters of scoring systems and the proposed models to diagnose appendicitis in the new dataset.
Table 2. The performance parameters of scoring systems and the proposed models to diagnose appendicitis in the new dataset.
Scoring System/ModelAUCAccuracySensitivitySpecificity
Alvarado0.6320.7440.3260.852
Lintula0.8160.8470.6500.892
PAS0.9260.8430.8720.858
RIPASA0.6830.7580.3210.795
LR0.9860.9750.9880.972
KNN0.9880.9790.9970.963
SVM0.9820.9830.9950.973
CART0.9940.9760.9970.967
RF0.9960.9920.9980.993
PAS: Pediatric Appendicitis Score; RIPASA: Raja Isteri Pengiran Anak Saleha Appendicitis; LR: Logistic Regression; KNN: kth Nearest Neighbor; SVM: Support Vector Machine; CART: Classification and Regression Tree; RF: Random Forest; AUC: Area Under the Receiver Operating Characteristic Curve.
Table 3. The performance parameters of models to diagnose appendicitis.
Table 3. The performance parameters of models to diagnose appendicitis.
ModelAUCAccuracySensitivitySpecificity
KNN0.9790.9460.9760.975
SVM0.9940.9410.9920.976
CART0.9870.9520.9130.962
RF0.9950.9920.9930.991
LR: Logistic Regression; KNN: kth Nearest Neighbor; SVM: Support Vector Machine; CART: Classification and Regression Tree; RF: Random Forest; AUC: Area Under the Receiver Operating Characteristic Curve.
Table 4. The performance parameters of scoring systems and the proposed models to diagnose appendicitis.
Table 4. The performance parameters of scoring systems and the proposed models to diagnose appendicitis.
Scoring System/ModelAUCAccuracySensitivitySpecificity
Alvarado0.7280.8710.4890.967
Lintula0.8630.9250.7600.966
PAS0.9440.9530.9270.960
RIPASA0.7140.8650.4690.967
LR0.9680.9620.9820.948
KNN0.9770.9690.9970.948
SVM0.9740.9670.9970.946
CART0.9820.9680.9940.949
RF0.9840.9680.9910.951
PAS: Pediatric Appendicitis Score; RIPASA: Raja Isteri Pengiran Anak Saleha Appendicitis; LR: Logistic Regression; KNN: kth Nearest Neighbor; SVM: Support Vector Machine; CART: Classification and Regression Tree; RF: Random Forest; AUC: Area Under the Receiver Operating Characteristic Curve.
Table 5. The performance parameters of models to classify appendicitis.
Table 5. The performance parameters of models to classify appendicitis.
ModelAUC (CV)AUCAccuracySensitivitySpecificity
KNN0.9620.9730.9130.9580.911
SVM0.9680.9700.9361.0000.933
CART0.9560.9710.9390.8330.943
RF0.9680.9730.9250.8750.927
LR: Logistic Regression; KNN: kth Nearest Neighbor; SVM: Support Vector Machine; CART: Classification and Regression Tree; RF: Random Forest; AUC: Area Under the Receiver Operating Characteristic Curve; CV: Cross Validation (5×).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aydın, E.; Sarnıç, T.E.; Türkmen, İ.U.; Khanmammadova, N.; Ateş, U.; Öztan, M.O.; Sekmenli, T.; Aras, N.F.; Öztaş, T.; Yalçınkaya, A.; et al. Diagnostic Accuracy of a Machine Learning-Derived Appendicitis Score in Children: A Multicenter Validation Study. Children 2025, 12, 937. https://doi.org/10.3390/children12070937

AMA Style

Aydın E, Sarnıç TE, Türkmen İU, Khanmammadova N, Ateş U, Öztan MO, Sekmenli T, Aras NF, Öztaş T, Yalçınkaya A, et al. Diagnostic Accuracy of a Machine Learning-Derived Appendicitis Score in Children: A Multicenter Validation Study. Children. 2025; 12(7):937. https://doi.org/10.3390/children12070937

Chicago/Turabian Style

Aydın, Emrah, Taha Eren Sarnıç, İnan Utku Türkmen, Narmina Khanmammadova, Ufuk Ateş, Mustafa Onur Öztan, Tamer Sekmenli, Necip Fazıl Aras, Tülin Öztaş, Ali Yalçınkaya, and et al. 2025. "Diagnostic Accuracy of a Machine Learning-Derived Appendicitis Score in Children: A Multicenter Validation Study" Children 12, no. 7: 937. https://doi.org/10.3390/children12070937

APA Style

Aydın, E., Sarnıç, T. E., Türkmen, İ. U., Khanmammadova, N., Ateş, U., Öztan, M. O., Sekmenli, T., Aras, N. F., Öztaş, T., Yalçınkaya, A., Özbek, M., Gökçe, D., Yalçın Cömert, H. S., Uzunlu, O., Kandırıcı, A., Ertürk, N., Süzen, A., Akova, F., Paşaoğlu, M., ... Karakuş, S. C. (2025). Diagnostic Accuracy of a Machine Learning-Derived Appendicitis Score in Children: A Multicenter Validation Study. Children, 12(7), 937. https://doi.org/10.3390/children12070937

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop