Next Article in Journal
Cognitively Inspired Federated Learning Framework for Interpretable and Privacy-Secured EEG Biomarker Prediction of Depression Relapse
Previous Article in Journal
DermaMamba: A Dual-Branch Vision Mamba Architecture with Linear Complexity for Efficient Skin Lesion Classification
Previous Article in Special Issue
A Three-Stage Fusion Neural Network for Predicting the Risk of Root Fracture—A Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance of ChatGPT-4 as an Auxiliary Tool: Evaluation of Accuracy and Repeatability on Orthodontic Radiology Questions

by
Mercedes Morales Morillo
1,
Nerea Iturralde Fernández
2,
Luis Daniel Pellicer Castillo
3,
Ana Suarez
2,
Yolanda Freire
2,* and
Victor Diaz-Flores García
4
1
School for Doctoral Studies and Research, Universidad Europea de Madrid, 28670 Villaviciosa de Odón, Spain
2
Department of Preclinical Dentistry II, Faculty of Biomedical and Health Sciences, Universidad Europea de Madrid, 28670 Villaviciosa de Odón, Spain
3
Department of Health Sciences, Miguel de Cervantes European University of Valladolid, 47012 Valladolid, Spain
4
Department of Preclinical Dentistry I, Faculty of Biomedical and Health Sciences, Universidad Europea de Madrid, 28670 Villaviciosa de Odón, Spain
*
Author to whom correspondence should be addressed.
Bioengineering 2025, 12(10), 1031; https://doi.org/10.3390/bioengineering12101031
Submission received: 26 August 2025 / Revised: 21 September 2025 / Accepted: 25 September 2025 / Published: 26 September 2025

Abstract

Background: Large language models (LLMs) are increasingly considered in dentistry, yet their accuracy in orthodontic radiology remains uncertain. This study evaluated the performance of ChatGPT-4 on questions aligned with current radiology guidelines. Methods: Fifty short, guideline-anchored questions were authored; thirty were pre-selected a priori for their diagnostic relevance. Using the ChatGPT-4 web interface in March 2025, we obtained 30 answers per item (900 in total) across two user accounts and three times of day, each in a new chat with a standardised prompt. Two blinded experts graded all responses on a 3-point scale (0 = incorrect, 1 = partially correct, 2 = correct); disagreements were adjudicated. The primary outcome was strict accuracy (proportion of answers graded 2). Secondary outcomes were partial-credit performance (mean 0–2 score) and inter-rater agreement using multiple coefficients. Results: Strict accuracy was 34.1% (95% CI 31.0–37.2), with wide item-level variability (0–100%). The mean partial-credit score was 1.09/2.00 (median 1.02; IQR 0.53–1.83). Inter-rater agreement was high (percent agreement: 0.938, with coefficients indicating substantial to almost-perfect reliability). Conclusions: In the conditions of this study, ChatGPT-4 demonstrated limited strict accuracy yet substantial reliability in expert grading when applied to orthodontic radiology questions. These findings underline its potential as a complementary educational and decision-support resource while also highlight its present limitations. Its role should remain supportive and informative, never replacing the critical appraisal and professional judgement of the clinician.

1. Introduction

Artificial intelligence (AI) is increasingly embedded in dentistry and, specifically, orthodontics, supporting diagnosis, planning, and patient communication [1]. This field benefits from advanced systems that replicate human processes such as learning, reasoning, and decision-making [2,3]. Large language models (LLMs) are AI systems trained on extensive datasets collected from diverse online sources, enabling them to generate texts, images, and other types of content with realistic and human-like characteristics [4].
Machine learning (ML) is a fundamental branch of AI that enables systems to learn from data, identify patterns, and make predictions without explicit programming. Within ML, deep learning (DL) represents an approach based on deep neural networks, which are capable of autonomously extracting complex features from data [5]. In orthodontics, these technologies have significantly enhanced diagnostic accuracy, treatment planning, and patient management [6,7,8].
The recent incorporation of automated machine learning (AutoML) platforms allows orthodontists to create complex predictive models for decision-making, facilitating the clinical adoption of AI by automating complex steps with the potential to improve the quality and efficiency of orthodontic treatment [9]. One of the most significant advances is the use of automated tools to analyze cephalometric radiographs. Through neural networks, the identification and annotation of cephalometric landmarks have become faster, more accurate, and more consistent–sometimes matching the quality of expert manual tracing. In addition, these tools can be used to predict bone growth, optimize the use of temporary anchorage devices (TADs), and assist in surgical planning for more complex orthodontic treatments [10]. This not only saves time on repetitive tasks but also reduces human error and enhances diagnostic standardization [11,12].
Such innovations are transforming clinical orthodontics by enabling more personalised patient care [13]. Remote monitoring powered by AI is reshaping teleorthodontics, allowing for more continuous and detailed follow-up without requiring the patient’s physical presence in the clinic [14].
In healthcare, models like ChatGPT have also proven useful in medical education and patient communication [15]. These tools are capable of generating evidence-based responses to complex questions in a clear and comprehensive manner [16].
However, the implementation of AI in orthodontics also presents important challenges and limitations. Concerns have been raised regarding its practical applicability in clinical contexts, as well as the ethical implications of deploying LLMs in healthcare settings [15]. Among these ethical concerns is the potential for bias, particularly when training datasets are disproportionately composed of specific demographic groups [17]. These issues highlight the need for further research and the development of ethical frameworks to guide the clinical integration of such technologies [15,18,19].
LLMs such as ChatGPT have shown potential in generating real-time responses, supporting clinical decision-making and patient education [20,21,22]. Prior studies have evaluated ChatGPT’s performance in areas like prosthodontics and oral surgery, identifying both its strengths in clinical explanation and its limitations in accuracy and consistency [20,23]. Nevertheless, the reliability and precision of these responses in the context of orthodontic radiodiagnosis have not yet been extensively investigated.
The aim of this study is to evaluate the diagnostic accuracy and reproducibility of ChatGPT’s responses to open-ended questions concerning orthodontic radiological diagnosis, by comparing them to established clinical standards.

2. Materials and Methods

According to the terms of the Declaration of Helsinki, this research did not require ethical approval as human participants were not involved.

2.1. Study Design

To assess the accuracy and repeatability of the answers provided by ChatGPT-4 to questions regarding orthodontic radiodiagnosis, 50 open-ended short questions were developed by two experienced authors (A.S., Y.F.) with reference to guidelines published by the British Orthodontic Society, the American Academy of Oral and Maxillofacial Radiology, the European academy of Paediatric Dentistry and The European Commission [24,25,26,27]. Two experienced orthodontists (M.M.M. and N.I.F.) independently assessed each question based on relevance for routine diagnosis/planning and coverage of common decision points (e.g., panoramic vs. CBCT indications, skeletally related imaging). The relevance of each question was assessed using a 3-point Likert scale (0 = disagree; 1 = neutral; 2 = agree). Any discrepancies were reviewed by a third expert independently (L.D.P.C.). As a result of this process, 30 questions were selected (Table 1).
Responses were generated in March 2025 using the ChatGPT web application with the “GPT-4” model label (OpenAI) using two accounts (M.M.M.; V.D.G.). To enter the questions, a specific prompt was designed. This specific prompt was designed with two main objectives: first, to simulate a realistic clinical interaction between an orthodontist and a general dentist; and second, to elicit concise and specific responses, thereby enabling a more objective and standardized evaluation by the expert panel. Therefore, the prompt used was as follows: “Imagine you are an orthodontist and I am a general dentist. Please answer the following question: (QUESTION) (only short answer).” (Figure 1) The rationale was to simulate a brief clinician-to-clinician exchange and elicit concise, scorable statements over discursive essays. In order to reduce any possible memory bias, each answer was collected in a new chat and sampled three times of day (morning/afternoon/evening). The responses were catalogued in an Excel© spreadsheet (Microsoft, Redmond, Washington, DC, USA). The answer generation process took place in March 2025.

2.2. Evaluation Criteria

The 900 answers generated were evaluated independently by two authors (N.I, LD.P.C) using a 3-point Likert scale (0 = incorrect; 1 = partially correct or incomplete; 2 = correct) (Table 2) and with access to the different guidelines used. Graders were provided with a written reference key consolidating guideline-anchored expectations per item. Any discrepancies between the two evaluations were resolved by an independent assessment from a third senior experienced author (V.D.G). For each of the 30 questions, the relative frequency (n) and absolute percentage (%) of answers graded as 0 (incorrect), 1 (incomplete or partially correct), and 2 (correct) were described.

2.3. Statistical Analysis

To assess the performance of ChatGPT, the accuracy and repeatability values were calculated. In order to analyze the accuracy of the answers generated by ChatGPT, the proportion of questions to which an answer of grade 2 (correct) was given was calculated for the total answers in the question set and for each individual question. This calculation was based on the Wald binomial method and took into account the 95% confidence interval.
The assessment of repeatability was conducted through concordance analyzes weighted for ordinal categories and multiple repetitions of the gradings given by the experts (including percentage agreement, Brennan and Prediger coefficient, Conger generalized Cohen kappa, Fleiss kappa, Gwet AC, and Krippendorff alpha) along with their corresponding 95% confidence intervals. The estimated coefficients were classified according to the benchmark scale proposed by Gwet in 2014 [28]. All of the statistical analyzes were carried out using a statistical software program (STATA version BE 14; StataCorp, College Station, TX, USA).

3. Results

A total of 900 answers were generated by ChatGPT, providing 30 responses for each of 30 questions. Table 3 shows the absolute and relative frequencies of the expert’s grading for each answer. To illustrate this heterogeneity more clearly, the proportion of correct (grade = 2) answers per question is displayed in Figure 1.
The range of percentage of correct repetitions was from 0% to 100%, depending on the specific question. The overall accuracy of the answers generated by ChatGPT was 34.1%, with a 95% confidence Interval ranging from 31.0% to 37.2%.
Among the questions asked, fifteen exhibited 100% repeatability, yielding the same score across all 30 answers of each of these questions. Four of the questions (7, 8, 13 and 30) indicated a 100% inaccuracy rate, as all answers generated were graded as incorrect. In contrast, questions 10, 11, 12, 15, 20, 21 and 27 exhibited a 100% rate of repeatability and accuracy with all responses correct. A total of 15 questions showed variability in terms of their degree of repeatability.
The results showed a substantial repeatability according to the benchmark scale used: <0.0 Poor, 0.0–0.2 Slight, 0.2–0.4 Fair, 0.4–0.6 Moderate, 0.6–0.8 Substantial, 0.8–1.0 Almost Perfect (Table 4). This overall pattern of performance is summarised in Figure 2.

4. Discussion

According to the results obtained, ChatGPT-4 achieved limited strict accuracy (34.1%) on text-only orthodontic radiology questions, with a significant degree of variability among questions and an assessment reliability rated as substantial to almost perfect. These findings suggest that, although ChatGPT may produce correct responses in certain contexts, question formulation and content domain strongly influence performance.
A pronounced heterogeneity was observed in the percentage of correct answers depending on the specific question. Among the 30 questions, 14 (46.7%) yielded no correct answers, while 10 (33.3%) surpassed a 50% correctness rate. The remaining 6 questions (20.0%) fell within an intermediate accuracy range, with correct responses between 1% and 50%. These results indicate that the model’s performance varied substantially based on the specific content evaluated.
The variability in results across different dental specialties is noteworthy. In this study, ChatGPT achieved an overall accuracy of 34.1% in the field of orthodontic radiodiagnosis. In contrast, in oral surgery, the tool achieved an accuracy of 71.7%, positioning it as a system with remarkable capabilities for information processing and understanding, and with consistency levels ranging from moderate to almost perfect [20]. Contrary to the findings of the present study, ChatGPT’s performance in endodontics has higher accuracy, with a rate of 57.3% and a consistency rate of 85.4% when answering binary questions [29]. In prosthodontics, despite using the same sample size, the model achieved only 25.6% accuracy, suggesting that the accuracy and repeatability of ChatGPT’s responses are influenced by the question format and the technical complexity [23]. While in paediatric dentistry [30], ChatGPT-4 achieved the highest average score (8.08 out of 10), highlighting a stark contrast with our findings. Therefore, it can be observed that different studies have reported varying accuracies depending on the specialty.
Another possible explanation for the discrepancy among studies may lie in the study design, as some studies evaluated only one answer per question. This methodology may constrain the interpretation of the findings, as relying on a single answer increases the risk of bias, particularly relevant given the phenomenon whereby large language models generate outputs that appear coherent and reliable but ultimately lack factual accuracy or clinical relevance [31]. In addition, the literature shows that specific and well-defined questions tend to yield better answers, while more open-ended or clinically demanding questions decrease the quality, a trend also observed in the results of the present study.
Regarding the model assessed, Makrygiannakis et al. [16] found that Microsoft Bing Chat outperformed ChatGPT-4 in orthodontics (7.1 vs. 4.7), followed by Google Bard (4.8) and ChatGPT 3.5 (3.8). These inter-specialty differences may be attributable to factors such as training dataset composition and the varying complexity of paediatric dentistry versus orthodontics.
A relevant point of comparison is the study by Tanaka et al. [10], which specifically assessed ChatGPT-4 in orthodontics, focusing on clear aligners, temporary anchorage devices (TADs), and digital imaging. That study reported that 71.6% of the responses were rated as “very good” and 15.1% as “good,” with median scores of 5.0 across all domains. However, the inter-rater agreement in that study was low (Fleiss’s Kappa = 0.004), suggesting considerable subjectivity. In contrast, our study demonstrated high inter-rater reliability, with 95.6% agreement between the two main reviewers (860 out of 900 responses), requiring a third grader in only 4.4% of cases. Weighted concordance analyses confirmed this consistency: agreement rate 0.938 (95% CI: 0.911–0.965), classified as “almost perfect”; Brennan & Prediger kappa index 0.833 (95% CI: 0.759–0.907); and Cohen/Conger kappa 0.813 (95% CI: 0.713–0.913), both interpreted as “substantial.” These metrics support the confidence in the accuracy of the evaluations and reinforce the reliability of our findings.
When comparing our results to those of Naureen and Kiani [32], who reported 92.6% average accuracy for ChatGPT-4 in orthodontic diagnosis and treatment, our findings reveal substantially lower performance (34.1%). However, their methodology relied on a single response per item evaluated with a five-point qualitative scale, whereas our design incorporated 30 repetitions per question, allowing us to assess not only accuracy but also response variability and model reliability. Cross-study comparisons remain difficult due to methodological differences and the heterogeneity of specialties assessed.
Recent research has begun comparing ChatGPT versions, particularly ChatGPT-4 versus ChatGPT-4o, with the latter showing improved performance in clinical contexts and lower rates of hallucinated responses [33,34]. However, applied to frequently asked orthodontic questions, the responses still fall short of expert-level knowledge [35]. Thus, continued refinement is necessary, particularly through strategies that promote clearer, more contextually grounded answers.
A potential limitation of this study is the use of a directed prompt. It has been observed that the prompt employed can influence the responses generated. Therefore, the use of a prompt oriented toward a specific answer could have conditioned the results obtained. For this reason, it is essential to conduct further studies analyzing the performance of ChatGPT in different contexts and with diverse prompts, which would allow for a broader exploration of the model’s capabilities and limitations [36].
It is also important to recognise the limitations identified in this study. In addition to the influence of question complexity on performance and reliability, the restricted number of items assessed may not capture the full range of orthodontic scenarios, and the simulated setting with standardised prompts may differ from real clinical interactions. Although inter-rater agreement was high, some degree of subjectivity in expert grading cannot be entirely excluded, and the model’s outputs remain sensitive to prompt formulation. Finally, as ChatGPT-4 itself is not specifically trained on orthodontic datasets, its accuracy and applicability to current clinical practice may be limited.
Considering the rapid evolution of these tools and their considerable potential, it is crucial to enhance their alignment with clinical reasoning and specialty-specific expertise. Current evidence does not support using these tools as a substitute for professional judgement. Furthermore, future studies should investigate the potential of ChatGPT as an auxiliary diagnostic tool in orthodontic radiology, employing robust and clinically realistic protocols. It is also important to recognise the limitations identified in this study, particularly the impact of question complexity on the model’s performance and reliability.

5. Conclusions

In the conditions of this study, ChatGPT-4 demonstrated limited strict accuracy yet substantial reliability in expert grading when applied to orthodontic radiology questions. These findings underline the model’s potential as a complementary educational and decision-support resource, while highlighting its present limitations. At this stage, its role should be regarded as supportive and informative, offering auxiliary guidance that may stimulate reflection or learning but never replacing the critical appraisal and professional judgement of the clinician.

Author Contributions

Conceptualization, M.M.M. and V.D.-F.G.; methodology, A.S., Y.F. and V.D.-F.G.; software, M.M.M. and V.D.-F.G.; validation, Y.F., A.S. and N.I.F.; formal analysis, M.M.M. and V.D.-F.G.; investigation, A.S., Y.F., N.I.F., L.D.P.C., M.M.M. and V.D.-F.G.; resources, M.M.M.; data curation, L.D.P.C. and N.I.F.; writing—original draft preparation, M.M.M. and V.D.-F.G.; writing—review and editing, A.S., Y.F., M.M.M. and V.D.-F.G.; visualization, N.I.F. and L.D.P.C.; supervision, V.D.-F.G.; project administration, V.D.-F.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

According to the terms of the Declaration of Helsinki, this research did not require ethical approval as human participants were not involved.

Informed Consent Statement

Not applicable.

Data Availability Statement

The relevant information for this article is available on the Open Science Framework repository (https://osf.io/k5tcp/?view_only=7fb04cfd0ca84b718cfadb2fb9c5cbcf, accessed on 20 September 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
LLMsLarge language models
MLMachine Learning
DLDeep learning
AutoMLAutomated Machine Learning
TADsTemporary Anchorage Devices

References

  1. Liu, J.; Zhang, C.; Shan, Z. Application of Artificial Intelligence in Orthodontics: Current State and Future Perspectives. Healthcare 2023, 11, 2760. [Google Scholar] [CrossRef]
  2. Bichu, Y.M.; Hansa, I.; Bichu, A.Y.; Premjani, P.; Flores-Mir, C.; Vaid, N.R. Applications of Artificial Intelligence and Machine Learning in Orthodontics: A Scoping Review. Prog. Orthod. 2021, 22, 18. [Google Scholar] [CrossRef] [PubMed]
  3. Lu, W.; Yu, X.; Li, Y.; Cao, Y.; Chen, Y.; Hua, F. Artificial Intelligence–Related Dental Research: Bibliometric and Altmetric Analysis. Int. Dent. J. 2025, 75, 166–175. [Google Scholar] [CrossRef]
  4. Thirunavukarasu, A.J.; Hassan, R.; Mahmood, S.; Sanghera, R.; Barzangi, K.; El Mukashfi, M.; Shah, S. Trialling a Large Language Model (ChatGPT) in General Practice with the Applied Knowledge Test: Observational Study Demonstrating Opportunities and Limitations in Primary Care. JMIR Med. Educ. 2023, 9, e46599. [Google Scholar] [CrossRef]
  5. Kazimierczak, N.; Kazimierczak, W.; Serafin, Z.; Nowicki, P.; Nożewski, J.; Janiszewska-Olszowska, J. AI in Orthodontics: Revolutionizing Diagnostics and Treatment Planning—A Comprehensive Review. J. Clin. Med. 2024, 13, 344. [Google Scholar] [CrossRef] [PubMed]
  6. Eggmann, F.; Weiger, R.; Zitzmann, N.U.; Blatz, M.B. Implications of Large Language Models Such as ChatGPT for Dental Medicine. J. Esthet. Restor. Dent. 2023, 35, 1098–1102. [Google Scholar] [CrossRef]
  7. Khurana, S.; Vaddi, A. ChatGPT From the Perspective of an Academic Oral and Maxillofacial Radiologist. Cureus 2023, 15, e40053. [Google Scholar] [CrossRef]
  8. Alkhamees, A. Evaluation of Artificial Intelligence as a Search Tool for Patients: Can ChatGPT-4 Provide Accurate Evidence-Based Orthodontic-Related Information? Cureus 2024, 16, e65820. [Google Scholar] [CrossRef]
  9. Shujaat, S. Automated Machine Learning in Dentistry: A Narrative Review of Applications, Challenges, and Future Directions. Diagnostics 2025, 15, 273. [Google Scholar] [CrossRef]
  10. Tanaka, O.M.; Gasparello, G.G.; Hartmann, G.C.; Casagrande, F.A.; Pithon, M.M. Assessing the Reliability of ChatGPT: A Content Analysis of Self-Generated and Self-Answered Questions on Clear Aligners, TADs and Digital Imaging. Dent. Press J. Orthod. 2023, 28, e2323183. [Google Scholar] [CrossRef] [PubMed]
  11. Ryu, J.; Kim, Y.-H.; Kim, T.-W.; Jung, S.-K. Evaluation of Artificial Intelligence Model for Crowding Categorization and Extraction Diagnosis Using Intraoral Photographs. Sci. Rep. 2023, 13, 5177. [Google Scholar] [CrossRef]
  12. Ye, H.; Cheng, Z.; Ungvijanpunya, N.; Chen, W.; Cao, L.; Gou, Y. Is Automatic Cephalometric Software Using Artificial Intelligence Better than Orthodontist Experts in Landmark Identification? BMC Oral Health 2023, 23, 467. [Google Scholar] [CrossRef]
  13. Albalawi, F.; Abalkhail, K. Trends and Application of Artificial Intelligence Technology in Orthodontic Diagnosis and Treatment Planning—A Review. Appl. Sci. 2022, 12, 11864. [Google Scholar] [CrossRef]
  14. Surovková, J.; Haluzová, S.; Strunga, M.; Urban, R.; Lifková, M.; Thurzo, A. The New Role of the Dental Assistant and Nurse in the Age of Advanced Artificial Intelligence in Telehealth Orthodontic Care with Dental Monitoring: Preliminary Report. Appl. Sci. 2023, 13, 5212. [Google Scholar] [CrossRef]
  15. Skryd, A.; Lawrence, K. ChatGPT as a Tool for Medical Education and Clinical Decision-Making on the Wards: Case Study. JMIR Form. Res. 2024, 8, e51346. [Google Scholar] [CrossRef] [PubMed]
  16. Makrygiannakis, M.A.; Giannakopoulos, K.; Kaklamanos, E.G. Evidence-Based Potential of Generative Artificial Intelligence Large Language Models in Orthodontics: A Comparative Study of ChatGPT, Google Bard, and Microsoft Bing. Eur. J. Orthod. 2024, 46, cjae017. [Google Scholar] [CrossRef]
  17. Chakraborty, C.; Pal, S.; Bhattacharya, M.; Dash, S.; Lee, S.-S. Overview of Chatbots with Special Emphasis on Artificial Intelligence-Enabled ChatGPT in Medical Science. Front. Artif. Intell. 2023, 6, 1237704. [Google Scholar] [CrossRef]
  18. De Angelis, L.; Baglivo, F.; Arzilli, G.; Privitera, G.P.; Ferragina, P.; Tozzi, A.E.; Rizzo, C. ChatGPT and the Rise of Large Language Models: The New AI-Driven Infodemic Threat in Public Health. Front. Public Health 2023, 11, 1166120. [Google Scholar] [CrossRef] [PubMed]
  19. Hulsen, T. Artificial Intelligence in Healthcare: ChatGPT and Beyond. AI 2024, 5, 550–554. [Google Scholar] [CrossRef]
  20. Suárez, A.; Jiménez, J.; Llorente De Pedro, M.; Andreu-Vázquez, C.; Díaz-Flores García, V.; Gómez Sánchez, M.; Freire, Y. Beyond the Scalpel: Assessing ChatGPT’s Potential as an Auxiliary Intelligent Virtual Assistant in Oral Surgery. Comput. Struct. Biotechnol. J. 2024, 24, 46–52. [Google Scholar] [CrossRef]
  21. Ryu, J.; Lee, Y.-S.; Mo, S.-P.; Lim, K.; Jung, S.-K.; Kim, T.-W. Application of Deep Learning Artificial Intelligence Technique to the Classification of Clinical Orthodontic Photos. BMC Oral Health 2022, 22, 454. [Google Scholar] [CrossRef]
  22. Alessandri-Bonetti, A.; Sangalli, L.; Salerno, M.; Gallenzi, P. Reliability of Artificial Intelligence-Assisted Cephalometric Analysis. A Pilot Study. BioMedInformatics 2023, 3, 44–53. [Google Scholar] [CrossRef]
  23. Freire, Y.; Santamaría Laorden, A.; Orejas Pérez, J.; Gómez Sánchez, M.; Díaz-Flores García, V.; Suárez, A. ChatGPT Performance in Prosthodontics: Assessment of Accuracy and Repeatability in Answer Generation. J. Prosthet. Dent. 2024, 131, 659.e1–659.e6. [Google Scholar] [CrossRef]
  24. Isaacson, K.G.; Thom, A.R.; Atack, N.E.; Horner, K.; Whaites, E. Guidelines for the Use of Radiographs in Clinical Orthodontics, 4th ed.; British Orthodontic Society: London, UK, 2015; ISBN 1-899297-09-X. Available online: https://bos.org.uk/wp-content/uploads/2022/03/Orthodontic-Radiographs-2016-2.pdf (accessed on 25 May 2025).
  25. Kühnisch, J.; Anttonen, V.; Duggal, M.S.; Spyridonos, M.L.; Rajasekharan, S.; Sobczak, M.; Stratigaki, E.; Van Acker, J.W.G.; Aps, J.K.M.; Horner, K.; et al. Best Clinical Practice Guidance for Prescribing Dental Radiographs in Children and Adolescents: An EAPD Policy Document. Eur. Arch. Paediatr. Dent. 2020, 21, 375–386. [Google Scholar] [CrossRef]
  26. European Commission. European Guidelines on Radiation Protection in Dental Radiology: The Safe Use of Radiographs in Dental Practice; European Commission, Ed.; Publications Office: Luxembourg, 2004.
  27. Etherington, G.; Bérard, P.; Blanchardon, E.; Breustedt, B.; Castellani, C.M.; Challeton-de Vathaire, C.; Giussani, A.; Franck, D.; Lopez, M.A.; Marsh, J.W.; et al. Technical recommendations for monitoring individuals for occupational intakes of radionuclides. Radiat. Prot. Dosim. 2016, 170, 8–12. [Google Scholar] [CrossRef] [PubMed]
  28. Gwet, K.L. Handbook of Inter-Rater Reliability: The Definitive Guide to Measuring the Extent of Agreement Among Raters. Volume 1: Analysis of Categorical Ratings, 5th ed.; AgreeStat Analytics: Gaithersburg, MD, USA, 2021; ISBN 978-1-7923-5463-2. [Google Scholar]
  29. Suárez, A.; Díaz-Flores García, V.; Algar, J.; Gómez Sánchez, M.; Llorente de Pedro, M.; Freire, Y. Unveiling the ChatGPT Phenomenon: Evaluating the Consistency and Accuracy of Endodontic Question Answers. Int. Endod. J. 2024, 57, 108–113. [Google Scholar] [CrossRef] [PubMed]
  30. Dermata, A.; Arhakis, A.; Makrygiannakis, M.A.; Giannakopoulos, K.; Kaklamanos, E.G. Evaluating the Evidence-Based Potential of Six Large Language Models in Paediatric Dentistry: A Comparative Study on Generative Artificial Intelligence. Eur. Arch. Paediatr. Dent. 2025, 26, 527–535. [Google Scholar] [CrossRef] [PubMed]
  31. Gajjar, K.; Balakumaran, K.; Kim, A.S. Reversible Left Ventricular Systolic Dysfunction Secondary to Pazopanib. Cureus 2018, 10, e3517. [Google Scholar] [CrossRef]
  32. Naureen, S.; Kiani, H. Assessing the Accuracy of AI Models in Orthodontic Knowledge: A Comparative Study Between ChatGPT-4 and Google Bard. J. Coll. Physicians Surg. Pak. 2024, 34, 761–766. [Google Scholar] [CrossRef]
  33. Shahriar, S.; Lund, B.D.; Mannuru, N.R.; Arshad, M.A.; Hayawi, K.; Bevara, R.V.K.; Mannuru, A.; Batool, L. Putting GPT-4o to the Sword: A Comprehensive Evaluation of Language, Vision, Speech, and Multimodal Proficiency. Appl. Sci. 2024, 14, 7782. [Google Scholar] [CrossRef]
  34. Luo, D.; Liu, M.; Yu, R.; Liu, Y.; Jiang, W.; Fan, Q.; Kuang, N.; Gao, Q.; Yin, T.; Zheng, Z. Evaluating the Performance of GPT-3.5, GPT-4, and GPT-4o in the Chinese National Medical Licensing Examination. Sci. Rep. 2025, 15, 14119. [Google Scholar] [CrossRef] [PubMed]
  35. Salmanpour, F.; Camcı, H.; Geniş, Ö. Comparative Analysis of AI Chatbot (ChatGPT-4.0 and Microsoft Copilot) and Expert Responses to Common Orthodontic Questions: Patient and Orthodontist Evaluations. BMC Oral Health 2025, 25, 896. [Google Scholar] [CrossRef] [PubMed]
  36. Freire, Y.; Santamaría Laorden, A.; Orejas Pérez, J.; Ortiz Collado, I.; Gómez Sánchez, M.; Thuissard Vasallo, I.J.; Díaz-Flores García, V.; Suárez, A. Evaluating the influence of prompt formulation on the reliability and repeatability of ChatGPT in implant-supported prostheses. PLoS ONE 2025, 20, e0323086. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Proportion of correct (grade = 2) answers for each of the 30 orthodontic radiology questions (n = 30 repetitions per item; total 900 answers).
Figure 1. Proportion of correct (grade = 2) answers for each of the 30 orthodontic radiology questions (n = 30 repetitions per item; total 900 answers).
Bioengineering 12 01031 g001
Figure 2. Overall distribution of expert grades across 900 ChatGPT-4 answers (0 = incorrect; 1 = partially correct; 2 = correct).
Figure 2. Overall distribution of expert grades across 900 ChatGPT-4 answers (0 = incorrect; 1 = partially correct; 2 = correct).
Bioengineering 12 01031 g002
Table 1. Questions selected for answer generation by ChatGPT-4.
Table 1. Questions selected for answer generation by ChatGPT-4.
Question NumberQuestion Description
1What is the function of panoramic radiography in orthodontics?
2Do we need to take an orthopantomography before the clinical examination for a correct diagnosis in orthodontics?
3Do we need for a correct diagnosis in orthodontics to take a CBCT before the clinical examination?
4In what type of radiographs can we be able to assess the state of skeletal maturation of the patient?
5What radiological tests would help in the diagnosis and treatment of impacted teeth?
6In patients with mixed dentition, is it indicated to routinely take a panoramic radiograph to assess tooth replacement?
7What radiographs should be taken at the end of an orthodontic treatment?
8Is it indicated to take post-treatment x-rays at the end of orthodontic cases?
9If third molars need to be extracted for orthodontic treatment, which radiodiagnostic test is the most indicated prior to extraction?
10Is pre-treatment lateral skull teleradiography indicated in patients under 10 years of age with class III?
11Is pre-treatment lateral skull teleradiography indicated in patients under 10 years of age with Class II?
12Is it appropriate to perform lateral skull radiography in patients between 10 and 18 years of age who are about to start treatment with functional appliances?
13Is it indicated to take radiographic records in a patient older than 10 years if the canines have not erupted, but by palpation we can intuit that they are in a favourable position?
14Is it considered necessary to take an X-ray in a patient over 14 years of age who has not erupted the second permanent molars?
15Is a lateral skull X-ray indicated before starting treatment in patients over 18 years of age with biprotrusion?
16How should we position the patient to take a correct lateral skull teleradiography?
17How should we prepare the patient to take a correct panoramic radiograph?
18How should we instruct the patient to bite in order to take a panoramic X-ray correctly?
19In a patient with symptoms associated with temporomandibular dysfunction, is it necessary to take an orthopantomography for a correct diagnosis?
20Is it correct to perform a CBCT in patients with temporomandibular dysfunction if we want to know the position of the articular disc?
21What test should we perform to assess the condition of the articular disc of the temporomandibular joint?
22What should the decision to perform a CBCT on an orthodontic patient be based on?
23What are the indications for CBCT in orthodontics?
24In which cases is it justified to perform a CBCT on an orthodontic patient?
25Can a cephalometric study be performed using the image obtained from a CBCT?
26What is the recommended field of view (FOV) if it is necessary to perform a CBCT on an orthodontic patient?
27What radiological test could you perform to measure bone volume when placing a mini-screw (TAD)?
28Is the use of CBCT routinely indicated in case planning where TADS are necessary?
29When would a CBCT be justified when extracting a supernumerary tooth?
30In which situation would a CBCT be indicated and justified for the evaluation of an impacted tooth?
Table 2. Grading System used for ChatGPT Answers.
Table 2. Grading System used for ChatGPT Answers.
GradingGrading Description
Incorrect (0)The answer provided is completely incorrect or unrelated to the question. It does not demonstrate an adequate understanding or knowledge of the topic.
Partially correct or incomplete (1)The answer shows some understanding or knowledge of the topic, but there are significant errors or missing elements. Although not completely incorrect, the answer is not sufficiently correct or complete to be considered certain or adequate.
Correct (2)The answer is completely accurate and shows a solid and precise understanding of the subject. All major components are addressed in a thorough and accurate manner.
Table 3. Distribution of experts grading for ChatGPT answers.
Table 3. Distribution of experts grading for ChatGPT answers.
IncorrectPartially Correct or IncompleteCorrect
QuestionnPercentage (%)nPercentage (%)nPercentage (%)
11756.671343.3300.00
200.002273.33826.67
326.671136.671756.67
400.001240.001860.005
500.002376.67723.33
62066.671033.3300.00
730100.0000.0000.00
830100.0000.0000.00
900.00310.002790.00
1000.0000.0030100.00
1100.0000.0030100.00
1200.0000.0030100.00
1330100.0000.0000.00
1400.0030100.0000,00
1500.0000.0030100.00
16516.672583.3300.00
1700.0030100.0000.00
182996.6700.0013.33
19413.332686.6700.00
2000.0000.0030100.00
2100.0000.0030100.00
22620.002480.0000.00
2300.001963.331136.67
2400.002376.67723.33
2500.0030100.0000.00
262066.671033.3300.00
2700.0000.0030100.00
2800.0030100.0000.00
2900.002996.6713.33
3030100.0000.0000.00
Table 4. Repeatability assessment for 30 repetitions of 30 questions generated by ChatGPT, based on expert grading.
Table 4. Repeatability assessment for 30 repetitions of 30 questions generated by ChatGPT, based on expert grading.
MethodsCoeficient95% Confidence IntervalBenchmark Scale
Percent Agreement0.9380.911–0.965Almost perfect
Brennan and Prediger0.8330.759–0.907Substantial
Cohen/Conger’s Kappa0.8130.713–0.913Substantial
Scott/Fleiss’ Kappa0.8130.768–0.909Substantial
Gwet’s AC0.8380.713–0.913Substantial
Krippendorff’s Alpha0.8130.911–0.965Substantial
Benchmark scale used: <0.0 Poor, 0.0–0.2 Slight, 0.2–0.4 Fair, 0.4–0.6 Moderate, 0.6–0.8 Substantial, 0.8–1.0 Almost Perfect.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Morales Morillo, M.; Iturralde Fernández, N.; Pellicer Castillo, L.D.; Suarez, A.; Freire, Y.; Diaz-Flores García, V. Performance of ChatGPT-4 as an Auxiliary Tool: Evaluation of Accuracy and Repeatability on Orthodontic Radiology Questions. Bioengineering 2025, 12, 1031. https://doi.org/10.3390/bioengineering12101031

AMA Style

Morales Morillo M, Iturralde Fernández N, Pellicer Castillo LD, Suarez A, Freire Y, Diaz-Flores García V. Performance of ChatGPT-4 as an Auxiliary Tool: Evaluation of Accuracy and Repeatability on Orthodontic Radiology Questions. Bioengineering. 2025; 12(10):1031. https://doi.org/10.3390/bioengineering12101031

Chicago/Turabian Style

Morales Morillo, Mercedes, Nerea Iturralde Fernández, Luis Daniel Pellicer Castillo, Ana Suarez, Yolanda Freire, and Victor Diaz-Flores García. 2025. "Performance of ChatGPT-4 as an Auxiliary Tool: Evaluation of Accuracy and Repeatability on Orthodontic Radiology Questions" Bioengineering 12, no. 10: 1031. https://doi.org/10.3390/bioengineering12101031

APA Style

Morales Morillo, M., Iturralde Fernández, N., Pellicer Castillo, L. D., Suarez, A., Freire, Y., & Diaz-Flores García, V. (2025). Performance of ChatGPT-4 as an Auxiliary Tool: Evaluation of Accuracy and Repeatability on Orthodontic Radiology Questions. Bioengineering, 12(10), 1031. https://doi.org/10.3390/bioengineering12101031

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop