Next Article in Journal
Three-Dimensional Dental Analysis in Subjects with Skeletal Malocclusion: A Retrospective Observational Study
Previous Article in Journal
Contemporary Magnetic Removable Partial Denture Utilizing a Novel Ultra-Thin Magnetic Attachment System
Previous Article in Special Issue
Evaluation of Large Language Model Performance in Answering Clinical Questions on Periodontal Furcation Defect Management
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluation of Chatbot Responses to Text-Based Multiple-Choice Questions in Prosthodontic and Restorative Dentistry

by
Reinhard Chun Wang Chau
1,
Khaing Myat Thu
1,
Ollie Yiru Yu
1,
Richard Tai-Chiu Hsung
1,2,
Denny Chon Pei Wang
1,
Manuel Wing Ho Man
1,
John Junwen Wang
1 and
Walter Yu Hang Lam
1,3,*
1
Faculty of Dentistry, The University of Hong Kong, Hong Kong 999077, China
2
Department of Computer Science, Hong Kong Chu Hai College, Hong Kong 999077, China
3
Musketeers Foundation Institute of Data Science, The University of Hong Kong, Hong Kong 999077, China
*
Author to whom correspondence should be addressed.
Dent. J. 2025, 13(7), 279; https://doi.org/10.3390/dj13070279
Submission received: 16 April 2025 / Revised: 3 June 2025 / Accepted: 19 June 2025 / Published: 21 June 2025
(This article belongs to the Special Issue Artificial Intelligence in Oral Rehabilitation)

Abstract

Background/Objectives: This study aims to evaluate the response accuracy and quality of three AI chatbots—GPT-4.0, Claude-2, and Llama-2—in answering multiple-choice questions in prosthodontic and restorative dentistry. Methods: A total of 191 text-based multiple-choice questions were selected from the prosthodontic and restorative dentistry sections of the United States Integrated National Board Dental Examination (INBDE) (n = 80) and the United Kingdom Overseas Registration Examination (ORE) (n = 111). These questions were inputted into the chatbots, and the AI-generated answers were compared with the official answer keys to determine their accuracy. Additionally, two dental specialists independently evaluated the rationales accompanying each chatbot response for accuracy, relevance, and comprehensiveness, categorizing them into four distinct ratings. Chi-square and post hoc Z-tests with Bonferroni adjustment were used to analyze the responses. The inter-rater reliability for evaluating the quality of the rationale ratings among specialists was assessed using Cohen’s Kappa (κ). Results: GPT-4.0 (65.4%; n = 125/191) demonstrated a significantly higher proportion of correctly answered multiple-choice questions when compared to Claude-2 (41.9%; n = 80/191) (p < 0.017) and Llama-2 (26.2%; n = 50/191) (p < 0.017). Significant differences were observed in the answer accuracy among all of the chatbots (p < 0.001). In terms of the rationale quality, GPT-4.0 (58.1%; n = 111/191) had a significantly higher proportion of “Correct Answer, Correct Rationale” than Claude-2 (37.2%; n = 71/191) (p < 0.017) and Llama-2 (24.1%; n = 46/191) (p < 0.017). Significant differences were observed in the rationale quality among all of the chatbots (p < 0.001). The inter-rater reliability was very high (κ = 0.83). Conclusions: GPT-4.0 demonstrated the highest accuracy and quality of reasoning in responding to prosthodontic and restorative dentistry questions. This underscores the varying efficacy of AI chatbots within specialized dental contexts.

1. Introduction

The advent of artificial intelligence (AI) has revolutionized various sectors, including healthcare and education [1,2], with significant advancements in Generative AI models such as ChatGPT (OpenAI, San Francisco, CA, USA), Claude-2 (Anthropic, San Francisco, CA, USA), and Llama-2 (Meta, Menlo Park, CA, USA). These models are renowned for their ability to generate complex, human-like responses [3,4]. GPT-4.0, a sophisticated multimodal model, excels in natural language understanding and generation, and is utilized in applications ranging from education to technical support, reportedly performing at a level comparable to the top 10% of professional and academic benchmarks [5]. Claude-2 is designed with a focus on safety, featuring a large context window of up to 100,000 tokens and strong reasoning abilities, making it suitable for complex tasks like those in healthcare and education [6]. Llama-2 is an open-source model, optimized for dialogue with an emphasis on safety, offering flexibility for both research and commercial applications [7].
In dentistry, AI chatbots have emerged as promising tools for enhancing patient communication and assisting in clinical decision-making [5,6,7]. These chatbots leverage advanced language models to interpret and generate human-like text, offering potential benefits in patient interactions and educational settings.
Dental licensing examinations, such as the Integrated National Board Dental Examination (INBDE [8]) in the United States (US) and the Overseas Registration Examination (ORE [9]) in the United Kingdom (UK), serve as essential benchmarks for evaluating the knowledge and competency of aspiring dentists. These examinations encompass a broad range of dental disciplines, with prosthodontic and restorative dentistry being the central component of the dental curriculum [10,11]. The integrity and reliability of these exams are paramount, as they ensure that licensed dentists possess the requisite expertise to provide quality daily patient care.
Despite the growing integration of AI in dental education and practice, from diagnostics and treatment planning to treatment provision [12,13], research investigating the proficiency of AI models in dental knowledge remains limited, especially regarding their proficiency in dental knowledge and the rationales they use to reach conclusions in dentistry. A previous study reported that some versions of AI chatbots could pass general licensing examinations (multiple choice) [14]; a few other studies reported that some of the AI chatbots could respond with relatively accurate answers for yes/no and multiple-choice questions regarding pediatric dentistry and endodontics [15,16,17]. Various studies have also reported the results in orthodontics [18]. However, their performance in prosthodontic and restorative dentistry, especially the correctness of their rationales for multiple-choice answers, remains unclear. Understanding the capabilities and limitations of the latest AI models in this context is crucial for educators, practitioners, and policymakers aiming to harness these technologies effectively.
This study aims to bridge this knowledge gap by evaluating the accuracy of three prominent AI chatbots—GPT-4.0, Claude-2, and Llama-2—in answering questions from prosthodontic and restorative dentistry sections of the INBDE and ORE. The null hypothesis posits that there are no significant differences in the correctness of multiple-choice answers and the rationales provided by these AI chatbots. By evaluating the performance of these chatbots, this research seeks to inform the potential role of AI in prosthodontic and restorative dentistry.

2. Materials and Methods

2.1. Study Design and Question Selection

This study employed a quantitative, comparative design to evaluate the performance of three AI chatbots in answering prosthodontic and restorative dentistry questions from dental licensing exams from the United States and the United Kingdom. These countries collectively represent 34% of the top 100 dental schools, according to the latest Quacquarelli Symonds (QS) ranking for 2024 [19].
A total of 191 multiple-choice questions were obtained from the top-selling INBDE Book [20] on the US online bookstore Amazon [21], and from the two books [22,23] in the ORE series, which are available through major online bookstores, including Book Depository in the UK [24]. These questions encompassed all the available and relevant questions from these sources, including 80 prosthodontic questions from the INBDE [8] and 111 restorative dentistry questions from the ORE [9]. To ensure consistency in the evaluation by the AI chatbots, questions that contained figures or tables were excluded to eliminate additional variables that might require interpretation.
The questions were input by an independent evaluator (RCWC) in the exact format as they appeared in the books between 29 January 2024 and 4 March 2024, with INBDE questions being input from 29 January 2024 to 8 February 2024, and ORE questions from 26 February 2024 to 4 March 2024.

2.2. Chatbot Selection and Configuration

Three AI chatbots were selected for evaluation: GPT-4.0 (OpenAI, San Francisco, CA, USA), Claude-2 (Anthropic, San Francisco, CA, USA), and Llama-2 (Meta, Menlo Park, CA, USA). These chatbots were selected due to their prominence and widespread recognition as leading models during the study period [25,26]. Each chatbot was accessed using default settings through a paid application program interface (API) service to ensure consistency in response generation [27]. The chatbots were instructed to provide answers with rationale, and no additional prompts were instructed to minimize variability. Each examination was answered by each AI chatbot in a separate chat session.

2.3. Evaluation of the Accuracy of AI-Generated Multiple-Choice Answers

The responses generated by the AI chatbots were collected and evaluated. The accuracy of the responses was determined by comparing them with the official multiple-choice answer keys of the INBDE and ORE. Additionally, each response was accompanied by a rationale generated by the respective AI chatbot, which was subsequently evaluated by a panel of two calibrated UK-trained dental specialists, A and B (DCPW and MWHM).

2.4. Evaluation of the Quality of Rationale Accompanying the Multiple-Choice Answers

The panel employed a standardized rubric to evaluate each rationale as Correct, Partly correct, or Wrong based on the following criteria [28,29]:
  • Accuracy: Whether the rationale is correct.
  • Relevance and comprehensiveness: Whether the rationale directly relates to the question and fully addresses all necessary elements to support the answer.
Responses were categorized and rated as follows:
  • Correct Answer, Correct Rationale: The AI chatbot answered accurately and considered all factors correctly, making its response both relevant and comprehensive.
  • Correct Answer, Partly Correct/Wrong Rationale: The AI chatbot answered accurately but did not consider all factors correctly, making its rationale either partly or completely irrelevant and/or incomplete.
  • Wrong Answer, Correct Rationale: The AI chatbot answered incorrectly but considered all factors correctly, making its rationale relevant and comprehensive.
  • Wrong Answer, Partly Correct/Wrong Rationale: The AI chatbot answered incorrectly and did not consider all factors correctly, making its rationale either partly or completely irrelevant and/or incomplete.

2.5. Statistical Analysis

Data were analyzed using IBM SPSS Statistics Version 29 (IBM, Armonk, NY, USA). The performance of the three AI chatbots—GPT-4.0, Claude-2, and Llama-2—was evaluated regarding answer accuracy and rationale quality. Descriptive statistics were used to summarize the performance metrics of each AI chatbot.
Chi-square tests of homogeneity were used to determine whether there were significant differences in the distribution of answer accuracy and rationale quality ratings across the three AI chatbots. Post hoc pairwise Z-tests with Bonferroni adjustments were performed to identify differences between chatbots. The significance level was adjusted to α = 0.017 (0.05/3) to control for Type I error across the three comparisons.
Inter-rater reliability between Specialist A and B was computed using Cohen’s Kappa (κ) to ensure consistency and agreement in evaluating rationale quality ratings.

3. Results

3.1. Accuracy of AI-Generated Multiple-Choice Answers

The Chi-square tests of homogeneity revealed significant differences in the answer accuracy among the chatbots, χ2(2, N = 573) = 60.416, p < 0.001 (Table 1).
GPT-4.0 demonstrated the highest accuracy, with 65.4% (n = 125/191) of its responses classified as “Correct Answer” and 34.6% (n = 66/191) as “Wrong Answer”. GPT-4.0’s accuracy is significantly higher than that of Claude-2 and Llama-2 (p < 0.017).
Claude-2 exhibited a lower accuracy, with 41.9% (n = 80/191) of its responses rated as “Correct Answer” and 58.1% (n = 111/191) as “Wrong Answer”. Claude-2’s accuracy is significantly lower compared to that of GPT-4.0 but higher than that of Llama-2 (p < 0.017).
Llama-2 showed the lowest accuracy, with only 26.2% (n = 50/191) of its responses classified as “Correct Answer” and 73.8% (n = 141/191) as “Wrong Answer”. Llama-2’s performance is significantly poorer compared to both GPT-4.0 and Claude-2 (p < 0.017) in the studied parameter.

3.2. Quality of Rationale Accompanying the Multiple-Choice Answers

The Chi-square tests of homogeneity indicated significant differences in the rationale quality ratings among the chatbots, χ2(6, N = 573) = 71.776, p < 0.001 (Table 2).
GPT-4.0 demonstrated the highest proficiency in delivering high-quality rationales, having 58.1% (n = 111/191) of its responses rated “Correct Answer, Correct Rationale”. There were significant differences between this category of GPT-4.0 and those of Claude-2 (p < 0.017) and Llama-2 (p < 0.017). This showcased GPT-4.0’s ability to provide accurate answers and quality rationales.
Claude-2 demonstrated a lower proficiency in delivering high-quality rationales, having 37.2% (n = 71/191) of its responses rated “Correct Answer, Correct Rationale”. In addition, there was a significant difference between Claude-2 and Llama-2 in “Correct Answer, Correct Rationale” (p < 0.017), indicating that while Claude-2 did not perform as well as GPT-4.0, it still had a higher ability to provide accurate answers and quality rationales compared to that of Llama-2.
Llama-2 exhibited the lowest performance in providing high-quality rationales, having only 24.1% (n = 46/191) of its responses rated “Correct Answer, Correct Rationale”. It also had 63.4% (n = 121/191) of its responses rated under the “Wrong Answer, Partly Correct/Wrong Rationale” category, indicating a significantly worse performance (p < 0.017) than the other chatbots. This highlights Llama-2’s need for improvement in generating accurate answers and quality rationales.
The results are illustrated in Figure 1.

3.3. Inter-Rater Reliability

The overall inter-rater reliability (Cohen’s Kappa, κ) between Specialist A and Specialist B was computed, resulting in a value of κ = 0.83, indicating almost perfect (range 0.81–0.99) agreement [30]. For GPT-4.0, the Cohen’s Kappa (κ) between the two specialists was 0.96, indicating almost perfect agreement. For Claude-2, the Cohen’s Kappa (κ) between the two specialists was 0.82, indicating almost perfect agreement [30]. For Llama-2, the Cohen’s Kappa (κ) between the two specialists was 0.72, indicating substantial agreement. These Cohen’s Kappa values affirmed the reliability of the ratings of the chatbots’ rationale quality.

4. Discussion

AI has transformative potential across various domains of dentistry. In diagnostics, AI algorithms can enhance the accuracy of radiographic interpretations and disease screening [31,32], enabling the early detection of possible oral pathologies that may be overlooked by the human eye. Treatment planning can be personalized through AI-driven predictive models that anticipate patient-specific outcomes, optimizing restorative procedures [12,33]. The rapid development of AI chatbots presents significant promise for medical education and patient management, particularly in providing accessible learning resources and enhancing patient communication. The accuracy and quality of the generated answers are of the utmost importance for these applications. As the consistency of chatbots in answering prosthodontic and restorative dentistry questions has been investigated [34,35], this study primarily aimed to evaluate the efficacy of three prominent AI chatbots—GPT-4.0, Claude-2, and Llama-2—in accurately addressing prosthodontic and restorative dentistry questions from dental licensing exams. Via a comparative evaluation of the accuracy and rationale quality of the AI chatbots in addressing prosthodontic and restorative dentistry multiple-choice questions from two dental licensing exams, this study aimed to shed light on the complexity involved in selecting a multiple-choice answer and understanding the rationale behind it.
The null hypothesis that there were no significant differences in the correctness of the multiple-choice answers and the rationales provided by these AI chatbots has been rejected, as a performance difference has been observed. GPT-4.0 exhibited the highest accuracy in answering the multiple-choice questions among the evaluated chatbots, correctly answering approximately 74% of NBDE and 60% of ORE questions, making it a likely preference for real-life applications. While these figures demonstrate a promising capability, particularly for GPT-4.0, the performance of Claude-2 and Llama-2 was significantly lower, with Llama-2 showing significant limitations in accuracy. This suggests that although advanced language models like GPT-4.0 are approaching the proficiency level of a qualified dentist and can serve as supplementary educational and clinical tools, there remains substantial room for improvement, especially in the accuracy and quality of responses from other advanced language models.
The evaluation of rationales provided by the chatbots revealed a similar trend. The findings indicate that GPT-4.0 significantly outperforms Claude-2 and Llama-2 in providing high-quality rationales for multiple-choice questions for prosthodontic and restorative dentistry. These differences are statistically significant, with a p-value of less than 0.001 and an effect size of 0.250. The ability to generate correct answers along with high-quality explanatory content is crucial for educational purposes and effective patient communication [36]. Furthermore, the results of this study align with previous research investigating the performance of AI chatbots across various areas of dentistry [15,17,37].
Several factors may contribute to the varying performance levels observed among the chatbots. The underlying training data, model architecture, and fine-tuning processes are pivotal in determining a chatbot’s proficiency in specialized domains like prosthodontic and restorative dentistry. GPT-4.0’s superior performance could be attributed to its more extensive training data and advanced algorithms, which enhance its contextual understanding and reasoning capabilities [38,39]. On the other hand, it is crucial to note that while GPT-4.0 performed better than its counterparts, it is still not perfect and is subject to errors and inaccuracies [16,26].
This study has limitations. The methodology did not assess response consistency across repeated trials, specific question types, or chatbot training data, thereby limiting insights into their full capabilities. Given the rapid evolution of AI chatbots, new models or updates may have emerged since the study’s completion (March 2024), necessitating updated research to remain relevant. Although the consistency of chatbots were not investigated in this study—areas addressed by previous research [34,35]—future studies should examine the consistency of new chatbot models alongside their accuracy and underlying rationale. This is especially important because new models are continually emerging, and there may be a lack of studies investigating their consistency.
AI-driven educational tools can revolutionize oral health education by offering interactive simulations and personalized feedback, potentially improving patient outcomes [40]. However, challenges like data privacy, bias mitigation, and AI consistency must be addressed [33,41]. Future research should include a broader range of questions, additional chatbot models, and response consistency assessments to identify improvement areas. Qualitative analyses should also evaluate the suitability of AI-generated information for patient communication.

5. Conclusions

This study provides a comparative evaluation of the accuracy and quality of AI chatbots in addressing prosthodontic and restorative dentistry questions from dental licensing examinations, based on a single round of evaluation. GPT-4.0 demonstrated the highest accuracy and rationale quality among the evaluated models, highlighting its potential as a valuable educational tool in the near future. Continued advancements in AI technology, combined with rigorous evaluation frameworks that include assessments of the response consistency, will be essential in realizing the full potential of chatbots in the dental field. AI chatbots may make mistakes, and it is imperative to verify important information.

Author Contributions

Conceptualization, R.C.W.C., K.M.T., O.Y.Y. and W.Y.H.L.; methodology, R.C.W.C., K.M.T., O.Y.Y. and W.Y.H.L.; validation, R.C.W.C., K.M.T., D.C.P.W., M.W.H.M. and W.Y.H.L.; formal analysis, R.C.W.C. and K.M.T.; investigation, R.C.W.C.; resources, R.C.W.C., K.M.T. and R.T.-C.H.; data curation, R.C.W.C., K.M.T., D.C.P.W. and M.W.H.M.; writing—original draft preparation, R.C.W.C.; writing—review and editing, R.C.W.C., K.M.T., R.T.-C.H., J.J.W. and W.Y.H.L.; visualization, R.C.W.C.; supervision, R.T.-C.H., J.J.W. and W.Y.H.L.; project administration, W.Y.H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to copyright restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
NBDEUnited States National Board Dental Examination
OREOverseas Registration Examination
AIArtificial Intelligence
USUnited States
UKUnited Kingdom
QSQuacquarelli Symonds
OHIOral Health Instructions

References

  1. Saraswat, D.; Bhattacharya, P.; Verma, A.; Prasad, V.K.; Tanwar, S.; Sharma, G.; Bokoro, P.N.; Sharma, R. Explainable AI for healthcare 5.0: Opportunities and challenges. IEEE Access 2022, 10, 84486–84517. [Google Scholar] [CrossRef]
  2. Zhang, K.; Aslan, A.B. AI technologies for education: Recent research & future directions. Comput. Educ. Artif. Intell. 2021, 2, 100025. [Google Scholar]
  3. Biswas, S.S. Role of Chat GPT in public health. Ann. Biomed. Eng. 2023, 51, 868–869. [Google Scholar] [CrossRef] [PubMed]
  4. Ali, S.R.; Dobbs, T.D.; Hutchings, H.A.; Whitaker, I.S. Using ChatGPT to write patient clinic letters. Lancet Digit. Health 2023, 5, e179–e181. [Google Scholar] [CrossRef]
  5. Harris, E. Large Language Models Answer Medical Questions Accurately, but Can’t Match Clinicians’ Knowledge. JAMA 2023, 330, 792–794. [Google Scholar] [CrossRef] [PubMed]
  6. Dashti, M.; Londono, J.; Ghasemi, S.; Moghaddasi, N. How much can we rely on artificial intelligence chatbots such as the ChatGPT software program to assist with scientific writing? J. Prosthet. Dent. 2023, 133, 1082–1088. [Google Scholar] [CrossRef]
  7. The Lancet Regional Health–Europe. Embracing Generative AI in Health Care; Elsevier: Amsterdam, The Netherlands, 2023; Volume 30, p. 100677. [Google Scholar] [CrossRef]
  8. INBDE® History and Purpose. Available online: https://jcnde.ada.org/inbde/inbde-history (accessed on 20 April 2023).
  9. General Dental Council. Overseas Registration Examination Part 1 Results for August 2022; General Dental Council: London, UK, 2022.
  10. McHarg, J.; Kay, E. Designing a dental curriculum for the twenty-first century. Br. Dent. J. 2009, 207, 493–497. [Google Scholar] [CrossRef]
  11. Klineberg, I.; Eckert, S. Functional Occlusion in Restorative Dentistry and Prosthodontics; Elsevier: Amsterdam, The Netherlands, 2015. [Google Scholar]
  12. Khanagar, S.B.; Al-Ehaideb, A.; Vishwanathaiah, S.; Maganur, P.C.; Patil, S.; Naik, S.; Baeshen, H.A.; Sarode, S.S. Scope and performance of artificial intelligence technology in orthodontic diagnosis, treatment planning, and clinical decision-making-a systematic review. J. Dent. Sci. 2021, 16, 482–492. [Google Scholar] [CrossRef]
  13. Thurzo, A.; Strunga, M.; Urban, R.; Surovková, J.; Afrashtehfar, K.I. Impact of artificial intelligence on dental education: A review and guide for curriculum update. Educ. Sci. 2023, 13, 150. [Google Scholar] [CrossRef]
  14. Chau, R.C.W.; Thu, K.M.; Yu, O.Y.; Hsung, R.T.-C.; Lo, E.C.M.; Lam, W.Y.H. Performance of Generative Artificial Intelligence in Dental Licensing Examinations. Int. Dent. J. 2024, 74, 616–621. [Google Scholar] [CrossRef]
  15. Künzle, P.; Paris, S. Performance of large language artificial intelligence models on solving restorative dentistry and endodontics student assessments. Clin. Oral Investig. 2024, 28, 575. [Google Scholar] [CrossRef] [PubMed]
  16. Suárez, A.; Díaz-Flores García, V.; Algar, J.; Gómez Sánchez, M.; Llorente de Pedro, M.; Freire, Y. Unveiling the ChatGPT phenomenon: Evaluating the consistency and accuracy of endodontic question answers. Int. Endod. J. 2024, 57, 108–113. [Google Scholar] [CrossRef]
  17. Rokhshad, R.; Zhang, P.; Mohammad-Rahimi, H.; Pitchika, V.; Entezari, N.; Schwendicke, F. Accuracy and consistency of chatbots versus clinicians for answering pediatric dentistry questions: A pilot study. J. Dent. 2024, 144, 104938. [Google Scholar] [CrossRef] [PubMed]
  18. Tomášik, J.; Zsoldos, M.; Oravcová, Ľ.; Lifková, M.; Pavleová, G.; Strunga, M.; Thurzo, A. AI and Face-Driven Orthodontics: A Scoping Review of Digital Advances in Diagnosis and Treatment Planning. AI 2024, 5, 158–176. [Google Scholar] [CrossRef]
  19. QS World University Rankings by Subject 2023: Dentistry. Available online: https://www.topuniversities.com/university-rankings/university-subject-rankings/2023/dentistry (accessed on 20 April 2023).
  20. Dowd, F.J. Mosby’s Review for the NBDE Part Two; Mosby Elsevier: St. Louis, MO, USA, 2007. [Google Scholar]
  21. Amazon.com: US NBDE Book. Available online: https://www.amazon.com/s?k=US+nbde+books&crid=254FFUWU4IBTB&sprefix=usnbde+books%2Caps%2C456&ref=nb_sb_noss (accessed on 20 April 2023).
  22. Hammond, D. Best of Fives for Dentistry, 3rd ed.; PasTest Ltd.: Knutsford, UK, 2014; 320p. [Google Scholar]
  23. Fan, K.F.M.; Jones, J.; Quinn, B.F.A. MCQs for Dentistry, 3rd ed.; PasTest Ltd.: Knutsford, UK, 2014; 320p. [Google Scholar]
  24. Books by Pastest. Available online: https://www.bookdepository.com/publishers/Pastest (accessed on 20 April 2023).
  25. Pilny, A.; Kelly, M.; Amanda, S.; Moore, K. From manual to machine: Assessing the efficacy of large language models in content analysis. Commun. Res. Rep. 2024, 41, 61–70. [Google Scholar] [CrossRef]
  26. Büttner, M.; Leser, U.; Schneider, L.; Schwendicke, F. Natural Language Processing: Chances and Challenges in Dentistry. J. Dent. 2024, 141, 104796. [Google Scholar] [CrossRef]
  27. Quora. About Poe. Available online: https://poe.com/about (accessed on 8 May 2025).
  28. Lin, P.-Y.; Tsai, Y.-H.; Chen, T.-C.; Hsieh, C.-Y.; Ou, S.-F.; Yang, C.-W.; Liu, C.-H.; Lin, T.-F.; Wang, C.-Y. The virtual assessment in dental education: A narrative review. J. Dent. Sci. 2024, 19, S102–S115. [Google Scholar] [CrossRef]
  29. Giannakopoulos, K.; Kavadella, A.; Aaqel Salim, A.; Stamatopoulos, V.; Kaklamanos, E.G. Evaluation of the performance of generative AI large language models ChatGPT, Google Bard, and Microsoft Bing Chat in supporting evidence-based dentistry: Comparative mixed methods study. J. Med. Internet Res. 2023, 25, e51580. [Google Scholar] [CrossRef]
  30. Viera, A.J.; Garrett, J.M. Understanding interobserver agreement: The kappa statistic. Fam. Med. 2005, 37, 360–363. [Google Scholar]
  31. Kishimoto, T.; Goto, T.; Matsuda, T.; Iwawaki, Y.; Ichikawa, T. Application of artificial intelligence in the dental field: A literature review. J. Prosthodont. Res. 2022, 66, 19–28. [Google Scholar] [CrossRef]
  32. Chau, R.C.W.; Cheng, A.C.C.; Mao, K.; Thu, K.M.; Ling, Z.; Tew, I.M.; Chang, T.H.; Tan, H.J.; McGrath, C.; Lo, W.-L. External Validation of an AI mHealth Tool for Gingivitis Detection among Older Adults at Daycare Centers: A Pilot Study. Int. Dent. J. 2025, 75, 1970–1978. [Google Scholar] [CrossRef] [PubMed]
  33. Shan, T.; Tay, F.; Gu, L. Application of artificial intelligence in dentistry. J. Dent. Res. 2021, 100, 232–244. [Google Scholar] [CrossRef] [PubMed]
  34. Gheisarifar, M.; Shembesh, M.; Koseoglu, M.; Fang, Q.; Afshari, F.S.; Yuan, J.C.-C.; Sukotjo, C. Evaluating the validity and consistency of artificial intelligence chatbots in responding to patients’ frequently asked questions in prosthodontics. J. Prosthet. Dent. 2025. [Google Scholar] [CrossRef] [PubMed]
  35. Ozdemir, Z.M.; Yapici, E. Evaluating the Accuracy, Reliability, Consistency, and Readability of Different Large Language Models in Restorative Dentistry. J. Esthet. Restor. Dent. 2025, 37, 1740–1752. [Google Scholar] [CrossRef]
  36. Khosravi, H.; Shum, S.B.; Chen, G.; Conati, C.; Tsai, Y.-S.; Kay, J.; Knight, S.; Martinez-Maldonado, R.; Sadiq, S.; Gašević, D. Explainable artificial intelligence in education. Comput. Educ. Artif. Intell. 2022, 3, 100074. [Google Scholar] [CrossRef]
  37. Danesh, A.; Pazouki, H.; Danesh, F.; Danesh, A.; Vardar-Sengul, S. Artificial intelligence in dental education: ChatGPT’s performance on the periodontic in-service examination. J. Periodontol. 2024, 95, 682–687. [Google Scholar] [CrossRef]
  38. Liu, J.; Liang, X.; Fang, D.; Zheng, J.; Yin, C.; Xie, H.; Li, Y.; Sun, X.; Tong, Y.; Che, H.; et al. The Diagnostic Ability of GPT-3.5 and GPT-4.0 in Surgery: Comparative Analysis. J. Med. Internet Res. 2024, 26, e54985. [Google Scholar] [CrossRef]
  39. Kipp, M. From GPT-3.5 to GPT-4.o: A Leap in AI’s Medical Exam Performance. Information 2024, 15, 543. [Google Scholar] [CrossRef]
  40. Chau, R.C.-W.; Thu, K.M.; Hsung, R.T.-C.; McGrath, C.; Lam, W.Y.-H. Self-monitoring of Oral Health Using Smartphone Selfie Powered by Artificial Intelligence: Implications for Preventive Dentistry. Oral Health Prev. Dent. 2024, 22, 327–340. [Google Scholar]
  41. Schwendicke, F.a.; Samek, W.; Krois, J. Artificial intelligence in dentistry: Chances and challenges. J. Dent. Res. 2020, 99, 769–774. [Google Scholar] [CrossRef]
Figure 1. Stacked bar count of chatbot by overall rationale scores.
Figure 1. Stacked bar count of chatbot by overall rationale scores.
Dentistry 13 00279 g001
Table 1. Distribution of answer accuracy across chatbots.
Table 1. Distribution of answer accuracy across chatbots.
CHATBOTCORRECT ANSWERWRONG ANSWER
GPT-4.0125/191 (65.4%) a66/191 (34.6%) a
CLAUDE-280/191 (41.9%) b111/191 (58.1%) b
LLAMA-250/191 (26.2%) c141/191 (73.8%) c
Each subscript letter denotes a subset of accuracy rating categories whose row proportions do not differ significantly from each other at the 0.001 level.
Table 2. Distribution of rationale quality across chatbots.
Table 2. Distribution of rationale quality across chatbots.
CHATBOTCORRECT ANSWERWRONG ANSWER
Correct RationalePartly Correct/Wrong RationaleCorrect RationalePartly Correct/Wrong Rationale
GPT-4.0111/191 (58.1%) a14/191 (7.3%) a22/191 (11.5%) a44/191 (23.0%) a
CLAUDE-271/191 (37.2%) b10/191 (5.2%) a,b29/191 (15.2%) a81/191 (42.4%) b
LLAMA-246/191 (24.1%) c4/191 (2.1%) b20/191 (10.5%) a121/191 (63.4%) c
Each subscript letter denotes a subset of accuracy rating categories whose row proportions do not differ significantly from each other at the 0.017 level.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chau, R.C.W.; Thu, K.M.; Yu, O.Y.; Hsung, R.T.-C.; Wang, D.C.P.; Man, M.W.H.; Wang, J.J.; Lam, W.Y.H. Evaluation of Chatbot Responses to Text-Based Multiple-Choice Questions in Prosthodontic and Restorative Dentistry. Dent. J. 2025, 13, 279. https://doi.org/10.3390/dj13070279

AMA Style

Chau RCW, Thu KM, Yu OY, Hsung RT-C, Wang DCP, Man MWH, Wang JJ, Lam WYH. Evaluation of Chatbot Responses to Text-Based Multiple-Choice Questions in Prosthodontic and Restorative Dentistry. Dentistry Journal. 2025; 13(7):279. https://doi.org/10.3390/dj13070279

Chicago/Turabian Style

Chau, Reinhard Chun Wang, Khaing Myat Thu, Ollie Yiru Yu, Richard Tai-Chiu Hsung, Denny Chon Pei Wang, Manuel Wing Ho Man, John Junwen Wang, and Walter Yu Hang Lam. 2025. "Evaluation of Chatbot Responses to Text-Based Multiple-Choice Questions in Prosthodontic and Restorative Dentistry" Dentistry Journal 13, no. 7: 279. https://doi.org/10.3390/dj13070279

APA Style

Chau, R. C. W., Thu, K. M., Yu, O. Y., Hsung, R. T.-C., Wang, D. C. P., Man, M. W. H., Wang, J. J., & Lam, W. Y. H. (2025). Evaluation of Chatbot Responses to Text-Based Multiple-Choice Questions in Prosthodontic and Restorative Dentistry. Dentistry Journal, 13(7), 279. https://doi.org/10.3390/dj13070279

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop