Next Article in Journal
Color Normalization in Breast Cancer Immunohistochemistry Images Based on Sparse Stain Separation and Self-Sparse Fuzzy Clustering
Previous Article in Journal
Prevalence of Tuberculosis in Central Asia and Southern Caucasus: A Systematic Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance of a Vision-Language Model in Detecting Common Dental Conditions on Panoramic Radiographs Using Different Tooth Numbering Systems

1
Oral and Maxillofacial Radiology, Applied Oral Sciences and Community Dental Care, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR 999077, China
2
Department of Diagnostic Radiology, Li Ka Shing Faculty of Medicine, The University of Hong Kong, Hong Kong SAR 999077, China
*
Authors to whom correspondence should be addressed.
Diagnostics 2025, 15(18), 2315; https://doi.org/10.3390/diagnostics15182315
Submission received: 7 August 2025 / Revised: 8 September 2025 / Accepted: 11 September 2025 / Published: 12 September 2025
(This article belongs to the Special Issue Advances in Head and Neck and Oral Maxillofacial Radiology)

Abstract

Objectives: The aim of this study was to evaluate the performance of GPT-4o in identifying nine common dental conditions on panoramic radiographs, both overall and at specific tooth sites, and to assess whether the use of different tooth numbering systems (FDI and Universal) in prompts would affect its diagnostic accuracy. Methods: Fifty panoramic radiographs exhibiting various common dental conditions including missing teeth, impacted teeth, caries, endodontically treated teeth, teeth with restorations, periapical lesions, periodontal bone loss, tooth fractures, cracks, retained roots, dental implants, osteolytic lesions, and osteosclerosis were included. Each image was evaluated twice by GPT-4o in May 2025, using structured prompts based on either the FDI or Universal tooth numbering system, to identify the presence of these conditions at specific tooth sites or regions. GPT-4o responses were compared to a consensus reference standard established by an oral-maxillofacial radiology team. GPT-4o’s performance was evaluated using balanced accuracy, sensitivity, specificity, and F1 score both at the patient and tooth levels. Results: A total of 100 GPT-4o responses were generated. At the patient level, balanced accuracy ranged from 46.25% to 98.83% (FDI) and 49.75% to 92.86% (Universal), with the highest accuracies for dental implants (92.86–98.83%). F1-scores and sensitivities were highest for implants, missing, and impacted teeth, but zero for caries, periapical lesions, and fractures. Specificity was generally high across conditions. Notable discrepancies were observed between patient- and tooth-level performance, especially for implants and restorations. GPT-4o’s performance was similar between using the two numbering systems. Conclusions: GPT-4o demonstrated superior performance in detecting dental implants and treated or restored teeth but inferior performance for caries, periapical lesions, and fractures. Diagnostic accuracy was higher at the patient level than at the tooth level, with similar performances for both numbering systems. Future studies with larger, more diverse datasets and multiple models are needed.

1. Introduction

With the rapid growth of technology, artificial intelligence (AI) models have become more involved in various aspects of daily life, including a wide range of medical and dental applications [1,2]. In dentistry, AI models has been proposed for tasks such as disease detection, segmentation, and prediction [3,4,5]. The advancement of AI research and its applications in these fields has been driven in large part by the development and availability of large, high-quality datasets [6,7,8,9,10,11,12,13,14]. Among recent AI innovations, large language models (LLMs) have emerged as powerful tools capable of understanding, generating, and using human language [15]. They learn to predict the next word in a sequence, allowing them to perform various language-related tasks like answering questions, summarizing text, translating languages, and even generating creative content [16]. Previous studies have introduced a cloud-based intelligent system built on LLMs to help patients with medical care and registration [17]. In addition, some studies have evaluated the performance of LLMs in completing licensing exams or generating radiology reports [18,19]. Several commercially available AI chatbots have shown the potential to perform at levels similar to clinicians, and they are providing valuable information for researchers [8,20].
Chat Generative Pre-trained Transformer (ChatGPT) has become one of the most popularly used LLMs since its release at the end of 2022 [21]. The early GPT versions could only process and interpret information through text and could not directly read images. Starting from GPT-4V, users can upload image files directly to the chat and have them interpreted instantly. Researchers have already evaluated GPT-4V’s performance in many dental fields, such as completing the national dental licensing exam and detecting supernumerary teeth [22,23]. With the release of GPT-4V, it opens up more opportunities for dental AI applications. Officially released on 31 August 2024, GPT-4o has a faster response time, lower resource demand, and better performance in various medical fields compared to GPT-3.5 or, in some cases, even resident physicians [24].
Panoramic radiography has played an important role in dental diagnostics. It provides a comprehensive view of the oral cavity, including teeth and bone structures [25]. Panoramic imaging is not only easily accessible in a wide range of dental settings compared to other advanced radiographic examinations, but it also exposes patients to relatively low levels of ionizing radiation [26]. These advantages make panoramic radiography a highly versatile and widely used imaging method in dental diagnostics [27]. Interpreting panoramic radiographs greatly depends on the clinicians’ experience and can be relatively challenging. Recent systematic reviews have summarized the potential of AI and LLMs in dental and maxillofacial imaging [28,29,30,31,32,33,34]. Previous studies have reported promising results for LLMs in generating radiology reports, detecting supernumerary teeth, and answering radiology-related questions [19,23,35]. However, the performance of GPT-4o in identifying common dental conditions on panoramic radiographs both overall and at specific tooth sites has not yet been evaluated, and the impact of different tooth numbering systems in prompts on its performance remains unknown. Therefore, the objectives of this study were to evaluate the performance of GPT-4o in identifying nine common dental conditions on panoramic radiographs, both overall and at specific tooth sites, and to assess whether the use of different tooth numbering systems (FDI and Universal) in prompts would affect its diagnostic accuracy.

2. Materials and Methods

2.1. Study Design

This observational study was designed to evaluate the performance of AI chatbots in detecting common dental conditions on panoramic radiographs. The study was conducted in full accordance with the Declaration of Helsinki 2013 (www.wma.net). The study protocol was submitted to and approved by the local institutional review board of the University of Hong Kong/Hospital Authority Hong Kong West Cluster (approval number UW 25-270).

2.2. Study Population

The primary objective of this pilot study was to evaluate the performance of GPT-4o in identifying various conditions, including caries, periapical lesions, periodontal bone loss, impacted third molars, restorations, implants, radiopaque, mixed-density, or radiolucent pathological jaw lesions, on panoramic radiographs. To ensure the inclusion of only images presenting the investigated conditions, this pilot study retrospectively retrieved 320 panoramic radiographs from patients who had undergone CBCT examinations between January 2020 and December 2022 in our Diagnostic Imaging Clinic, for dentoalveolar, endodontic, periodontal, dental implant evaluations with clinical records indicating the presence of the investigated conditions. Subsequently, two examiners, a Board-Certified Radiologist (AN) and a Professor in the Oral Radiology Department (KFH), reviewed the 320 panoramic images for inclusion. Panoramic radiographs were excluded if any of the following exclusion criteria were met: (1) patients < 18 years of age; (2) images not presenting at least one of the investigated conditions; (3) images where the presence of the investigated conditions was doubtful and consensus could not be reached between the examiners; or (4) images with severe positioning errors, ghost images, or motion artifacts. Afterwards, a total of 50 panoramic radiographs were included in this pilot study. The number of cases for each condition is provided in Table 1. All panoramic radiographs were acquired using the Veraview X800 (J. Morita, Kyoto, Japan) at the Diagnostic Imaging Clinic at Prince Philip Dental Hospital, Faculty of Dentistry, the University of Hong Kong.

2.3. Prompt Development and Response Generation

This study was conducted using GPT-4o accessed via the University of Hong Kong. Structured prompts were developed to evaluate GPT-4o’s capability in detecting common dental conditions on panoramic radiographs and to assess whether incorporating different tooth numbering systems (FDI or Universal) into the prompts would impact its performance. For each image, two prompts were used with the only difference being the tooth numbering system referenced. Each image was uploaded as a new file, and a new chat topic was initiated for every case to prevent data interruption and potential context contamination. If a response did not specify the exact tooth number, a follow-up question was issued requesting this information. All responses were generated in May 2025. The prompts provided clear instructions to focus solely on visible patterns and to avoid making clinical diagnoses. The prompts were as follows:
  • FDI numbering system prompt:
  • “This is an anonymized dental panoramic radiograph from a research study. You are assisting in identifying observable dental anomalies as listed below, and specify the FDI tooth number of identified items. Focus only on visual patterns, not diagnoses.
  • Developmental anomalies:
  • Hypodontia (missing teeth) or anodontia.
  • Dental anomalies:
  • Caries (radiolucent lesions in enamel/dentin).
  • Periapical lesions (e.g., radiolucency at root apex indicating infection or granuloma).
  • Tooth fractures, cracks, or retained roots.
  • Impacted teeth (e.g., third molars, canines).
  • Bony anomalies:
  • Periodontal bone loss (horizontal/vertical reduction in alveolar bone height).
  • Osteolytic lesions (e.g., cysts, tumors like ameloblastoma or odontogenic keratocyst).
  • Osteosclerosis (abnormal bone density, e.g., condensing osteitis).
  • Iatrogenic/treatment-related findings:
  • Endodontically treated teeth or teeth with restorations (Fillings, Crowns or bridges).
  • Dental implants (screw-shaped radiopaque structures).”
  • Universal numbering system prompt:
  • The second prompt was identical in structure and content, except that it instructed ChatGPT to specify the Universal tooth number for identified findings.

2.4. Data Collection

After collecting all the responses, the data were stored in an Excel spreadsheet (Microsoft, Redmond, DC, USA). In the Excel spreadsheet, all the investigated conditions were listed for each tooth. When GPT-4o indicated that there were conditions and specified the tooth number, a “positive” outcome was recorded for that specific tooth sites. Otherwise, any conditions that were not mentioned in the GPT-4o’s response were recorded as “negative”. Both the response using the FDI and the Universal numbering system were recorded in separate files. A postgraduate dental student in dento-maxillofacial radiology (ZL) independently generated all responses from GPT-4o using the prompts and recorded GPT-4o’s findings. The reference standard for the presence or absence of each dental condition was established by consensus among a postgraduate student, an oral-maxillofacial radiologist, and a professor in the Oral Radiology department. True positives, true negatives, false positives, and false negatives were then determined by comparing GPT’s outputs to this reference standard.
Panoramic images have reported accuracy levels of 70–85% for diagnosing periodontal bone loss [36,37]. In order to evaluate GPT-4o′s ability to correctly identify moderate to severe bone loss at the tooth-specific level, a case was considered positive for “periodontal bone loss” only if the alveolar bone level was located apical to the midpoint of the root.
For the “osteolytic lesions” and “osteosclerosis” conditions, their locations were assessed in six regions, including the anterior maxilla, anterior mandible, left posterior maxilla, right posterior maxilla, left posterior mandible, and right posterior mandible, rather than specific tooth sites.

2.5. Statistical Analysis

All collected data were initially analysed descriptively to summarize the distribution of positive and negative cases in each dental condition category. To provide a more reliable assessment of GPT-4o’s performance, the diagnostic performance of GPT-4o was evaluated for each condition and numbering system (FDI and Universal) using sensitivity, specificity, balanced accuracy (average of sensitivity and specificity), and F1-score (harmonic mean of precision and recall). These evaluation metrics were determined by comparing GPT-4o’s predictions to the reference standard using counts of true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN) for each condition. For tooth-level analysis, each tooth or region was treated as an independent unit. For patient-level analysis, a patient was considered positive for a condition if GPT-4o correctly identified the condition on any tooth. Otherwise, the patient was considered negative. All statistical analyses were performed using SPSS software (Version 29.0; IBM Corp., Armonk, NY, USA).

3. Results

A total of 100 responses were recorded for fifty images. Table 1 shows the count of positive and negative cases for each dental condition at both the patient and tooth levels as identified by the reference standard and by GPT-4o using either the FDI or universal numbering system.
Table 2 demonstrates the diagnostic performance of GPT-4o, including balanced accuracy, sensitivity, specificity, and F1-score for each condition at both the patient and tooth levels.
At the patient level, balanced accuracy ranged from 46.25% to 98.83% with the FDI system and from 49.75% to 92.86% with the universal system. Tooth-level balanced accuracy was generally lower but more consistent, ranging from 49.66% to 65.04% (FDI) and 49.75% to 77.89% (universal). The highest patient-level balanced accuracies were observed for Dental Implants (98.83% FDI, 92.86% universal) and Endodontically Treated Teeth or Teeth with Restorations (80.92% FDI, 79.61% universal). At the tooth level, the highest balanced accuracies were found for Dental Implants (65.04% FDI, 77.89% universal), Impacted Teeth (61.35% FDI, 68.81% universal), and Missing Teeth (61.65% FDI, 63.96% universal).
F1-score showed considerable variability across conditions, with the highest scores at the patient level for Dental Implants (99.33% FDI, 92.31% universal), Endodontically Treated Teeth or Teeth with Restorations (89.19% FDI, 87.67% universal), and Missing Teeth (73.68% FDI, 77.92% universal). At the tooth level, F1-scores were generally lower, with the highest values for Missing Teeth (34.25% FDI, 38.69% universal), Impacted Teeth (49.52% FDI, 45.87% universal), and Dental Implants (28.57% FDI, 48% universal). Notably, several conditions exhibited an F1-score of 0 at both the patient and tooth levels, including Caries, Periapical Lesions, and Tooth Fractures/Cracks/Retained Roots.
Sensitivity at the patient level ranged from 0% to 100% (FDI) and 0% to 85.71% (universal), with the highest values for Dental Implants (100% FDI, 85.71% universal), Endodontically Treated Teeth or Teeth with Restorations (86.84% FDI, 84.21% universal), and Missing Teeth (80% FDI, 85.71% universal). At the tooth level, sensitivity was lower overall, ranging from 0% to 48.15% (FDI) and 0% to 46.30% (universal). The highest tooth-level sensitivities were for Impacted Teeth (48.15% FDI, 46.30% universal), Dental Implants (30.77% FDI, 46.15% universal), and Missing Teeth (28.72% FDI, 33.33% universal).
Specificity was generally high, particularly at the tooth level (mostly above 91%). At the patient level, the highest specificities were for Dental Implants (97.67% FDI, 100% universal), Osteolytic Lesions/Osteosclerosis (97.30% FDI/universal), and Caries (95.35% FDI, 97.67% universal). At the tooth level, Periapical Lesions (99.87% FDI, 100% universal), Caries (99.81% FDI, 99.87% universal), and Tooth Fractures, Cracks, or Retained Roots (99.56% FDI, 99.49% universal) had the highest specificities.
The largest discrepancies in balanced accuracy between patient and tooth levels were observed for Dental Implants (FDI: 33.8%, universal: 14.97%), Endodontically Treated Teeth or Teeth with Restorations (FDI: 20.7%, universal: 21.08%), and Missing Teeth (FDI: 15.0%, universal: 11.1%). In contrast, the smallest differences were found for Periodontal Bone Loss (FDI: 0.17%, universal: 0.59%), Impacted Teeth (FDI: 1.4%, universal: 3.75%), and Caries (FDI: 2.24%, universal: 1.10%).
Figure 1 and Figure 2 exhibit representative cases to show both the input images and GPT-4o’s outputs.

4. Discussion

This study evaluated the performance of GPT-4o in identifying common dental conditions on panoramic radiographs, focusing on diagnostic accuracy both overall and at specific tooth sites. Additionally, it assessed whether using different tooth numbering systems (FDI and Universal) in prompts would affect GPT-4o’s diagnostic performance. The results showed that GPT-4o’s performance changed largely across different conditions and between patient and tooth levels. At the patient level, balanced accuracy ranged from 46.25% to 98.83% with the FDI system and from 49.75% to 92.86% with the universal system, while at the tooth level, balanced accuracy was generally lower but more consistent, ranging from 49.66% to 65.04% (FDI) and 49.75% to 77.89% (universal). Sensitivity demonstrated considerable variability, with higher values for certain conditions at the patient level, but generally lower and less variable results at the tooth level. In contrast, specificity was consistently high across most conditions, particularly at the tooth level, frequently exceeding 91%. F1-scores also reflected these trends, with higher values observed for some conditions at the patient level, but overall lower scores at the tooth level and several conditions exhibiting F1-scores of zero. These findings indicate that while GPT-4o achieved high specificity in excluding negative cases, its ability to consistently and accurately detect positive cases remains highly variable depending on the dental condition.
In this study, GPT-4o demonstrated the highest diagnostic performance for dental implants, endodontically treated teeth or teeth with restorations, and missing teeth, particularly at the patient level. Patient-level balanced accuracy was approximately 95% for dental implants and 80% for endodontically treated or restored teeth. GPT-4o successfully identified these conditions, which have radiopaque appearances and easily recognizable features on panoramic images, in most cases at the patient level. In contrast, GPT-4o’s performance was limited for caries, periapical lesions, and tooth fractures, cracks, or retained roots, with F1-scores and sensitivities of zero at both patient and tooth levels. These findings suggest that at the current stage GPT-4o is more capable of detecting radiopaque dental conditions than radiolucent ones, likely due to the distinct radiographic characteristics that make radiopaque entities more readily identifiable by AI models [38]. Previous studies have shown that AI models achieved high accuracy in detecting dental implants in panoramic radiographs [39,40]. On the contrary, more subtle radiolucent conditions (e.g., caries, periapical lesions, and tooth fractures) are relatively challenging to detect because of their less noticeable radiographic appearance and the limited detail provided by panoramic images. Zadrożny et al. found that while deep learning models obtained high sensitivity for dental fillings, endodontically treated teeth, residual roots, periodontal bone loss, missing teeth, and prosthetic restorations, these models showed low sensitivity for periapical lesions, caries, and over- and underfilled canals [41]. Similarly, Zhu et al. reported that the sensitivity and specificity of deep learning models for detecting caries in panoramic radiographs were not favourable. These findings suggest that AI-assisted diagnosis of subtle radiolucent lesions in panoramic images remains challenging [42].
In this study, large differences were found in GPT-4o’s performance between patient and tooth levels for certain conditions, particularly for dental implants, endodontically treated teeth or teeth with restorations, and missing teeth. For these conditions, the discrepancies in balanced accuracy between patient and tooth levels were considerable. The higher performance at the patient level indicates that GPT-4o could recognize the presence of these conditions in the image. However, it was often unable to accurately localize and assign the findings to the correct tooth sites, resulting in reduced tooth-level accuracy. This limitation in localization may be partially attributed to limitations of panoramic radiographs, such as overlapping anatomical structures, tooth crowding, and image distortion, which could affect GPT-4o’s ability in precise localization of specific teeth [43,44].
It was noted that the majority of the existing teeth misclassified as missing (i.e., false negative) were third molars. This finding suggests that GPT-4o had difficulty in accurately detecting third molars, which might be due to the high variability in their appearance and positioning on panoramic radiographs [45]. Third molars often present with inconsistent radiographic features owing to a wide range of anatomical variations and may be partially erupted, impacted, or superimposed by other anatomical structures, which makes them challenging for AI to identify [46]. Suárez et al. found that ChatGPT-4o frequently provided incomplete or fabricated information in complex third molar cases, particularly those involving overlapping anatomical structures or underdeveloped roots [47]. In this study, GPT-4o was observed to classify around fifteen third molars as both “missing teeth” and “impacted teeth”. This finding indicates that GPT-4o may consider third molars to be missing if they are impacted and not aligned at the same occlusal level as adjacent molars, which suggests a lack of clear understanding regarding the mutually exclusive nature of these conditions for the same tooth. Another study reported that GPT-4 performed poorly in detecting resorptive changes of impacted canines on panoramic radiographs, which suggests that accurately identifying changes in tooth structure changes remains challenging for commercially available LLMs [48].
While LLMs were expected to be able to detect large and well-defined radiopaque or radiolucent jaw lesions, in this study GPT-4o performed poorly in detecting the jaw lesions compared to its performance in identifying relatively less obvious conditions, such as periodontal bone loss. The sensitivity for osteolytic lesions or osteosclerosis was low at both the patient level (7.69%) and sextant level (3.33%) with F1-scores also near zero, which indicates that GPT-4o could not identify almost any true positive cases. This surprising outcome may be partially attributed to the prompt design without clear guidance for lesion detection. Notably, previous studies evaluating commercially available LLMs for radiographic image interpretation have rarely provided detailed descriptions of their prompt design, and those that did typically used a single prompting strategy rather than directly comparing multiple approaches. A recent study by Silva et al. evaluated a GPT model’s performance in describing radiolucent lesions in panoramic radiographs using a series of structured, stepwise prompts [49]. This approach first asked the model to provide a detailed radiographic description (covering internal structure, periphery, location, and effects on adjacent structures), then to offer impressions regarding the lesion’s nature, and finally to generate a list of likely differential diagnoses. Although this structured strategy gave clear direction for identifying both bony and dental anomalies, the reported accuracy in characterizing lesions remained limited, ranging from 25% to 67.85%, highlighting the ongoing challenges in applying LLMs for lesion detection and differential diagnosis in clinical practice. Based on previous studies, our study further assessed the potential of GPT in detecting common dental conditions using checklist-style prompts, which have not yet been investigated in previous studies. The prompting approach used in this study directed the model to search for specific anomalies, focusing mainly on visual patterns without requiring interpretative steps. This alternative approach may provide additional information regarding how prompt design may influence LLM performance in dental radiographic analysis.
Notably, for certain conditions such as caries and periapical lesions, GPT-4o obtained zero sensitivity but maintained high accuracy and specificity at both the patient and tooth levels. This phenomenon indicates that GPT-4o could not identify any true positive cases from the included images, resulting in a great amount of correctly classified negative cases. As the number of positive and negative cases for all conditions were relatively imbalanced at the tooth level, conventional evaluation metrics, such as accuracy and specificity, can be misleading, as high values may simply reflect the predominance of negative cases rather than true diagnostic capability. Instead of these metrics, sensitivity, balanced accuracy, and F1-score could provide a more reliable assessment of AI’s ability to correctly identify positive cases for analysis, with positive cases fewer than negative cases [50]. It is important to mention that GPT-4o was not specifically developed for medical or dental diagnostic purposes and should not be used for clinical decision-making, particularly in dental diagnostics. Our findings clearly demonstrate that GPT-4o could not detect subtle but clinically important radiolucent conditions at both the patient and tooth levels.
This study has several limitations. First, the sample size was relatively small, comprising only fifty images and nine common dental conditions, which may limit the generalizability of the findings. However, it is worth noting that the sample size in this study was greater than that of previous studies assessing the performance of ChatGPT in interpreting panoramic radiographs (which used between 28 and 30 images), and that each of the fifty images in this study was assessed twice using prompts with different tooth numbering systems, which was not evaluated in previous studies [47,49]. Moreover, this study did not evaluate other LLMs as GPT-4o was the only publicly accessible LLM with image reading capabilities (i.e., vision-language models) available during the data collection period. Future studies should include larger and more diverse datasets with a more balanced distribution of positive and negative cases at the tooth level. Additionally, the images selected in this study were deemed to have minimal imaging or patient positioning artifacts to reduce imaging quality, so that GPT-4o’s ability could be assessed from higher quality images. Furthermore, studies specifically designed to evaluate the performance of multiple LLMs across a wider range of dental conditions and image qualities, as well as to compare LLM performance with dental students and practitioners of different qualifications, will be needed to more comprehensively understand the potential and limitations of these models in dental diagnostics.

5. Conclusions

Based on the findings of this study, the following conclusions can be drawn:
  • GPT-4o generally demonstrated low accuracy in identifying many of the investigated radiolucent conditions, with relatively higher accuracy only for easily recognizable radiopaque features such as dental implants, endodontically treated teeth, and teeth with restorations, indicating that it was unable to reliably detect subtle but clinically important radiolucent conditions at both the patient and tooth levels.
  • GPT-4o achieved higher diagnostic performance at the patient level than at the tooth level for dental implants, endodontically treated teeth, and missing teeth, suggesting that while the model could detect the presence of these conditions in the image, it might not accurately localize them.
  • GPT-4o demonstrated similar diagnostic performance when using either the FDI or Universal numbering systems in identifying most dental conditions.
  • Future studies should use larger and more diverse datasets, evaluate the performance of multiple language models across a wider range of dental conditions and image qualities, and include comparisons with dental students and practitioners of varying qualifications to more comprehensively understand the potential and limitations of LLMs in dental diagnostics.

Author Contributions

Conceptualisation: K.F.H., A.N. and Q.Y.H.A.; Methodology: K.F.H., A.N. and Q.Y.H.A.; Data collection: Z.L., K.F.H. and A.N.; Data analysis: Z.L. and K.F.H.; Writing—original draft preparation: Z.L. and K.F.H.; Writing—review and editing: A.N., A.W.K.Y., Q.Y.H.A. and R.T. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the HKU URC Seed Fund for Basic Research for New Staff at the University of Hong Kong (103036002).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of the University of Hong Kong/Hospital Authority Hong Kong West Cluster (UW 25-270; 19 June 2025).

Informed Consent Statement

Patient consent was waived due to the retrospective nature of this study. In addition, patients had already signed and authorized the hospital or the University to use any anonymized photographs, radiographs, images, or study models for education, research, and publication purposes.

Data Availability Statement

Data will be provided by the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial intelligence
LLMLarge language model
ChatGPTChat Generative Pre-trained Transformer

References

  1. Wei, H.; Dai, Y.; Yuan, K.; Li, K.Y.; Hung, K.F.; Hu, E.M.; Lee, A.H.C.; Chang, J.W.W.; Zhang, C.; Li, X. Ai-powered problem- and case-based learning in medical and dental education: A systematic review and meta-analysis. Int. Dent. J. 2025, 75, 100858. [Google Scholar] [CrossRef] [PubMed]
  2. Sadr, S.; Rokhshad, R.; Daghighi, Y.; Golkar, M.; Kheybari, F.T.; Gorjinejad, F.; Kojori, A.M.; Rahimirad, P.; Shobeiri, P.; Mahdian, M.; et al. Deep learning for tooth identification and numbering on dental radiography: A systematic review and meta-analysis. Dentomaxillofacial Radiol. 2024, 53, 5–21. [Google Scholar] [CrossRef]
  3. Hao, J.; Wong, L.M.; Shan, Z.; Ai, Q.Y.H.; Shi, X.; Tsoi, J.K.H.; Hung, K.F. A semi-supervised transformer-based deep learning framework for automated tooth segmentation and identification on panoramic radiographs. Diagnostics 2024, 14, 1948. [Google Scholar] [CrossRef]
  4. Chau, K.K.; Zhu, M.; AlHadidi, A.; Wang, C.; Hung, K.; Wohlgemuth, P.; Lam, W.Y.H.; Liu, W.; Yuan, Y.; Chen, H. A novel ai model for detecting periapical lesion on cbct: Cbct-sam. J. Dent. 2025, 153, 105526. [Google Scholar] [CrossRef]
  5. Khubrani, Y.H.; Thomas, D.; Slator, P.J.; White, R.D.; Farnell, D.J.J. Detection of periodontal bone loss and periodontitis from 2d dental radiographs via machine learning and deep learning: Systematic review employing appraise-ai and meta-analysis. Dentomaxillofacial Radiol. 2025, 54, 89–108. [Google Scholar] [CrossRef]
  6. Hao, J.; Nalley, A.; Yeung, A.W.K.; Tanaka, R.; Ai, Q.Y.H.; Lam, W.Y.H.; Shan, Z.; Leung, Y.Y.; AlHadidi, A.; Bornstein, M.M.; et al. Characteristics, licensing, and ethical considerations of openly accessible oral-maxillofacial imaging datasets: A systematic review. NPJ Digit. Med. 2025, 8, 412. [Google Scholar] [CrossRef] [PubMed]
  7. Wen, D.; Khan, S.M.; Xu, A.J.; Ibrahim, H.; Smith, L.; Caballero, J.; Zepeda, L.; de Blas Perez, C.; Denniston, A.K.; Liu, X.; et al. Characteristics of publicly available skin cancer image datasets: A systematic review. Lancet Digit. Health 2022, 4, e64–e74. [Google Scholar] [CrossRef] [PubMed]
  8. Khan, S.M.; Liu, X.; Nath, S.; Korot, E.; Faes, L.; Wagner, S.K.; Keane, P.A.; Sebire, N.J.; Burton, M.J.; Denniston, A.K. A global review of publicly available datasets for ophthalmological imaging: Barriers to access, usability, and generalisability. Lancet Digit. Health 2021, 3, e51–e66. [Google Scholar] [CrossRef]
  9. Panetta, K.; Rajendran, R.; Ramesh, A.; Rao, S.; Agaian, S. Tufts dental database: A multimodal panoramic x-ray dataset for benchmarking diagnostic systems. IEEE J. Biomed. Health Inform. 2022, 26, 1650–1659. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Ye, F.; Chen, L.; Xu, F.; Chen, X.; Wu, H.; Cao, M.; Li, Y.; Wang, Y.; Huang, X. Children’s dental panoramic radiographs dataset for caries segmentation and dental disease detection. Sci. Data 2023, 10, 380. [Google Scholar] [CrossRef]
  11. Zhou, W.; Lu, X.; Zhao, D.; Jiang, M.; Fan, L.; Zhang, W.; Li, F.; Wang, D.; Yin, W.; Liu, X. A dual-labeled dataset and fusion model for automatic teeth segmentation, numbering, and state assessment on panoramic radiographs. BMC Oral Health 2024, 24, 1201. [Google Scholar] [CrossRef]
  12. Sengupta, N.; Sarode, S.C.; Sarode, G.S.; Ghone, U. Scarcity of publicly available oral cancer image datasets for machine learning research. Oral Oncol. 2022, 126, 105737. [Google Scholar] [CrossRef]
  13. Ribeiro-de-Assis, M.C.F.; Soares, J.P.; de Lima, L.M.; de Barros, L.A.P.; Grão-Velloso, T.R.; Krohling, R.A.; Camisasca, D.R. Ndb-ufes: An oral cancer and leukoplakia dataset composed of histopathological images and patient data. Data Brief 2023, 48, 109128. [Google Scholar] [CrossRef]
  14. Silva, A.B.; Martins, A.S.; Tosta, T.A.A.; Loyola, A.M.; Cardoso, S.V.; Neves, L.A.; de Faria, P.R.; Nascimento, M.Z.D. Oralepitheliumdb: A dataset for oral epithelial dysplasia image segmentation and classification. J. Imaging Inform. Med. 2024, 37, 1691–1710. [Google Scholar] [CrossRef] [PubMed]
  15. Yang, R.; Tan, T.F.; Lu, W.; Thirunavukarasu, A.J.; Ting, D.S.W.; Liu, N. Large language models in health care: Development, applications, and challenges. Health Care Sci. 2023, 2, 255–263. [Google Scholar] [CrossRef]
  16. Bedi, S.; Liu, Y.; Orr-Ewing, L.; Dash, D.; Koyejo, S.; Callahan, A.; Fries, J.A.; Wornow, M.; Swaminathan, A.; Lehmann, L.S.; et al. Testing and evaluation of health care applications of large language models: A systematic review. JAMA 2025, 333, 319–328. [Google Scholar] [CrossRef] [PubMed]
  17. Wang, J.; Zhang, G.; Wang, W.; Zhang, K.; Sheng, Y. Cloud-based intelligent self-diagnosis and department recommendation service using chinese medical bert. J. Cloud Comput. Adv. Syst. Appl. 2021, 10, 4. [Google Scholar] [CrossRef]
  18. Su, M.C.; Lin, L.E.; Lin, L.H.; Chen, Y.C. Assessing question characteristic influences on chatgpt’s performance and response-explanation consistency: Insights from taiwan’s nursing licensing exam. Int. J. Nurs. Stud. 2024, 153, 104717. [Google Scholar] [CrossRef]
  19. Le Guellec, B.; Lefèvre, A.; Geay, C.; Shorten, L.; Bruge, C.; Hacein-Bey, L.; Amouyel, P.; Pruvo, J.-P.; Kuchcinski, G.; Hamroun, A. Performance of an open-source large language model in extracting information from free-text radiology reports. Radiol. Artif. Intell. 2024, 6, e230364. [Google Scholar] [CrossRef]
  20. Hirosawa, T.; Harada, Y.; Yokose, M.; Sakamoto, T.; Kawamura, R.; Shimizu, T. Diagnostic accuracy of differential-diagnosis lists generated by generative pretrained transformer 3 chatbot for clinical vignettes with common chief complaints: A pilot study. Int. J. Environ. Res. Public Health 2023, 20, 3378. [Google Scholar] [CrossRef] [PubMed]
  21. OpenAI. Gpt-4o System Card. 2024. Available online: https://openai.com/index/gpt-4o-system-card/ (accessed on 11 September 2025).
  22. Masaki, M.; Hikaru, F.; Kosuke, M.; Taiji, N.; Masanari, H.; Izumi, Y.; Kentaro, O.; Shuji, A. Evaluating gpt-4v’s performance in the japanese national dental examination: A challenge explored evaluating gpt-4v’s performance in the japanese national dental examination: A challenge explored. J. Dent. Sci. 2024, 19, 1595–1600. [Google Scholar]
  23. Aşar, E.M.; İpek, İ.; Lge, K.B. Customized gpt-4v(ision) for radiographic diagnosis: Can large language model detect supernumerary teeth? BMC Oral Health 2025, 25, 756. [Google Scholar] [CrossRef] [PubMed]
  24. Hoppe, J.M.; Auer, M.K.; Strüven, A.; Massberg, S.; Stremmel, C. Chatgpt with gpt-4 outperforms emergency department physicians in diagnostic accuracy: Retrospective analysis. J. Med. Internet Res. 2024, 26, e56110. [Google Scholar] [CrossRef]
  25. Rushton, M.N.; Rushton, V.E. A study to determine the added value of 740 screening panoramic radiographs compared to intraoral radiography in the management of adult (>18 years) dentate patients in a primary care setting. J. Dent. 2012, 40, 661–669. [Google Scholar] [CrossRef]
  26. Sams, C.M.; Dietsche, E.W.; Swenson, D.W.; DuPont, G.J.; Ayyala, R.S. Pediatric panoramic radiography: Techniques, artifacts, and interpretation. Radiographics 2021, 41, 595–608. [Google Scholar] [CrossRef]
  27. El Maaytah, M.; Jerjes, W.; Swinson, B.; Upile, T.; Thompson, G.; Gittelmon, S.; Baldwin, D.; Hadi, H.; Vourvachis, M.; Abizadeh, N.; et al. Inferior alveolar nerve injury and surgical difficulty prediction in third molar surgery: The role of dental panoramic tomography. Br. J. Oral Maxillofac. Surg. 2007, 45, e10–e11. [Google Scholar] [CrossRef]
  28. Zheng, J.; Ding, X.; Pu, J.J.; Chung, S.M.; Ai, Q.Y.H.; Hung, K.F.; Shan, Z. Unlocking the potentials of large language models in orthodontics: A scoping review. Bioengineering 2024, 11, 1145. [Google Scholar] [CrossRef]
  29. Mao, K.; Wong, L.M.; Zhang, R.; So, T.Y.; Shan, Z.; Hung, K.F.; Ai, Q.Y.H. Radiomics analysis in characterization of salivary gland tumors on mri: A systematic review. Cancers 2023, 15, 4918. [Google Scholar] [CrossRef]
  30. Manek, M.; Maita, I.; Silva, D.F.B.; de Melo, D.P.; Major, P.W.; Jaremko, J.L.; Almeida, F.T. Temporomandibular joint assessment in mri images using artificial intelligence tools: Where are we now? A systematic review. Dentomaxillofacial Radiol. 2025, 54, 1–11. [Google Scholar] [CrossRef] [PubMed]
  31. Moreira, G.C.; Ribeiro, C.S.D.C.; Verner, F.S.; Lemos, C.A.A. Performance of artificial intelligence in evaluating maxillary sinus mucosal alterations in imaging examinations: Systematic review. Dentomaxillofacial Radiol. 2025, 54, 342–349. [Google Scholar] [CrossRef] [PubMed]
  32. Querrer, R.; Vieira, L.S.; Teodoro, A.B.; Melo, N.S.; Mesquita, C.R.M.; Silva, M.A.G.; Figueiredo, P.T.S.; Leite, A.F. Deep learning for osteoporosis screening in dental practice: A systematic review. Dentomaxillofacial Radiol. 2025, twaf052. [Google Scholar] [CrossRef]
  33. Mao, K.; Thu, K.M.; Hung, K.F.; Yu, O.Y.; Hsung, R.T.; Lam, W.Y. Artificial intelligence in detecting periodontal disease from intraoral photographs: A systematic review. Int. Dent. J. 2025, 75, 100883. [Google Scholar] [CrossRef] [PubMed]
  34. Liu, Z.; Nalley, A.; Hao, J.; QY, H.A.; Yeung, A.W.K.; Tanaka, R.; Hung, K.F. The performance of large language models in dentomaxillofacial radiology: A systematic review. Dentomaxillofacial Radiol. 2025, twaf060. [Google Scholar] [CrossRef] [PubMed]
  35. Mago, J.; Sharma, M. The potential usefulness of chatgpt in oral and maxillofacial radiology. Curēus 2023, 15, e42133. [Google Scholar] [CrossRef]
  36. Machado, V.; Proença, L.; Morgado, M.; Mendes, J.J.; Botelho, J. Accuracy of panoramic radiograph for diagnosing periodontitis comparing to clinical examination. J. Clin. Med. 2020, 9, 2313. [Google Scholar] [CrossRef]
  37. Anbiaee, N.; Pahlavanzadeh, P. Evaluation of panoramic radiography diagnostic accuracy in the assessment of interdental alveolar bone loss using cbct. Clin. Exp. Dent. Res. 2024, 10, e70042. [Google Scholar] [CrossRef]
  38. Çelik, B.; Çelik, M.E. Automated detection of dental restorations using deep learning on panoramic radiographs. Dentomaxillofacial Radiol. 2022, 51, 20220244. [Google Scholar] [CrossRef] [PubMed]
  39. Lee, D.W.; Kim, S.Y.; Jeong, S.N.; Lee, J.H. Artificial intelligence in fractured dental implant detection and classification: Evaluation using dataset from two dental hospitals. Diagnostics 2021, 11, 233. [Google Scholar] [CrossRef]
  40. Balel, Y.; Sağtaş, K.; Teke, F.; Kurt, M.A. Artificial intelligence-based detection and numbering of dental implants on panoramic radiographs. Clin. Implant. Dent. Relat. Res. 2025, 27, e70000. [Google Scholar] [CrossRef]
  41. Zadrożny, Ł.; Regulski, P.; Brus-Sawczuk, K.; Czajkowska, M.; Parkanyi, L.; Ganz, S.; Mijiritsky, E. Artificial intelligence application in assessment of panoramic radiographs. Diagnostics 2022, 12, 224. [Google Scholar] [CrossRef]
  42. Zhu, J.; Chen, Z.; Zhao, J.; Yu, Y.; Li, X.; Shi, K.; Zhang, F.; Yu, F.; Shi, K.; Sun, Z.; et al. Artificial intelligence in the diagnosis of dental diseases on panoramic radiographs: A preliminary study. BMC Oral Health 2023, 23, 358. [Google Scholar] [CrossRef] [PubMed]
  43. Cosson, J. Interpreting an orthopantomogram. Aust. J. Gen. Pract. 2020, 49, 550–555. [Google Scholar] [CrossRef]
  44. Wenzel, A.; Matzen, L.H.; Spin-Neto, R.; Schropp, L. Effect of computer-assisted-learning and simulation clinics on dental students’ cognitive and performance skills: Panoramic image errors related to patient’s head position. Dentomaxillofacial Radiol. 2020, 49, 20200154. [Google Scholar] [CrossRef]
  45. Santos, K.K.; Lages, F.S.; Maciel, C.A.B.; Glória, J.C.R.; Douglas-de-Oliveira, D.W. Prevalence of mandibular third molars according to the pell & gregory and winter classifications. J. Maxillofac. Oral Surg. 2022, 21, 627–633. [Google Scholar] [CrossRef]
  46. Güller, M.T.; Kumbasar, N.; Miloğlu, Ö. Evaluation of the effectiveness of panoramic radiography in impacted mandibular third molars on deep learning models developed with findings obtained with cone beam computed tomography. Oral Radiol. 2025, 41, 260–275. [Google Scholar] [CrossRef] [PubMed]
  47. Suárez, A.; Arena, S.; Calzada, A.H.; Varón, A.I.C.; García, V.D.-F.; Freire, Y. Decoding wisdom: Evaluating chatgpt’s accuracy and reproducibility in analyzing orthopantomographic images for third molar assessment. Comput. Struct. Biotechnol. J. 2025, 28, 141–147. [Google Scholar] [CrossRef] [PubMed]
  48. Salmanpour, F.; Akpınar, M. Performance of chat generative pretrained transformer-4.0 in determining labiolingual localization of maxillary impacted canine and presence of resorption in incisors through panoramic radiographs: A retrospective study. Am. J. Orthod. Dentofac. Orthop. 2025, 168, 220–231. [Google Scholar] [CrossRef]
  49. Silva, T.P.; Andrade-Bortoletto, M.F.S.; Ocampo, T.S.C.; Alencar-Palha, C.; Bornstein, M.M.; Oliveira-Santos, C.; Oliveira, M.L. Performance of a commercially available generative pre-trained transformer (gpt) in describing radiolucent lesions in panoramic radiographs and establishing differential diagnoses. Clin. Oral Investig. 2024, 28, 204. [Google Scholar] [CrossRef]
  50. Owusu-Adjei, M.; Hayfron-Acquah, J.B.; Frimpong, T.; Abdul-Salaam, G. Imbalanced class distribution and performance evaluation metrics: A systematic review of prediction accuracy for determining model performance in healthcare systems. PLoS Digit. Health 2023, 2, e0000290. [Google Scholar] [CrossRef]
Figure 1. A representative case, displaying both the input image (above) and GPT-4o’s output (below). In this case, GPT-4o did not detect some dental conditions at the correct tooth sites. Teeth 18, 28, 38, 47, and 48 are missing (yellow doted rectangles) in this image, but only tooth 38 was identified as missing by GPT-4o. A dental implant is present at the site of tooth 46 (blue doted rectangle), which was accurately detected by GPT-4o. While GPT-4o correctly detected the presence of three teeth with restorations (green doted rectangles) in the posterior mandible, it misidentified their locations as 36, 46, and 47 instead of the correct sites, 36, 37, and 46. Neither the filling (orange doted rectangle) in teeth 14, 15, 16 nor the endodontic treatment (red doted rectangle) on tooth 37 was identified by GPT-4o.
Figure 1. A representative case, displaying both the input image (above) and GPT-4o’s output (below). In this case, GPT-4o did not detect some dental conditions at the correct tooth sites. Teeth 18, 28, 38, 47, and 48 are missing (yellow doted rectangles) in this image, but only tooth 38 was identified as missing by GPT-4o. A dental implant is present at the site of tooth 46 (blue doted rectangle), which was accurately detected by GPT-4o. While GPT-4o correctly detected the presence of three teeth with restorations (green doted rectangles) in the posterior mandible, it misidentified their locations as 36, 46, and 47 instead of the correct sites, 36, 37, and 46. Neither the filling (orange doted rectangle) in teeth 14, 15, 16 nor the endodontic treatment (red doted rectangle) on tooth 37 was identified by GPT-4o.
Diagnostics 15 02315 g001
Figure 2. A representative case, displaying both the input image (above) and GPT-4o’s output (below). In this case, GPT-4o detected some dental conditions that are not present in the image, demonstrating hallucination. Although GPT-4o identified multiple missing teeth and a third molar (yellow doted rectangles), these findings are false positives, as all teeth are present in the image. GPT-4o correctly identified the presence of an osteolytic lesion in the mandible (blue doted rectangle), but mislocated it to the right posterior region instead of the left posterior region.
Figure 2. A representative case, displaying both the input image (above) and GPT-4o’s output (below). In this case, GPT-4o detected some dental conditions that are not present in the image, demonstrating hallucination. Although GPT-4o identified multiple missing teeth and a third molar (yellow doted rectangles), these findings are false positives, as all teeth are present in the image. GPT-4o correctly identified the presence of an osteolytic lesion in the mandible (blue doted rectangle), but mislocated it to the right posterior region instead of the left posterior region.
Diagnostics 15 02315 g002
Table 1. The count of positive and negative cases for each dental condition identified in the included images at both the patient and tooth levels as determined by the reference standard and by GPT-4o using either the FDI or universal numbering system.
Table 1. The count of positive and negative cases for each dental condition identified in the included images at both the patient and tooth levels as determined by the reference standard and by GPT-4o using either the FDI or universal numbering system.
ConditionLevelPositive Count (Reference Standard)Negative Count (Reference Standard)FDI Numbering SystemUniversal Numbering System
Positive Count (GPT-4o)Negative Count (GPT-4o)Positive Count (GPT-4o)Negative Count (GPT-4o)
Missing teethPatient3515419428
Tooth195140513214681411459
Impacted teethPatient232733173119
Tooth543465134955345
CariesPatient743248149
Tooth2115793159721598
Endodontically treated teeth or teeth with restorations (Fillings, Crowns or bridges)Patient381236143515
Tooth202139810314971101490
Periapical lesionsPatient1040347050
Tooth1715832159801600
Periodontal bone lossPatient2327842545
Tooth3051295411559281572
Tooth fractures, cracks, or retained rootsPatient1040248347
Tooth1615847159381592
Dental implantsPatient743842644
Tooth131587151585121588
Osteolytic lesions or
Osteosclerosis
Patient1337248248
Region1429523072309
Table 2. The diagnostic performance of GPT-4o for detecting each condition at both the patient and tooth levels using either the FDI or universal numbering system.
Table 2. The diagnostic performance of GPT-4o for detecting each condition at both the patient and tooth levels using either the FDI or universal numbering system.
ConditionLevelF1-Score (%)Balanced Accuracy (%)Sensitivity (%)Specificity (%)
FDIUniv.FDIUniv.FDIUniv.FDIUniv.
Missing teethPatient73.6877.9246.6752.868085.7113.3320
Tooth34.2538.6961.6563.9628.7233.3394.5994.59
Impacted teethPatient64.2966.676065.0678.2678.2644.4451.85
Tooth49.5245.8761.3568.8148.1546.3092.7791.33
CariesPatient0047.6748.840095.3597.67
Tooth0049.9149.940099.8199.87
Endodontically treated teeth or teeth with restorationsPatient89.1987.6780.9279.6186.8484.217575
Tooth32.1328.2160.258.5324.2621.7896.1495.28
Periapical lesionsPatient0046.25500092.5100
Tooth0049.94500099.87100
Periodontal bone lossPatient25.8121.4351.2952.8217.3913.0485.1992.59
Tooth8.679.7651.4652.234.925.3297.9999.15
Tooth fractures, cracks, or retained rootsPatient015.3847.552.50109595
Tooth0049.7749.750099.5699.49
Dental implantsPatient99.3392.3198.8392.8610085.7197.67100
Tooth28.574865.0477.8930.7746.1599.3199.62
Osteolytic lesions or OsteosclerosisPatient13.3313.3352.4952.497.697.6997.3097.30
Region011.7649.6653.1606.6799.3299.66
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Z.; Ai, Q.Y.H.; Yeung, A.W.K.; Tanaka, R.; Nalley, A.; Hung, K.F. Performance of a Vision-Language Model in Detecting Common Dental Conditions on Panoramic Radiographs Using Different Tooth Numbering Systems. Diagnostics 2025, 15, 2315. https://doi.org/10.3390/diagnostics15182315

AMA Style

Liu Z, Ai QYH, Yeung AWK, Tanaka R, Nalley A, Hung KF. Performance of a Vision-Language Model in Detecting Common Dental Conditions on Panoramic Radiographs Using Different Tooth Numbering Systems. Diagnostics. 2025; 15(18):2315. https://doi.org/10.3390/diagnostics15182315

Chicago/Turabian Style

Liu, Zekai, Qi Yong H. Ai, Andy Wai Kan Yeung, Ray Tanaka, Andrew Nalley, and Kuo Feng Hung. 2025. "Performance of a Vision-Language Model in Detecting Common Dental Conditions on Panoramic Radiographs Using Different Tooth Numbering Systems" Diagnostics 15, no. 18: 2315. https://doi.org/10.3390/diagnostics15182315

APA Style

Liu, Z., Ai, Q. Y. H., Yeung, A. W. K., Tanaka, R., Nalley, A., & Hung, K. F. (2025). Performance of a Vision-Language Model in Detecting Common Dental Conditions on Panoramic Radiographs Using Different Tooth Numbering Systems. Diagnostics, 15(18), 2315. https://doi.org/10.3390/diagnostics15182315

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop