Detection of Dental Anomalies in Digital Panoramic Images Using YOLO: A Next Generation Approach Based on Single Stage Detection Models
Abstract
1. Introduction
2. Materials and Methods
2.1. Dataset and Preprocessing
Image Quality Assessment and Selection Criteria
- Complete Anatomical Coverage: The radiograph had to provide a complete view of the dentition, extending from the right third molar region to the left third molar region, while clearly visualizing both mandibular condyles and the inferior border of the mandible.
- Correct Patient Positioning: Images had to exhibit proper patient positioning, characterized by a slight “smile line” of the occlusal plane and no superimposition of the spinal column over the anterior teeth.
- Sufficient Image Contrast and Density: The image density and contrast levels had to be adequate to allow for clear differentiation between enamel, dentin, and pulpal tissues.
- Conversely, images were excluded if they presented one or more of the following significant issues:
- Major Positioning Errors: Severe errors such as a slumped patient position or incorrect head tilt that resulted in significant distortion or magnification of the dental arches.
- Obscuring Artifacts: Presence of artifacts from earrings, eyeglasses, or lead aprons that superimposed over critical anatomical structures and impeded diagnostic evaluation.
- Motion Artifacts: Blurring across the majority of the image due to patient movement during exposure, rendering fine details diagnostically unusable.
2.2. Model Selection and Architecture
2.3. Model Training and Parameters
2.3.1. Model Training and Validation Strategy
- Tier 1: Internal Training and Hyperparameter Tuning on the Primary Dataset
- Tier 2: External Validation on an Independent Public Dataset
2.3.2. Training Configuration and Hyperparameters
2.4. Performance Evaluation Metrics
2.4.1. Loss Functions
- Bounding Box (Box) Regression Loss: This component quantifies the accuracy of the predicted bounding box locations and sizes. Modern YOLO versions utilize advanced IoU-based losses, such as Complete IoU (CIoU) Loss, which accounts for overlap area, central point distance, and aspect ratio, leading to more stable training.
- Classification (CLS) Loss: This measures the correctness of the class predictions for each detected object. It is typically calculated using Binary Cross-Entropy (BCE) with logits, which is effective for multi-label classification tasks.
- Objectness (OBJ) or Distribution Focal (DFL) Loss: This component helps the model distinguish between foreground objects and the background. It also uses a BCE-based loss to predict the confidence score for each bounding box.
2.4.2. Classification Performance Metrics
- True Positive (TP): An instance where the model correctly identifies a positive class.
- False Positive (FP): An instance where the model incorrectly identifies a positive class.
- True Negative (TN): An instance where the model correctly identifies a negative class.
- False Negative (FN): An instance where the model incorrectly identifies a negative class.
- From these components, the following metrics were calculated:
- Accuracy: The ratio of all correct predictions (both positive and negative) to the total number of instances, as calculated in Formula (1).
- Precision: The ratio of correctly identified positive detections (True Positives, TP) to the total number of positive detections made by the model (TP + False Positives, FP), as shown in Formula (2). It measures the accuracy of the predictions.
- Recall: The ratio of correctly identified positive detections (TP) to the total number of actual positive instances in the data (TP + False Negatives, FN), calculated as shown in Formula (3). It measures the model’s ability to find all relevant objects.
- F1-Score: The harmonic mean of Precision and Recall. It provides a single, balanced measure of a model’s performance, which is particularly useful when there is an uneven class distribution.
2.4.3. Object Detection Metrics
- Average Precision (AP): Calculated as the area under the Precision-Recall curve for a single class, as shown in Formula (5). It summarizes the model’s performance on that specific class across all recall levels
- Mean Average Precision (mAP): The mean of the AP values calculated across all classes, as defined by Formula (6). This provides a single, aggregate score for the model’s overall performance. In this study, we report mAP50 (mAP at an IoU threshold of 0.5) and mAP50-95 (mAP averaged over IoU thresholds from 0.5 to 0.95).
2.5. Implementation Details
3. Results
3.1. Comparative Performance and Model Selection on the Internal Validation Set
3.2. Final Performance on the Independent External Test Set
4. Discussion
4.1. Interpretation of Findings and Clinical Implications
4.2. Comparison with Existing Literature
4.3. Strengths and Limitations
4.4. Future Directions
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Özler, C.Ö.; İleri Keçeli, T.; Uzamış Tekçiçek, M. Pediatric Dentistry and General Anesthesia. Ank. Med. J. 2019, 19, 658–664. [Google Scholar] [CrossRef]
- Sivari, E.; Senirkentli, G.B.; Bostanci, E.; Guzel, M.S.; Acici, K.; Asuroglu, T. Deep Learning in Diagnosis of Dental Anomalies and Diseases: A Systematic Review. Diagnostics 2023, 13, 2512. [Google Scholar] [CrossRef] [PubMed]
- Peres, M.A.; Macpherson, L.M.D.; Weyant, R.J.; Daly, B.; Venturelli, R.; Mathur, M.R.; Listl, S.; Celeste, R.K.; Guarnizo-Herreño, C.C.; Kearns, C.; et al. Oral Diseases: A Global Public Health Challenge. Lancet 2019, 394, 249–260. [Google Scholar] [CrossRef]
- Ali, A.H.; Koller, G.; Foschi, F.; Andiappan, M.; Bruce, K.D.; Banerjee, A.; Mannocci, F. Self-Limiting versus Conventional Caries Removal: A Randomized Clinical Trial. J. Dent. Res. 2018, 97, 1207–1213. [Google Scholar] [CrossRef]
- Muresan, M.P.; Barbura, A.R.; Nedevschi, S. Teeth Detection and Dental Problem Classification in Panoramic X-Ray Images using Deep Learning and Image Processing Techniques. In Proceedings of the 2020 IEEE 16th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 3–5 September 2020; pp. 457–463. [Google Scholar] [CrossRef]
- Schwendicke, F.; Golla, T.; Dreher, M.; Krois, J. Convolutional Neural Networks for Dental Image Diagnostics: A Scoping Review. J. Dent. 2019, 91, 103226. [Google Scholar] [CrossRef]
- Macdonald, D.; Yu, W. Incidental Findings in a Consecutive Series of Digital Panoramic Radiographs. Imaging Sci. Dent. 2020, 50, 53–64. [Google Scholar] [CrossRef]
- Nguyen, T.T. Use of Artificial Intelligence in Dentistry: Current Clinical Trends and Research Advances. Dent. News 2021, 28, 50–57. [Google Scholar] [CrossRef]
- Park, W.J.; Park, J.B. History and Application of Artificial Neural Networks in Dentistry. Eur. J. Dent. 2018, 12, 594–601. [Google Scholar] [CrossRef]
- Saglam, H.; Ari, T.; Bayrakdar, İ.Ş.; Bilgir, E.; Uğurlu, M.; Çelik, Ö.; Orhan, K. Artificial Intelligence In Dentistry. Sağlık Bilimlerinde Yapay Zeka Dergisi. J. Artif. Intell. Health Sci. 2021, 1, 26–33. [Google Scholar]
- Mumuni, A.; Mumuni, F. Automated Data Processing and Feature Engineering for Deep Learning and Big Data Applications: A Survey. J. Inf. Intell. 2024, 3, 113–153. [Google Scholar] [CrossRef]
- Zhou, T.; Ruan, S.; Canu, S. A Review: Deep Learning for Medical Image Segmentation Using Multi-Modality Fusion. Array 2020, 3–4, 100004. [Google Scholar] [CrossRef]
- Hamdan, M.H.; Tuzova, L.; Mol, A.; Tawil, P.Z.; Tuzoff, D.; Tyndall, D.A. The Effect of a Deep-Learning Tool on Dentists’ Performances in Detecting Apical Radiolucencies on Periapical Radiographs. DentoMaxillofac. Radiol. 2022, 51, 20220122. [Google Scholar] [CrossRef]
- Prados-Privado, M.; Villalón, J.G.; Martínez-Martínez, C.H.; Ivorra, C. Dental Images Recognition Technology and Applications: A Literature Review. Appl. Sci. 2020, 10, 2856. [Google Scholar] [CrossRef]
- Naqvi, R.A.; Arsalan, M.; Qaiser, T.; Khan, T.M.; Razzak, I. Sensor Data Fusion Based on Deep Learning for Computer Vision Applications and Medical Applications. Sensors 2022, 22, 8058. [Google Scholar] [CrossRef]
- Vishwanathaiah, S.; Fageeh, H.N.; Khanagar, S.B.; Maganur, P.C. Artificial Intelligence Its Uses and Application in Pediatric Dentistry: A Review. Biomedicines 2023, 11, 788. [Google Scholar] [CrossRef]
- Çolak, H.; Dülgergil, Ç.; Dalli, M.; Hamidi, M. Early Childhood Caries Update: A Review of Causes, Diagnoses, and Treatments. J. Nat. Sci. Biol. Med. 2013, 4, 29. [Google Scholar] [CrossRef]
- Kaya, E.; Gunec, H.G.; Aydin, K.C.; Urkmez, E.S.; Duranay, R.; Ates, H.F. A Deep Learning Approach to Permanent Tooth Germ Detection on Pediatric Panoramic Radiographs. Imaging Sci. Dent. 2022, 52, 275–281. [Google Scholar] [CrossRef] [PubMed]
- Kılıc, M.C.; Bayrakdar, I.S.; Çelik, Ö.; Bilgir, E.; Orhan, K.; Aydın, O.B.; Kaplan, F.A.; Sağlam, H.; Odabaş, A.; Aslan, A.F.; et al. Artificial Intelligence System for Automatic Deciduous Tooth Detection and Numbering in Panoramic Radiographs. DentoMaxillofac. Radiol. 2021, 50, 20200172. [Google Scholar] [CrossRef] [PubMed]
- Zhang, Y.; Ye, F.; Chen, L.; Xu, F.; Chen, X.; Wu, H.; Cao, M.; Li, Y.; Wang, Y.; Huang, X. Children’s Dental Panoramic Radiographs Dataset for Caries Segmentation and Dental Disease Detection. Sci. Data 2023, 10, 380. [Google Scholar] [CrossRef] [PubMed]
- Viola, P.; Jones, M. Rapid Object Detection Using a Boosted Cascade of Simple Features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Kauai, HI, USA, 8–14 December 2001; Volume 1. [Google Scholar] [CrossRef]
- Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; pp. 886–893. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 1137–1149. [Google Scholar] [CrossRef]
- Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar] [CrossRef]
- Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2019; pp. 10778–10787. [Google Scholar] [CrossRef]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
- Sapkota, R.; Meng, Z.; Churuvija, M.; Du, X.; Ma, Z.; Karkee, M. Comprehensive Performance Evaluation of YOLO11, YOLOv10, YOLOv9, and YOLOv8 on Detecting and Counting Fruitlet in Complex Orchard Environments. arXiv 2024, arXiv:2407.12040v5. [Google Scholar]
- Bumann, E.E.; Al-Qarni, S.; Chandrashekar, G.; Sabzian, R.; Bohaty, B.; Lee, Y. A Novel Collaborative Learning Model for Mixed Dentition and Fillings Segmentation in Panoramic Radiographs. J. Dent. 2024, 140, 104779. [Google Scholar] [CrossRef]
Frequency (Label) | Relative Frequency (%) | Cumulative (%) | |
---|---|---|---|
Dental Caries | 830 | 43.57 | 43.57 |
Deciduous Tooth | 791 | 41.52 | 85.09 |
Root Canal Treatment | 171 | 8.98 | 94.07 |
Pulpotomy | 113 | 5.93 | 100.00 |
Algorithms | Pixels | Parameters (M) | FLOPs (G) |
---|---|---|---|
YOLOv8x | 640 | 68.16 | 311.2 |
YOLOv9e | 640 | 58.1 | 192.5 |
YOLOv10x | 640 | 29.5 | 160.4 |
YOLOv11x | 640 | 56.9 | 194.9 |
Category | Parameter | Value/Description |
---|---|---|
Dataset & Input | Input Image Size | 640 × 640 pixels |
Batch Size | 16 | |
Training Regimen | Number of Epoch | 500 |
Optimizer & Learning Rate | Optimizer | SGD (Stochastic Gradient Descent) |
Momentum | 0.937 | |
Weight Decay | 0.0005 | |
Initial Learning Rate (lr0) | 0.01 | |
Learning Rate Scheduler | Cosine Annealing | |
Data Augmentation | Mosaic | Applied to combine four training images into one. |
MixUp | Applied to create composite images by linearly interpolating two images and their labels. | |
Random Affine Transformations | Included random rotations, scaling, and translations. | |
Loss & Evaluation | Training IoU Threshold | 0.5 |
Loss Function Components | A composite function including:
|
Accuracy | Recall | Precision | F1-Score | |
---|---|---|---|---|
YOLOv8x | 0.86 | 0.89 | 0.92 | 0.90 |
YOLOv9e | 0.87 | 0.89 | 0.91 | 0.90 |
YOLOv10x | 0.87 | 0.89 | 0.92 | 0.90 |
YOLOv11x | 0.91 | 0.92 | 0.94 | 0.93 |
mAP50 | mAP50-95 | |
---|---|---|
YOLOv8x | 0.89 | 0.71 |
YOLOv9e | 0.90 | 0.70 |
YOLOv10x | 0.89 | 0.71 |
YOLOv11x | 0.91 | 0.69 |
Box Loss | CLS Loss | DFL Lose | |
---|---|---|---|
YOLOv8x | 0.93 | 0.83 | 1.27 |
YOLOv9e | 0.92 | 0.84 | 1.49 |
YOLOv10x | 1.84 | 2.38 | 2.23 |
YOLOv11x | 0.90 | 0.82 | 1.10 |
Training Time Per Epoch (s) | Validation Time per Epoch (s) | |
---|---|---|
YOLOv8x | 0.92 | 0.89 |
YOLOv9e | 0.91 | 0.89 |
YOLOv10x | 0.92 | 0.89 |
YOLOv11x | 0.90 | 0.90 |
Accuracy | Recall | Precision | F1-Score | |
---|---|---|---|---|
YOLOv11x | 0.82 | 0.83 | 0.80 | 0.81 |
mAP50 | mAP50-95 | |
---|---|---|
YOLOv11x | 0.82 | 0.62 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Şevik, U.; Mutlu, O. Detection of Dental Anomalies in Digital Panoramic Images Using YOLO: A Next Generation Approach Based on Single Stage Detection Models. Diagnostics 2025, 15, 1961. https://doi.org/10.3390/diagnostics15151961
Şevik U, Mutlu O. Detection of Dental Anomalies in Digital Panoramic Images Using YOLO: A Next Generation Approach Based on Single Stage Detection Models. Diagnostics. 2025; 15(15):1961. https://doi.org/10.3390/diagnostics15151961
Chicago/Turabian StyleŞevik, Uğur, and Onur Mutlu. 2025. "Detection of Dental Anomalies in Digital Panoramic Images Using YOLO: A Next Generation Approach Based on Single Stage Detection Models" Diagnostics 15, no. 15: 1961. https://doi.org/10.3390/diagnostics15151961
APA StyleŞevik, U., & Mutlu, O. (2025). Detection of Dental Anomalies in Digital Panoramic Images Using YOLO: A Next Generation Approach Based on Single Stage Detection Models. Diagnostics, 15(15), 1961. https://doi.org/10.3390/diagnostics15151961