Next Article in Journal
Efficacy of Triphala and Chlorhexidine Mouthwashes on Gingival Inflammation and Dental Plaque in Children: A Systematic Review
Next Article in Special Issue
A Pre-Adoption Assessment of a Collaborative Care Approach to Dental-Fear Treatment
Previous Article in Journal
Association Between Hashimoto’s Thyroiditis and Periodontal Disease: A Narrative Review
Previous Article in Special Issue
Usage of Chelating Gels in Root Canal Preparation: A Survey of Australian Clinicians
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Elementary White Mucosal Lesions by an AI System: A Pilot Study

by
Gaetano La Mantia
1,2,3,*,
Federico Kiswarday
4,
Giuseppe Pizzo
1,2,4,
Giovanna Giuliana
2,
Giacomo Oteri
3,
Mario G. C. A. Cimino
5,
Olga Di Fede
2 and
Giuseppina Campisi
1,6
1
Unit of Oral Medicine and Dentistry for Fragile Patients, Department of Rehabilitation, Fragility and Continuity of Care University Hospital “Policlinico Paolo Giaccone”, 90127 Palermo, Italy
2
Department of Precision Medicine in Medical, Surgical and Critical Care (Me.Pre.C.C.), University of Palermo, 90127 Palermo, Italy
3
Department of Biomedical and Dental Sciences, Morphological and Functional Images, University of Messina, 98124 Messina, Italy
4
Degree in Dentistry, School of Medicine and Surgery, University of Palermo, 90127 Palermo, Italy
5
Department of Information Engineering, University of Pisa, 56122 Pisa, Italy
6
Department of Biomedicine, Neuroscience and Advanced Diagnostics (BIND), University of Palermo, 90127 Palermo, Italy
*
Author to whom correspondence should be addressed.
Oral 2024, 4(4), 557-566; https://doi.org/10.3390/oral4040043
Submission received: 14 August 2024 / Revised: 4 November 2024 / Accepted: 6 November 2024 / Published: 14 November 2024
(This article belongs to the Special Issue Current Issues in Oral Health)

Abstract

:
Aim: Accurately identifying primary lesions in oral medicine, particularly elementary white lesions, is a significant challenge, especially for trainee dentists. This study aimed to develop and evaluate a deep learning (DL) model for the detection and classification of elementary white mucosal lesions (EWMLs) using clinical images. Materials and Methods: A dataset was created by collecting photographs of various oral lesions, including oral leukoplakia, OLP plaque-like and reticular forms, OLL, oral candidiasis, and hyperkeratotic lesions from the Unit of Oral Medicine. The SentiSight.AI (Neurotechnology Co.®, Vilnius, Lithuania) AI platform was used for image labeling and model training. The dataset comprised 221 photos, divided into training (n = 179) and validation (n = 42) sets. Results: The model achieved an overall precision of 77.2%, sensitivity of 76.0%, F1 score of 74.4%, and mAP of 82.3%. Specific classes, such as condyloma and papilloma, demonstrated high performance, while others like leucoplakia showed room for improvement. Conclusions: The DL model showed promising results in detecting and classifying EWMLs, with significant potential for educational tools and clinical applications. Expanding the dataset and incorporating diverse image sources are essential for improving model accuracy and generalizability.

1. Introduction

Diagnosing oral white lesions is complex due to the broad spectrum of conditions they encompass, ranging from benign reactive lesions to potentially serious dysplastic or neoplastic ones. Characteristic features can help distinguish these lesions, but similarities often complicate the diagnosis, necessitating biopsies for confirmation. Establishing a definitive diagnosis is crucial to avoid delays in treating patients with more severe conditions [1,2].
Elementary white mucosal lesions (EWMLs) constitute only 5% of oral pathologies, yet some, such as leukoplakia and lichen planus, carry significant malignant potential (up to 0.5–100%). Therefore, a thorough clinical diagnostic approach is essential to rule out malignancy or premalignancy [3].
The clinical signs of these conditions vary, and only senior oral medicine experts have demonstrated high reliability in distinguishing them. General dentists, especially trainees, may exhibit lower diagnostic accuracies. An AI system tailored to assist dental students and trainees could be a valuable tool in effectively distinguishing EWMLs [3,4,5].
EWMLs can result from a thickened keratotic layer or the accumulation of non-keratotic material, arising from factors such as local frictional irritation, immunologic reactions, or more serious processes like premalignant or malignant transformation [3].
Frictional hyperkeratosis, a benign lesion, is characterized by epithelial thickening due to chronic irritation, commonly from ill-fitting dental appliances or habitual cheek biting [6]. Oral leukoplakia (OL) presents as a white patch or plaque that cannot be clinically or histologically attributed to other diseases, often associated with irritants like tobacco and alcohol [7]. Oral lichen planus (OLP) is a chronic mucocutaneous disorder believed to be immune-mediated, presenting with various patterns, including reticular and erosive forms [8,9]. Oral lichenoid reactions (OLR) resemble OLP but arise from medications or dental materials [9,10,11]. Oral condylomas and papillomas are benign HPV-related lesions with whitish growths on the mucosa [12,13,14]. Oral candidiasis, or thrush, is a fungal infection characterized by forms such as pseudomembranous and hyperplastic [15,16,17].
Several mobile applications have been proposed to improve the clinical recognition of oral lesions among medical and dental professionals [18,19]. Machine learning (ML), a subset of artificial intelligence (AI), has emerged as a potent tool for supporting diagnostic tasks in various fields, including healthcare and dentistry. Deep learning (DL) is a subset of ML that involves neural networks with many layers [8,9]. These neural networks are designed to automatically learn and extract features from raw data through a process of hierarchical representation [10]. In the context of this study, DL refers to the use of such advanced neural network architectures to analyze and classify clinical images of oral lesions, enabling the model to detect and differentiate between various types of lesions with high accuracy [11]. DL employs multi-layered neural networks, such as convolutional neural networks (CNNs), to discern intricate patterns in data, particularly in complex structures like medical images. CNNs have played a crucial role in detecting a variety of medical conditions, such as breast cancer in mammograms, skin cancer in clinical screenings, and diabetic retinopathy in retinal images. Similarly, in dentistry, CNNs have shown remarkable effectiveness in identifying different conditions like periodontal bone loss, apical lesions, and caries lesions, achieving high levels of accuracy [8,9,12].
AI has transformed medical and dental diagnostics by analyzing and classifying data, enabling precise diagnoses. DL models, especially CNNs, extract features like gradient differences and analyze shape, contour, and pattern, facilitating efficient detection and differentiation of EWMLs [20,21,22,23,24,25]. The relevance of AI to oral health is profound. By improving early detection of potentially malignant lesions, AI can enable timely interventions and better patient outcomes. The consistency and objectivity of AI analyses reduce variability and enhance diagnostic accuracy, making the diagnostic process more reliable. Moreover, AI tools can streamline workflows, saving time and resources for practitioners and patients. In educational settings, AI aids in training future dental professionals to effectively diagnose and manage various oral conditions. As of today, there is a lack of AI systems tailored to differentiate EWMLs. Our research aims to fill this critical gap in current diagnostic practices by exploring the effectiveness of a deep learning model in detecting and classifying EWMLs from oral images. The central research question driving this study is: How effective is a deep learning model in detecting and classifying EWMLs in oral images? We hypothesize that the model will achieve high accuracy, potentially outperforming traditional diagnostic methods, and effectively assist trainee dentists and dental students in differentiating EWMLs, thereby improving their diagnostic skills and contributing to more accurate clinical outcomes.

2. Materials and Methods

The initial dataset consisted of 231 clinical oral photographs selected from the database of the Oral Medicine Unit at University Hospital “Policlinico Paolo Giaccone” Palermo, Italy.
The study protocol conformed with the ethical guidelines of the 1964 Declaration of Helsinki and its later amendments or comparable ethical standards.
These photographs were selected from the database during the period from 29 March 2024 to 8 July 2024. The photographs, which had been originally taken prior to this period, were included in the dataset after obtaining informed consent from patients (ethical approval n.12 07/05/2024). The dataset aimed to represent various oral conditions, including oral leukoplakia, striae and reticular figurations, morsicatio buccarum, linea alba, hyperplastic candidiasis, papilloma, and condyloma. Out of these, 10 photos were discarded during the labeling phase, resulting in a final dataset of 221 photos. The discarded photos did not meet the fundamental requirements for accurate lesion detection. Specifically, these images were removed due to issues such as poor focus, inadequate lighting, or repetitive content. These quality concerns were deemed critical to ensure that the dataset used for training and evaluation was of high integrity, thus supporting reliable model performance. These photographs were meticulously categorized after definitive histopathologic diagnoses (when necessary) into five distinct classes, depicting various manifestations of oral mucosal lesions.
The classification comprised 60 images for striate-reticular lesions, 30 for morsicatio buccarum, 24 for linea alba, 39 for leukoplakia, 40 for condyloma-papilloma, and 28 for oral candidiasis (Table 1). To ensure diversity and randomness in the dataset, photographs of both male and female patients were randomly selected.
A Nikon D7200 digital camera, equipped with a 105 mm lens and a macro flash SB-R200, was utilized for capturing the oral mucosal lesions. The resolution of the photographs varied, with dimensions of 6000 × 4000 and 4800 × 3200 pixels. A horizontal resolution of 300 dpi and a vertical resolution of 300 dpi were maintained, with a bit depth of 24 and color representation in sRGB. The input images were resized while preserving their original aspect ratio, with the new maximum dimension set to 1024 pixels and the minimum dimension set to 600 pixels. These specifications ensured high-quality images suitable for detailed analysis and classification.

2.1. Labeling Phase

For labeling the photographs in the dataset, a manual annotation process was employed. The image labeling and model training processes were conducted using the SentiSight.AI (Neurotechnology Co.®, Vilnius, Lithuania) web-based tool. Patient privacy was ensured through a confidentiality agreement with Neurotechnology, confirming that images on the SentiSight.AI platform were used only for model training, not shared with third parties, and stored securely. Each photograph underwent a thorough examination by trained professional senior experts in oral medicine. Utilizing a manual annotation process, clinical images were labeled with bounding boxes outlining regions of interest (ROIs) corresponding to specific lesions depicted. These ROIs were subsequently assigned appropriate class identifiers corresponding to the type of oral mucosal lesion present.
In instances where the lesion spanned a significant area, necessitating the delineation of a large region that could potentially include non-relevant elements such as teeth, clinicians opted to create multiple bounding boxes to optimize the utilization of the photograph without introducing undesired elements irrelevant to the lesion. This approach ensured that the model received optimal training data, minimizing the risk of incorporating extraneous information during the training process.
In some cases, clinicians utilized the inversion of image colors (negative) to facilitate the recognition of certain lesions.

2.2. Model Training

The algorithm used for training was Faster R-CNN with ResNet-101, a powerful CNN architecture. This model, with 44.5 million parameters and 347 layers, served as the backbone network for the detection and classification tasks. Pre-trained weights from the MS COCO dataset were utilized to initialize the network, leveraging the rich feature representations learned from a diverse range of objects and scenes. Subsequently, the entire network was fine-tuned on our dataset of oral mucosal lesion images.
The total training time amounted to 60 min and 27 s. For each of the five classes, an 80% random subset (179 photos) served as the training set, while the remaining 20% (42 photos) constituted the validation set. During fine-tuning, a flat learning rate of 10 × 103 and a large model size were employed throughout training, ensuring stable convergence while updating the network parameters. The Stochastic Gradient Descent (SGD) optimizer with momentum was utilized for efficient optimization, facilitating rapid convergence towards a locally optimal solution. Additionally, a batch size of one image was used, allowing for dynamic adjustments during training to accommodate variations in image characteristics and lesion manifestations. The selection of the model checkpoint was based on the highest mean average precision (mAP) metric achieved on the validation set, which occurred at 47 min and 22 s (Figure 1). The cross-entropy loss function was used. Regarding the hardware used for training, the SentiSight.AI web-based tool was utilized. It is important to note that the hardware choice does not influence the accuracy of predictions and, therefore, is not relevant to the scientific study. The focus lies on the methodology, algorithms, and data quality rather than the specific hardware configuration. The best model occurred at 47 min and 22 s.

2.3. Model Validation and Evaluation

The model performance was evaluated on the validation set, which comprised 42 images that were unseen during training, ensuring an independent assessment. A threshold intersection over union (IoU) value of 0.5 was utilized to assess the prediction performance. IoU measures the overlap between the predicted bounding box and the ground truth box, with higher values indicating a closer match. IoU was used to classify predictions as true positive (TP), false positive (FP), and false negative (FN). Specifically, if the predicted bounding box had an IoU ≥ 0.5 with the ground truth bounding box, it was considered a TP. Conversely, if the predicted bounding box had an IoU < 0.5 with the ground truth bounding box, it was considered an FP. Additionally, if the ground truth bounding box was not detected by the model, it was considered an FN.
Global metrics for the validation set were computed using precision, sensitivity, F1 score (harmonic mean of precision and sensitivity, balancing the two metrics), and mAP (Equation (1)).
P r e c i s i o n = T P T P + F P
S e n s i t i v i t y = T P T P + F N
F 1   s c o r e = 2 · P r e c i s i o n · S e n s i t i v i t y P r e c i s i o n + S e n s i t i v i t y
m A P = 1 N i = 1 N A P i
Equation (1): Metrics of performance for CNN models. Abbreviations: FN, false negative; FP, false positive; TP, true positive; AP, average precision; mAP, mean average precision.

3. Results

The evaluation of the five-class object detection model was conducted on the validation set. Overall statistics for the entire validation test are summarized in Table 2, while detailed statistics for each class are reported in Table 3. Globally, the model achieved a precision of 77.2%, sensitivity of 76.0%, F1 score of 74.4%, and an mAP of 82.3%. To provide a clearer overview of the results, a confusion matrix was utilized, enabling a more comprehensive understanding of the model’s performance in terms of correctly and incorrectly classified instances for each class (Figure 2). Additionally, Table 4 shows the counts of TP, FP, and FN for each class. The confusion matrix provides a detailed overview of the model’s performance in correctly classifying different oral lesions in the study. The results show a high accuracy for the striate-reticular lesions, linea alba, and condyloma–papilloma classes, with a value of 1.00, indicating that all images in these classes were classified correctly. The morsicatio buccarum class achieved an accuracy of 0.67, with 0.17 of the images misclassified as striate-reticular lesions and an additional 0.17 not detected. The oral candidiasis class obtained an accuracy of 0.80, with 0.20 of the lesions not detected, while the leukoplakia class showed a lower accuracy of 0.57, with 0.43 of the lesions not detected. This matrix highlights the strengths and critical areas of the model, providing a clearer understanding of the results obtained.

4. Discussion

In recent years, there has been significant interest in developing artificial intelligence (AI) systems for computer-aided diagnosis of oral lesions using clinical images. These AI systems, particularly deep learning algorithms, are designed to differentiate malignant oral lesions from benign ones by utilizing various imaging modalities, including clinical photographs and histopathological images [16,17,18,19,20].
AI techniques utilizing CNNs for the diagnosis of diseases have been extensively explored across various fields, particularly in histopathology and dermatology [18,19,20]. Similarly, in dentistry, the utilization of CNNs shows promise; in fact, AI-based detection in dentistry using clinical photography has demonstrated a high diagnostic odds ratio [18,21].
In this pilot study, we developed object recognition models capable of detecting the presence of EWMLs in images, a task that, while not difficult for an expert dentist, is significant for trainee dentists and dental students.
Several studies have demonstrated the potential of AI-based systems for classifying oral lesions. CNN-based models have shown diagnostic performance comparable to expert clinicians in classifying oral squamous cell carcinoma (OSCC) and oral potentially malignant disorders (OPMDs) using oral photographs [19,20].
To date, however, AI systems enabling the detection and classification of EWMLs are still lacking.
The development of an effective AI system for detecting and diagnosing oral lesions faces several challenges. One of the primary difficulties is obtaining large, diverse, and well-annotated datasets of oral lesion images. Current datasets are often small and limited to specific types of lesions, making it challenging to train robust AI models [22,23]. Additionally, integrating these AI systems into clinical workflows is essential to assist clinicians, especially general practitioners. These systems should improve the accuracy of early cancer diagnosis and support expert-level decision-making in screening programs for oral malignant and potentially malignant lesions. AI functions as a complementary tool rather than a replacement for clinical expertise, reaffirming that dentists remain the key decision-makers in the diagnostic process [24,25,26].
Another significant challenge is related to the medico-legal issues arising from data privacy concerns and lack of explainability for the systems used [17,27,28].
The performance of CNN models is significantly influenced by image quality and the number of samples [29]. Capturing images of the oral cavity is particularly challenging due to the lack of natural lighting, the need to remove mucous membranes for visibility in certain regions, and difficulties in standardizing angles, distance, framing, and sharpness [30].
The condyloma, papilloma, and hyperplastic candidiasis classes achieved outstanding performance across all statistics, demonstrating the model’s ability to distinguish these lesions accurately. High values in some indicators (e.g., precision, sensitivity, F1 score, and mAP) for these classes underscore the robustness and reliability of the model in these areas.
The striae, reticular, and morsicatio buccarum classes show relatively balanced performance, with striae and reticular achieving an impressive sensitivity of 84.6%. This highlights the model’s ability to detect these lesions effectively, even if further refinement could enhance precision and reduce false positives.
The leucoplakia class exhibited a precision of 60.0% and a sensitivity of 25.0%. This discrepancy suggests the need for more balanced and representative data for this class to improve the model ability to generalize.
The linea alba class also shows very promising results with a sensitivity of 100.0% and a strong precision of 80.0%. This indicates the efficacy of model capability in detection for this class.
The results highlight how the recognition of linea alba, condyloma, and papilloma lesions is straightforward for AI, as their morphological characteristics are easily distinguishable clinically from other analyzed classes, achieving high values across all statistics.
This study explored the process of gathering and labeling images taken from the oral cavity, specifically those exhibiting confirmed EWMLs by the clinical team. It also showcased the outcomes of using DL to automate the early detection of EWMLs. Our initial findings illustrate promising findings, indicating that DL methods could be effective in addressing this complex challenge. However, some questions are open; first of all, who should perform the image segmentation by designating the bounding boxes useful for model training. Furthermore, it is important to understand whether it is necessary to instruct oral medicine professionals in image segmentation. In our opinion, this training could also benefit dental trainees and dental students by helping them learn to recognize lesions and learn from them. The integration of DSLR photographs into this study might raise concerns about the applicability and future scalability of the AI model in real-world contexts. Although DSLRs provide high-quality images, it is likely that smartphone photos will become more prevalent in the future, especially with the development of apps for diagnosing oral lesions. However, smartphone images can vary in quality due to differences in models, camera specifications, lighting conditions, and user handling, which could impact AI performance if not trained on a diverse dataset [4,31,32,33].
To date, there are no studies comparing the performance of AI systems with images acquired from DSLRs and smartphones. Therefore, it is crucial to expand datasets with smartphone images to enhance the model’s generalization in real-world scenarios, such as screening programs or field diagnostics.

5. Limitations and Strengths

The deep learning models developed for detecting EWMLs have initially yielded satisfactory results, with overall outcomes being quite promising. Nevertheless, a significant limitation is the relatively small size of the dataset, which presents a considerable challenge during the model training phase. Although the model performs well across several classes, there are specific areas that require further improvement. To enhance the model’s robustness and generalizability, additional research with larger datasets is necessary.
The study highlights several notable strengths. The use of a Faster R-CNN with ResNet-101 architecture, one of the most advanced models available, showcases a sophisticated application of AI in oral medicine diagnosis. Another strength lies in the high-performance metrics achieved for certain lesion classes, such as condyloma, papilloma, and hyperplastic candidiasis. Additionally, the study suggests that AI-based tools could be especially valuable for training dental students and trainees, enhancing their ability to recognize and understand lesions. This educational aspect could improve training programs and support the development of diagnostic skills. Another interesting aspect of this work is the innovative use of image color inversion (negative), employed in certain cases to enhance the recognition of specific lesions.

6. Conclusions

The challenge of precise lesion localization in object detection can be addressed by expanding datasets, as larger datasets improve deep learning models’ ability to detect complex patterns and enhance generalizability. Image augmentation techniques like rotation, scaling, and color adjustments can further enhance model robustness. Incorporating images from various devices ensures adaptability to different image qualities and lighting conditions, while regular updates and user feedback help improve model accuracy. Today, there is a noticeable lack of apps that could facilitate the learning path for trainee dentists and dental students, and developing such applications in the future would be beneficial. Additionally, this tool could potentially be extended to identify red, pigmented, and red-and-white lesions. Early exposure to AI systems in dental education is crucial, as it will prepare students to integrate these technologies into future practice. Future research should focus on developing automatic segmentation tools to streamline image annotation, enhancing scalability and efficiency. While AI can assist in lesion detection, dentists are the ultimate decision-makers in diagnosis, ensuring that AI serves as a complementary tool rather than a replacement for their clinical expertise.

Author Contributions

Conceptualization, G.C. and O.D.F.; methodology, G.C., G.G., M.G.C.A.C. and O.D.F.; software, F.K., G.L.M. and M.G.C.A.C.; validation, G.P., F.K. and G.L.M.; investigation, G.L.M.; resources, G.O. and G.G.; data curation, G.P. and M.G.C.A.C.; formal analysis, M.G.C.A.C.; writing—original draft preparation, G.C., G.L.M. and F.K.; writing—review and editing, G.C., G.L.M. and O.D.F.; visualization, G.G., G.O. and G.P.; supervision, G.C. and M.G.C.A.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Local Ethics Committee of the University Hospital “Policlinico Paolo Giaccone”, Palermo, Italy (protocol code: 12/2024; date of approval: 7 May 2024).

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

We gratefully acknowledge the free technical support provided by the algorithm developers at Neurotechnology Co.®, Vilnius, Lithuania, for the analysis of the trained models during the completion of this study. The work was partially supported by the project “OCAX—Oral Cancer eXplained by DL-enhanced case-based classification” PRIN 2022 code P2022KMWX3.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. McKinney, R.; Olmo, H.; McGovern, B. Benign Chronic White Lesions of the Oral Mucosa. In StatPearls; StatPearls Publishing: Treasure Island, FL, USA, 2024. [Google Scholar]
  2. Simi, S.M.; Nandakumar, G.; Anish, T.S. White Lesions in the Oral Cavity: A Clinicopathological Study from a Tertiary Care Dermatology Centre in Kerala, India. Indian J. Dermatol. 2013, 58, 269. [Google Scholar] [CrossRef] [PubMed]
  3. Shiragur, S.S.; Srinath, S.; Yadav, S.T.; Purushothaman, A.; Chavan, N. V Spectrum of White Lesions in the Oral Cavity—A Review. J. Oral Med. Oral Surg. Oral Pathol. Oral Radiol. 2024, 10, 3–13. [Google Scholar] [CrossRef]
  4. Moosa, Y.; Alizai, M.H.K.; Tahir, A.; Zia, S.; Sadia, S.; Fareed, M.T. Artificial Intelligence in Oral Medicine. Int. J. Health Sci. 2023, 7, 1476–1488. [Google Scholar] [CrossRef]
  5. Albagieh, H.; Alzeer, Z.O.; Alasmari, O.N.; Alkadhi, A.A.; Naitah, A.N.; Almasaad, K.F.; Alshahrani, T.S.; Alshahrani, K.S.; Almahmoud, M.I. Comparing Artificial Intelligence and Senior Residents in Oral Lesion Diagnosis: A Comparative Study. Cureus 2024, 16, e51584. [Google Scholar] [CrossRef] [PubMed]
  6. Di Fede, O.; Panzarella, V.; Buttacavoli, F.; La Mantia, G.; Campisi, G. Doctoral: A Smartphone-Based Decision Support Tool for the Early Detection of oral Potentially Malignant Disorders. Digit Health 2023, 9, 20552076231177141. [Google Scholar] [CrossRef] [PubMed]
  7. Zain, R.B.; Pateel, D.G.S.; Ramanathan, A.; Kallarakkal, T.G.; Wong, G.R.; Yang, Y.H.; Zaini, Z.M.; Ibrahim, N.; Kohli, S.; Durward, C. Effectiveness of “OralDETECT”: A Repetitive Test-Enhanced, Corrective Feedback Method Competency Assessment Tool for Early Detection of Oral Cancer. J. Cancer Educ. 2022, 37, 319–327. [Google Scholar] [CrossRef]
  8. Fatima, A.; Shafi, I.; Afzal, H.; Díez, I.D.L.T.; Lourdes, D.R.S.M.; Breñosa, J.; Espinosa, J.C.M.; Ashraf, I. Advancements in Dentistry with Artificial Intelligence: Current Clinical Applications and Future Perspectives. Healthcare 2022, 10, 2188. [Google Scholar] [CrossRef] [PubMed]
  9. Krishnan, G.; Singh, S.; Pathania, M.; Gosavi, S.; Abhishek, S.; Parchani, A.; Dhar, M. Artificial Intelligence in Clinical Medicine: Catalyzing a Sustainable Global Healthcare Paradigm. Front. Artif. Intell. 2023, 6, 1227091. [Google Scholar] [CrossRef]
  10. Rakhmatulin, I.; Dao, M.S.; Nassibi, A.; Mandic, D. Exploring Convolutional Neural Network Architectures for EEG Feature Extraction. Sensors 2024, 24, 877. [Google Scholar] [CrossRef]
  11. Tanriver, G.; Soluk Tekkesin, M.; Ergen, O.; Koljenovickoljenovic, S.; Bossy, P.; Bouaoud, J. Automated Detection and Classification of Oral Lesions Using Deep Learning to Detect Oral Potentially Malignant Disorders. Cancers 2021, 13, 2766. [Google Scholar] [CrossRef]
  12. Ding, H.; Wu, J.; Zhao, W.; Matinlinna, J.P.; Burrow, M.F.; Tsoi, J.K.H. Artificial Intelligence in Dentistry—A Review. Front. Dent. Med. 2023, 4, 1085251. [Google Scholar] [CrossRef]
  13. Parola, M.; La Mantia, G.; Galatolo, F.; Cimino, M.G.C.A.; Campisi, G.; Di Fede, O. Image-Based Screening of Oral Cancer via Deep Ensemble Architecture. In Proceedings of the 2023 IEEE Symposium Series on Computational Intelligence (SSCI), Mexico City, Mexico, 5–8 December 2023; pp. 1572–1578. [Google Scholar] [CrossRef]
  14. Chen, Y.-W.; Stanley, K.; Att, W. Artificial Intelligence in Dentistry: Current Applications and Future Perspectives. Quintessence Int. 2020, 51, 248–257. [Google Scholar] [CrossRef]
  15. Alzaid, N.; Ghulam, O.; Albani, M.; Alharbi, R.; Othman, M.; Taher, H.; Albaradie, S.; Ahmed, S. Revolutionizing Dental Care: A Comprehensive Review of Artificial Intelligence Applications Among Various Dental Specialties. Cureus 2023, 15, e47033. [Google Scholar] [CrossRef] [PubMed]
  16. Hegde, S.; Ajila, V.; Zhu, W.; Zeng, C. Artificial Intelligence in Early Diagnosis and Prevention of Oral Cancer. Asia-Pac. J. Oncol. Nurs. 2022, 9, 100133. [Google Scholar] [CrossRef]
  17. Dixit, S.; Kumar, A.; Srinivasan, K. A Current Review of Machine Learning and Deep Learning Models in Oral Cancer Diagnosis: Recent Technologies, Open Challenges, and Future Research Directions. Diagnostics 2023, 13, 1353. [Google Scholar] [CrossRef]
  18. Al-Rawi, N.; Sultan, A.; Rajai, B.; Shuaeeb, H.; Alnajjar, M.; Alketbi, M.; Mohammad, Y.; Shetty, S.R.; Mashrah, M.A. The Effectiveness of Artificial Intelligence in Detection of Oral Cancer. Int. Dent. J. 2022, 72, 436. [Google Scholar] [CrossRef] [PubMed]
  19. Sathishkumar, R.; Govindarajan, M. A Comprehensive Study on Artificial Intelligence Techniques for Oral Cancer Diagnosis: Challenges and Opportunities. In Proceedings of the 2023 International Conference on System, Computation, Automation and Networking, ICSCAN 2023, Puducherry, India, 17–18 November 2023. [Google Scholar] [CrossRef]
  20. Khanagar, S.B.; Alkadi, L.; Alghilan, M.A.; Kalagi, S.; Awawdeh, M.; Bijai, L.K.; Vishwanathaiah, S.; Aldhebaib, A.; Singh, O.G. Application and Performance of Artificial Intelligence (AI) in Oral Cancer Diagnosis and Prediction Using Histopathological Images: A Systematic Review. Biomedicines 2023, 11, 1612. [Google Scholar] [CrossRef]
  21. Decroos, F.; Springenberg, S.; Lang, T.; Päpper, M.; Zapf, A.; Metze, D.; Steinkraus, V.; Böer-Auer, A. A Deep Learning Approach for Histopathological Diagnosis of Onychomycosis: Not Inferior to Analogue Diagnosis by Histopathologists. Acta Derm. Venereol. 2021, 101, 107. [Google Scholar] [CrossRef]
  22. Rajendran, S.; Lim, J.H.; Yogalingam, K.; Kallarakkal, T.G.; Zain, R.B.; Jayasinghe, R.D.; Rimal, J.; Kerr, A.R.; Amtha, R.; Patil, K.; et al. The Establishment of a Multi-Source Dataset of Images on Common Oral Lesions; Preprint; Research Square: Durham, NC, USA, 2021. [Google Scholar] [CrossRef]
  23. Dinesh, Y.; Ramalingam, K.; Ramani, P.; Deepak, R.M. Machine Learning in the Detection of Oral Lesions with Clinical Intraoral Images. Cureus 2023, 15, e44018. [Google Scholar] [CrossRef]
  24. Bonny, T.; Al Nassan, W.; Obaideen, K.; Al Mallahi, M.N.; Mohammad, Y.; El-Damanhoury, H.M. Contemporary Role and Applications of Artificial Intelligence in Dentistry. F1000Res 2023, 12, 1179. [Google Scholar] [CrossRef]
  25. Batra, A.M.; Reche, A. A New Era of Dental Care: Harnessing Artificial Intelligence for Better Diagnosis and Treatment. Cureus 2023, 15, e49319. [Google Scholar] [CrossRef]
  26. Roganović, J.; Radenković, M.; Miličić, B. Responsible Use of Artificial Intelligence in Dentistry: Survey on Dentists’ and Final-Year Undergraduates’ Perspectives. Healthcare 2023, 11, 1480. [Google Scholar] [CrossRef] [PubMed]
  27. Di Fede, O.; La Mantia, G.; Cimino, M.G.C.A.; Campisi, G. Protection of Patient Data in Digital Oral and General Health Care: A Scoping Review with Respect to the Current Regulations. Oral 2023, 3, 155–165. [Google Scholar] [CrossRef]
  28. Parola, M.; Galatolo, F.A.; La Mantia, G.; Cimino, M.G.C.A.; Campisi, G.; Di Fede, O. Towards Explainable Oral Cancer Recognition: Screening on Imperfect Images via Informed Deep Learning and Case-Based Reasoning. Comput. Med. Imaging Graph. 2024, 117, 102433. [Google Scholar] [CrossRef]
  29. Guo, J.; Wang, H.; Xue, X.; Li, M.; Ma, Z. Real-Time Classification on Oral Ulcer Images with Residual Network and Image Enhancement. IET Image Process 2022, 16, 641–646. [Google Scholar] [CrossRef]
  30. Fabiane, R.; Gomes, T.; Schmith, J.; Marques De Figueiredo, R.; Freitas, S.A.; Nunes Machado, G.; Romanini, J.; Coelho Carrard, V. Use of Artificial Intelligence in the Classification of Elementary Oral Lesions from Clinical Images. Int. J. Environ. Res. Public Health 2023, 20, 3894. [Google Scholar] [CrossRef]
  31. Ciecierski-Holmes, T.; Singh, R.; Axt, M.; Brenner, S.; Barteit, S. Artificial Intelligence for Strengthening Healthcare Systems in Low- and Middle-Income Countries: A Systematic Scoping Review. npj Digit. Med. 2022, 5, 162. [Google Scholar] [CrossRef] [PubMed]
  32. Gore, J.C. Artificial Intelligence in Medical Imaging. Magn. Reson. Imaging 2020, 68, A1–A4. [Google Scholar] [CrossRef]
  33. Mintz, Y.; Brodie, R. Introduction to Artificial Intelligence in Medicine. Minim. Invasive Ther. Allied Technol. 2019, 28, 73–81. [Google Scholar] [CrossRef]
Figure 1. The graph shows the mAP during the training of the Faster R-CNN model with ResNet-101. Initially, the validation mAP is low, but it rises as the model learns. The mAP fluctuates due to the complexity of the data but peaks at 47 min and 22 s, indicating the best model performance. This peak is marked by a dashed red line. After this point, the mAP stabilizes or slightly declines, suggesting that further training may lead to overfitting. In summary, the graph highlights the optimal model performance at 47 min and 22 s, balancing accuracy and generalization.
Figure 1. The graph shows the mAP during the training of the Faster R-CNN model with ResNet-101. Initially, the validation mAP is low, but it rises as the model learns. The mAP fluctuates due to the complexity of the data but peaks at 47 min and 22 s, indicating the best model performance. This peak is marked by a dashed red line. After this point, the mAP stabilizes or slightly declines, suggesting that further training may lead to overfitting. In summary, the graph highlights the optimal model performance at 47 min and 22 s, balancing accuracy and generalization.
Oral 04 00043 g001
Figure 2. Normalized confusion matrix for multiclass performance. The values on the diagonal represent the TP, indicating the proportion of correct classifications for each class. The proportions of misclassifications for each class are shown in the far-right column, labeled as “NC”, which represents FN. These are cases where the model failed to classify an image into any of the specified classes. The other cells outside the diagonal represent FP, where the model incorrectly classified an image as belonging to a particular class when it did not. Abbreviations: STR/RET, striate-reticular lesions; MB, morsicatio buccarum; LA, linea alba; LEU, leukoplakia; CON/PAP, condyloma–papilloma; CAN, oral candidiasis; NC, not classified.
Figure 2. Normalized confusion matrix for multiclass performance. The values on the diagonal represent the TP, indicating the proportion of correct classifications for each class. The proportions of misclassifications for each class are shown in the far-right column, labeled as “NC”, which represents FN. These are cases where the model failed to classify an image into any of the specified classes. The other cells outside the diagonal represent FP, where the model incorrectly classified an image as belonging to a particular class when it did not. Abbreviations: STR/RET, striate-reticular lesions; MB, morsicatio buccarum; LA, linea alba; LEU, leukoplakia; CON/PAP, condyloma–papilloma; CAN, oral candidiasis; NC, not classified.
Oral 04 00043 g002
Table 1. Number of photos per each class, divided into training and validation. Abbreviations: STR/RET, striate-reticular lesions; MB, morsicatio buccarum; LA, linea alba; LEU, leukoplakia; CON/PAP, condyloma–papilloma; CAN, oral candidiasis.
Table 1. Number of photos per each class, divided into training and validation. Abbreviations: STR/RET, striate-reticular lesions; MB, morsicatio buccarum; LA, linea alba; LEU, leukoplakia; CON/PAP, condyloma–papilloma; CAN, oral candidiasis.
STR/RETMBLALEUCON/PAPCAN
Training482420323223179
Validation126478542
Total603024394028221
Table 2. Statistics calculated for the whole dataset globally.
Table 2. Statistics calculated for the whole dataset globally.
ParameterValue
Precision77.2%
Sensitivity76.0%
F1 score74.4%
mAP82.3%
Table 3. Statistics calculated for each class individually.
Table 3. Statistics calculated for each class individually.
ClassPrecisionSensitivityF1-ScoremAP
Striae-reticular54.5%84.6%66.3%62.1%
Morsicatio buccarum80.0%66.7%72.7%88.9%
Linea alba80.0%100.0%88.9%90.0%
Leukoplakia60.0%25.0%35.3%56.8%
Condyloma–papilloma88.9%100.0%94.1%95.8%
Hyperplasticcandidiasis100.0%80.0%88.9%100.0%
Table 4. Summary of classification counts for multiclass performance. This table shows the counts of TP, FP, and FN for each class.
Table 4. Summary of classification counts for multiclass performance. This table shows the counts of TP, FP, and FN for each class.
Class TPFPFN
Striae-reticular1200
Morsicatio buccarum411
Linea alba400
Leukoplakia403
Condyloma-papilloma800
Hyperplasticcandidiasis401
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

La Mantia, G.; Kiswarday, F.; Pizzo, G.; Giuliana, G.; Oteri, G.; Cimino, M.G.C.A.; Di Fede, O.; Campisi, G. Detection of Elementary White Mucosal Lesions by an AI System: A Pilot Study. Oral 2024, 4, 557-566. https://doi.org/10.3390/oral4040043

AMA Style

La Mantia G, Kiswarday F, Pizzo G, Giuliana G, Oteri G, Cimino MGCA, Di Fede O, Campisi G. Detection of Elementary White Mucosal Lesions by an AI System: A Pilot Study. Oral. 2024; 4(4):557-566. https://doi.org/10.3390/oral4040043

Chicago/Turabian Style

La Mantia, Gaetano, Federico Kiswarday, Giuseppe Pizzo, Giovanna Giuliana, Giacomo Oteri, Mario G. C. A. Cimino, Olga Di Fede, and Giuseppina Campisi. 2024. "Detection of Elementary White Mucosal Lesions by an AI System: A Pilot Study" Oral 4, no. 4: 557-566. https://doi.org/10.3390/oral4040043

APA Style

La Mantia, G., Kiswarday, F., Pizzo, G., Giuliana, G., Oteri, G., Cimino, M. G. C. A., Di Fede, O., & Campisi, G. (2024). Detection of Elementary White Mucosal Lesions by an AI System: A Pilot Study. Oral, 4(4), 557-566. https://doi.org/10.3390/oral4040043

Article Metrics

Back to TopTop