You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

31 July 2023

Teeth Lesion Detection Using Deep Learning and the Internet of Things Post-COVID-19

,
,
,
,
,
and
1
College of Electrical and Mechanical Engineering, National University of Sciences and Technology (NUST), Islamabad 44000, Pakistan
2
Abasyn University Islamabad Campus, Islamabad 44000, Pakistan
3
Higher Polytechnic School, Universidad Europea del Atlántico, Isabel Torres 21, 39011 Santander, Spain
4
Universidad Internacional Iberoamericana, Campeche 24560, Mexico
This article belongs to the Special Issue Machine Learning for Biomedical Sensing and Healthcare

Abstract

With a view of the post-COVID-19 world and probable future pandemics, this paper presents an Internet of Things (IoT)-based automated healthcare diagnosis model that employs a mixed approach using data augmentation, transfer learning, and deep learning techniques and does not require physical interaction between the patient and physician. Through a user-friendly graphic user interface and availability of suitable computing power on smart devices, the embedded artificial intelligence allows the proposed model to be effectively used by a layperson without the need for a dental expert by indicating any issues with the teeth and subsequent treatment options. The proposed method involves multiple processes, including data acquisition using IoT devices, data preprocessing, deep learning-based feature extraction, and classification through an unsupervised neural network. The dataset contains multiple periapical X-rays of five different types of lesions obtained through an IoT device mounted within the mouth guard. A pretrained AlexNet, a fast GPU implementation of a convolutional neural network (CNN), is fine-tuned using data augmentation and transfer learning and employed to extract the suitable feature set. The data augmentation avoids overtraining, whereas accuracy is improved by transfer learning. Later, support vector machine (SVM) and the K-nearest neighbors (KNN) classifiers are trained for lesion classification. It was found that the proposed automated model based on the AlexNet extraction mechanism followed by the SVM classifier achieved an accuracy of 98%, showing the effectiveness of the presented approach.

1. Introduction

Dental diseases such as periodontal disease, dental caries, and fluorosis affect a large population worldwide. However, their prevalence varies among different geographical distributions. Based on the research conducted by World Health Organization, dental diseases affect 621 million children, and almost 2.40 billion adults are affected by dental caries [1]. Furthermore, results obtained from the National Health and nutrition examination survey show that almost 41% of children between the age of 2–11, 42% of adolescents from 6–19 years of age, and 90% of adults over the age of 20 are affected by dental caries [2].
Periapical radiographs, a kind of intraoral radiograph, are widely used in endodontics. Periapical diseases are substantially inflammatory lesions that lead to dental caries causing detrimental injury to the teeth. These diseases are classified as apical cysts, abscesses, and granulomas affecting the dental pulp protected by enamel, cementum, and dentin. If left untreated, chronic damage to the pulp chamber can lead to inflammation, eventually turning into pulp necrosis and periradicular pathosis. Scenarios triggering periapical radiolucency include trauma, tooth wear, or caries. However, it is possible to prevent their spread through non-surgical endodontic treatment. Most periapical lesions heal through meticulous nonsurgical endodontic treatments. A timeframe of 6 to 12 months is required for assessing the healing potential, after which root canal treatment should be considered [3,4]. Additionally, complete healing of the lesion might take up to four years if not diagnosed in time. This requires frequent visits and interactions with the dentist and associated staff, which has become difficult in the post-COVID-19 environment. Postponing treatment can increase the risk of tooth fracture [5]. Therefore, it is essential to diagnose teeth lesions in a timely way through routine oral examination of the teeth and gums as well as the soft tissues in and around the mouth. Conventional periapical radiographs are obtained by exposing X-ray radiation that is then processed chemically to produce images. This film-based conventional method has certain drawbacks, and digital radiography has been introduced to overcome these drawbacks. This process involves acquiring images digitally which are then manipulated using a computer. There are several ways of obtaining such images, including intraoral sensors, charge-coupled devices, and scanning of radiographs [6]. Due to the emergence and spread of low-cost connectivity in underdeveloped and developing countries, the Internet of Things (IoT) has enabled dental services to be more broadly leveraged, providing better dental health in developing countries and decreasing the overall demand for dental care. Looking at the successful adoption of the IoT in different sectors [7,8], it can serve as an attractive area of adoption in dentistry. IoT-enabled devices can obtain information, including pathological parameters, to predict oral health and make decisions regarding the treatment of the disease at earlier stages. With the growth in IoT-based healthcare services, it has become possible to achieve significant enhancement in dental healthcare, including early detection and prediction [9].
Recent advancements in the microcontroller-based wearable and implantable smart electronic devices has given rise to touchless technologies, which are gaining momentum and are regarded as the future of medical science. Such wearables are data transmission devices worn on the human body, such as the eyes, ears, knees, feet, fingers, and hands. These devices can detect, process, analyze, receive, and transmit related vital body signals and information such as the pulse rate, heart rate, body temperature, and other ambient data which do not require human intervention and may allow immediate biofeedback to the wearer [10,11]. They are an essential part of an IoT scenario in which they allow the exchange of high-quality data through the global internet via processor and relay agents to the operator, data collector, company, and/or manufacturer.
As a result, effective biomedical monitoring systems can be designed for medical diagnosis, physiological and psychological health monitoring, and evaluation through continuous measurement of critical biomarkers from various human body parts. A possible range of wearable devices in the form of basic clothing, common accessories, attachments, and body implants is shown in Figure 1.
Figure 1. Wearable healthcare devices.
The information available from these sensors can be transmitted through modern communication links, as shown in Figure 2, for intelligent processing and storage in an IoT environment. The IoT infrastructure makes possible the monitoring and recording of real-time temporal characteristics of chronic/acute patients, allowing for both immediate diagnosis and further research around finding and tracking disease trends.
Figure 2. IoT communication infrastructure.
Wearable technology is expanding into newer applications, and has gained significant space in consumer electronics through smartwatches, activity trackers, advanced e-textiles, healthcare products, navigation systems, and geofencing applications. Within the healthcare domain, parameters such as blood pressure, exercise time, steps walked/ran, running speed, calories burned, heart rate, body temperature, pulse rate, seizures, physical stress/strain, and level of certain body chemicals can be measured and used to evaluate users’ health. The issue of power sourcing for wearables has been suitably addressed by various methods as well. Smart wristbands can be used to generate their own electrical signals through thermoelectric generation technology (TEG) [12]. Power can be generated using the temperature differential between human skin and the environmental temperature. These implants have been used to monitor the diet and nutrition content of meals, and play a great role in improving diet plans, leading to a more healthy and fit life [13]. These devices collect information about salt, glucose, and even alcohol consumption, can share these data with other smart devices that a person uses, and are capable of updating physicians about diet routines and health conditions.
Additionally, IoT infrastructure allows monitoring and recording of the real-time temporal characteristics of chronic/acute patients for immediate diagnosis and tracing of disease trends. For diagnosing dental caries, digital radiographs are employed by dentists to assess and evaluate the extent of caries and to determine the need for treatment. Among the common intraoral types of radiographs (periapical, bitewing, and panoramic), periapical radiographs are the more common, as they provide localized information related to the length and adequacy of caries and periodontal ligament space [14]. Furthermore, challenges faced in interpreting conventional radiographs create the potential for inconsistencies among dentists due to a number of factors, including contrast variation, magnification, and lack of experience. Nevertheless, due to the clinical reliability of periapical radiographs these have become a popular choice for diagnosing dental caries. Additionally, the above-mentioned challenges pave the way for utilizing automated solutions for improving diagnosis and standardization of care.
Recently, deep learning-based techniques have demonstrated excellent performance in computer vision, including object detection, tracking, and recognition, improving the ability to build software for automated analysis and evaluation of images. Different deep learning techniques have demonstrated the potential for automated identification of radiological and pathological features. Furthermore, different image processing and recognition procedures have been adopted for medical segmentation, with high accuracy and efficiency observed in the classification of different diseases using deep learning-based models, including cystic lesions [15], skin cancer [16], COVID-19 [17], and thoracic diseases [18], demonstrating improved accuracy and efficiency.
On the contrary, few studies have been observed based on deep learning, specifically deep convolutional neural networks (CNNs), in the dental field, limiting research investigating the diagnosis and detection of dental caries. Additionally, limited attempts at automatic dental radiograph analysis using deep learning have been observed from previous studies [19,20,21]. Other than this, variable accuracy is observed in the methods currently utilized, highlighting the need for further research in this field. Approaches involving CNN and transfer learning have been used in dental diseases based on dental X-rays [22,23].
However, due to ease of classification, most approaches consider specific issues such as periodontitis, periapical lesions, and dental caries. Other teeth lesions, such as endo-perio and perio-endo lesions, are more challenging to diagnose due to their closely matching attributes. Therefore, they have been grouped into five significant lesions: primary endodontic with or without secondary periodontal involvement, primary periodontal with or without secondary endodontic involvement, and true combined lesion [24,25].
There are several limitations and challenges in the existing approaches. First, most of the work relies on traditional machine learning techniques, which may not be effective in detecting complex lesions due to the closely matching attributes of dental lesions, making them difficult to diagnose using traditional methods. Second, data insufficiency is a common challenge, as there may not be enough data available to train accurate models.
Keeping in view the limitations of previous studies, the present study aims to predict five types of endo-perio and perio-endo lesions in periapical radiographs by adopting a hybrid approach using transfer learning and a fine-tuning method after acquiring radiographs from a wearable device through an IoT infrastructure. Additionally, the presented work compares the diagnosis efficacy of different machine learning-based techniques, including support vector machine (SVM) and K-nearest neighbors (KNN), on features extracted using a pretrained AlexNet. Finally, the objective is to evaluate the efficacy of CNN algorithms and transfer learning to detect and classify dental caries in periapical radiographs and to overcome data insufficiency using data augmentation techniques. The main contributions of this research include:
  • Effective detection and classification of dental caries in periapical radiographs by leveraging deep learning techniques such as transfer learning and image augmentation;
  • Integrating IoT technology through an IoT-enabled device to capture tooth lesions from radiographs, allowing for remote monitoring and management of dental health, which is significant in situations where in-person visits to dental clinics might be limited.
The rest of the paper is structured in five sections, with Section 2 presenting a literature review on wearables in the healthcare sector and machine, deep, and transfer learning concepts and techniques used for diagnosis of teeth lesions in dental X-rays, Section 3 explaining the methodology of the proposed framework, Section 4 discussing the results and providing a comparative analysis, and Section 5 concluding the paper.

3. Methods and Materials

Figure 3 illustrates the proposed system for automatic dental lesion detection. The automatic classification of teeth lesions is divided into five steps: (1) IoT-enabled data retrieval and dataset preparation (2) data augmentation; (3) feature extraction; (4) classification; and (5) improved accuracy and detection.
Figure 3. The architecture of the proposed framework.

3.1. Materials

3.1.1. IoT to Enable Data Retrieval

The Internet of things (IoT) is a collection of physical objects which enable collection and exchange of information over the network. IoT devices allow gathering real-time data from sensors to allow better decision-making [42]. The IoT-enabled dental film utilized in the present work involves a circuit board with a proximity sensor to ensure that the film is placed securely on the patient’s teeth, another sensor to measure the rotation of the head in order to acquire a proper angle, and a WiFi module for data transmission. The IoT-enabled dental film is placed on the extension cone parallel (XCP) device to take periapical radiographs of the posterior teeth [43]. The placement for periapical radiographs and the regions captured is illustrated below in Figure 4. The face of the dental film is positioned perpendicular to the interdental space between the teeth. The film is guided towards the midline, the patient bites into the bite block, and then the X-ray is taken [44].
Figure 4. IoT-enabled dental film placement for periapical radiographs.
After acquiring the radiographs, the WiFi module (ESP8266) mounted inside the circuit transmits the images automatically to the laptop/smartphone for further analysis. This module is specifically designed for IoT for wireless connectivity [45]. The proposed system can help to reduce the proximal surface overlap in periapical radiography [46]. Experimentation was carried out at a dental institute in Rawalpindi, Pakistan, and 534 periapical radiographs labeled by experienced clinicians were collected. Figure 5 below depicts the process of wireless radiograph transmission from IoT-enabled dental film to a laptop/tablet/smartphone screen or any smart device through WiFi.
Figure 5. Wireless radiograph transmission.

3.1.2. Data Preparation

The periapical radiograph dataset is divided into two sets for training and testing, containing 453 and 81 images, respectively. The data division ratio is shown below in Figure 6.
Figure 6. Traing and test dataset ratio.
Radiographs captured and transmitted through IoT-enabled devices were labeled for their classes by experts. The class distribution for the training and test datasets is shown in Table 2.
Table 2. Periapical radiographs dataset.

3.2. Methods

3.2.1. Data Augmentation

Considering the relatively small size of the dataset, different data augmentation techniques were employed prior to training and evaluation of the different deep learning architectures’ performance. Data augmentation is a common preprocessing technique that is employed before feeding the data samples to the neural network for training. It entails creating artificial data from known data. This helps to mitigate the problem of overtraining and the adverse impact of class imbalance, leading to improved model accuracy. Figure 7 depicts how the augmented image datastore was used to transform the training data for each epoch. During this process, one randomly augmented version of each image was used during each training epoch.
Figure 7. Data augmentation workflow.
The training dataset was augmented using an augmentation technique defined through suitable function classes, such as RandXReflection and RandYReflection, in Matlab. In addition to applying the data augmentation techniques, images input to Alexnet were resized and converted to color from grayscale. Using the data augmentation techniques, the number of training images was increased from 453 to 1359. Using RandYReflection, each image was reflected horizontally as shown in Figure 8.
Figure 8. Original image and horizontally random reflected image.
Figure 9 shows the RandXReflection-generated vertical reflection of each image. The images were augmented during training and were not saved in memory, and when the network parameters were trained, the augmented images were automatically discarded.
Figure 9. Original image and random vertically reflected image.

3.2.2. Feature Extraction (Pre-Trained AlexNet)

A pre-trained AlexNet was used in the framework for feature extraction. This network comprises a large network structure with 60 million parameters and 650,000 neurons. Several improvements have been made to train the parameters [47]. One of the most significant improvements is the activation function. These functions provide nonlinearity within the neural network. One of the activation functions that helps in avoiding gradient vanishing problems is the rectified linear unit (ReLU) function. The calculation process of the ReLU function is
R e L U ( x ) = m a x ( x , o ) .
Furthermore, the ReLU function tends to converge faster than other activation functions in deep networks. A dropout layer is employed to avoid overfitting by forcing neurons to cooperate with others, resulting in improved generalization. The fully connected layers within the network are used for classification. The activation function utilized with the final layer was the softmax function, which constrains the output to a range of (0,1). Equation (2) below expresses the softmax function:
s o f t m a x ( x ) = e x p ( x i ) j = 1 n e x p ( x j ) .
The AlexNet model used in this study was made up of 25 layers, containing one input layer, five convolution layers, seven ReLU layers, two normalization layers, three max-pooling layers, and three fully connected layers.

3.2.3. AlexNet Fine-Tuning Using Transfer Learning

The periapical radiographs dataset contained only a few hundred samples, which is insufficient to train a deep network of large volume such as AlexNet. Therefore, transfer learning was employed to make the network more suitable for feature selection and reduce the time required for classification. The entire structure of the network was divided into two parts: pretrained networks and transferred networks. The last three layers of the AlexNet were replaced: a fully connected layer with 1000 neurons to receive the extracted features and map them to output categories, another SoftMax layer, and a classification layer to output the classes; the rest of the model was preserved. The network parameters were already pretrained on ImageNet, making the extracted features effective for classification.
Transfer learning is a process that involves taking a pretrained network, modifying it, and retraining it on new data. This technique requires little data and computational time, as shown in Figure 10.
Figure 10. Effort vs. flexibility for transfer learning.
Transfer learning is an effective and convenient method for training deep neural networks even when the system lacks sufficient labeled samples. Furthermore, transfer learning can be implemented using ordinary personal computers, making it more suitable for training networks with computational resource constraints. The workflow of transfer learning is shown in Figure 11.
Figure 11. Workflow of the transfer learning approach.
The learning rate was fixed at 0.0001 and the momentum value was kept the same as in the pretrained AlexNet, i.e., 0.9. The mini-batch size was set to 8 and the number of epochs was kept at 100. Using an NVIDIA graphics card (GeForce 930MX), the total training time was computed at around 50 min and 51 s, as depicted in Figure 12.
Figure 12. Training process of the model.

3.2.4. Classification

Using the fine-tuned AlexNet, 4096 features were extracted from each periapical radiograph. The extracted features were then used to train two classifiers: support vector machine (SVM) and K-nearest neighbors (KNN). SVM allows for solving classification and regression-related problems by utilizing multi-dimensional hyperplanes to separate the data with more significant gaps between them, while KNN is a supervised data classifier based on lazy learning. It allows for storing data classes and classifying unseen data based on feature similarities from the known classes. Additionally, different functions can be used for feature comparisons, such as Manhattan, Euclidean, and Minkowski. Finally, the anonymous data are classified according to the nearest neighbor, with K being the number of neighbors.

3.2.5. Performance Evaluation Metrics

The performance of the proposed approach was evaluated using accuracy for five classes by dividing the number of correct predictions by the total number of test X-rays. Equation (3) was used to evaluate accuracy [48]:
A c c u r a c y = T P + T N T P + T N + F P + F N .
Accuracy is an important evaluation indicator in neural network models. To calculate the accuracy, the concepts of true positive (TP), true negative (TN), false positive (FP), and false negative (FN) are introduced [49].

3.2.6. Error Analysis

In order to calculate the error between the predicted labels and the predicted value from the network, the Huber loss function [50] was used. The Adam optimizer was applied after the error was calculated to update the network biases and weights in each layer. The Hubert loss function leads to robust training and has proven to be effective in tracking outliers. It is defined as
f h u b e r = σ a 2 2 | a | 1 2 a > σ , a < σ ,
where σ is the threshold.

4. Experiments, Results, and Discussions

4.1. Settings

The proposed method was implemented using Matlab R2017a with an i5-4770 CPU and NVIDIA GeForce 930MX graphics card. The trained structure can run on any personal computer with Matlab. Additionally, 80% of the dataset was used for training, and the rest was used for testing. The details of the dataset after augmentation are listed in Table 3.
Table 3. Training and testing data et of periapical X-rays.

4.2. Experiments

4.2.1. First Set of Experiments

The features were extracted using the pretrained Alexnet and the extracted features were used to train SVM and KNN classifiers. Feature extraction using the pretrained deep convolution neural network was effortless and quick. A suitable function from the statistic and machine learning toolbox was used to train the multi-class SVM. Similarly, an appropriate function from the statistic and machine learning toolbox was used to train the KNN on the features extracted from the training data. The confusion matrix function was used to show the confusion matrix for the predictions of both classifiers. The SVM classifier performed better than the KNN for this set of experiments. The classification accuracy achieved by the two classifiers is shown in Figure 13.
Figure 13. Confusion matrices for first set of experiments.

4.2.2. Second Set of Experiments

In the second set of experiments, the pretrained Alexnet network was fine-tuned with the augmented data. This retrained network was then used to classify the test examples using the softmax function and the same retrained model was used for feature extraction. The ReLU layer (21st layer) of the retrained model was used to output the extracted features. The SVM and KNN were then retrained on these extracted features by changing the activation function.
The classification accuracy in Figure 14 shows that the SVM trained on the features extracted by the retrained network fine-tuned on augmented data performed better than the other two classifiers, i.e., the KNN classifier and the retrained (softmax classifier).
Figure 14. Results for the second set of experiments.
The classification of test data by the retrained model misclassified two examples. It wrongly predicted a primary endo lesion instead of a primary perio with secondary endo lesion and a primary periodontal lesion for a primary endo lesion, as shown in Figure 14a. The next figure shows the confusion matrix of the SVM trained on features extracted by the retrained model, demonstrating that the trained SVM is unable to classify one test example. The KNN classifier trained on the features extracted by the retrained model wrongly classified four test examples, as shown in Figure 14c.

4.3. Fine-Tuned AlexNet without Data Augmentation

The pretrained network (AlexNet) was fine-tuned using the training data without augmentation and its performance was compared against the AlexNet trained with data augmentation in order to establish the efficacy of the former in classifying test images using the softmax function. The training set was kept at 453 training examples and the network’s performance versus the amount of data provided for training was computed. It can be observed that the performance of the pretrained Alexnet deteriorated without the data augmentation approach.
A total of 81 test examples were passed through the retrained model for classification, and 35 examples were misclassified by the retrained network. The same test data were presented to the SVM and KNN classifiers trained on this retrained AlexNet model. The results show that lower accuracy was achieved by the classifiers when data augmentation was not performed. The classification accuracy depicted in Table 4 indicates that the SVM trained on features extracted by the retrained model fine-tuned with the augmented data performed better than the other classifiers. Additionally, Table 5 shows the precision, recall, and f1-score achieved using the proposed classifier. Figure 15 shows the accuracy of classifiers for different scenarios.
Table 4. Accuracy Results for Classifiers.
Table 5. Precision, recall, and F1 score results for classifiers.
Figure 15. Percentage accuracy of all classifiers.
The IoDT has great potential in opening up new horizons in the dental field. The proposed work incorporates a IoT-based dental film to capture and transfer X-rays automatically to ensure that radiographs can be analyzed effectively. The fine-tuned AlexNet was trained using periapical radiographs belonging to different classes. The minimum batch size was set to 8 to avoid low memory issues. The minimum learning rate was set to 0.0001 to mimic a slow learning pace, allowing newly added layers to catch up to existing layers. To show the efficacy of the proposed approach, it was compared with several state-of-the-art methods, as shown in Table 4. The results show that the SVM trained on features extracted by the retrained model fine-tuned with the augmented data performs better compared to other classifiers. Furthermore, the loss ratio is small even when using a smaller dataset. These good results are achieved through transfer learning. Testing was carried out using 81 images, and confusion matrices were produced for the different sets of experiments.
Furthermore, other advanced deep learning-based methods can be tried for periapical radiographs. More robust data augmentation could be applied to further enhance the dataset to achieve improved results. The model can perform multi-class classification on periapical radiographs. However, other intraoral radiographs could be used to further validate the efficacy of the proposed model. The future management of dental caries will be based on three factors: early disease detection, assessment of risks, and prevention. The aim of this study is to provide affordable alternatives for patients that can be used effectively at the community level.

4.4. Performance Analysis Using State-of-the-Art Approaches

To corroborate the efficacy of the proposed approach, a performance appraisal was carried out with other state-of-the-art approaches; the results are provided in Table 6. The proposed method achieves an accuracy of 98% using data augmentation and 75% without data augmentation. These results confirm the superior performance of the proposed approach over recent approaches used for diagnosis of teeth lesions.
Table 6. Performance comparison with state-of-the art models.

4.5. Strengths and Limitations of the Proposed Approach

The proposed approach has several strengths and limitations that are worth considering when evaluating the real-time deployment of the proposed system.
The proposed approach utilizes deep learning-based techniques such as data augmentation and transfer learning, which are more effective in detecting complex dental lesions. Furthermore, more accurate and reliable data collection can be ensured by integrating IoT-enabled devices. The need for expert involvement is significantly reduced, making the system more accessible to patients.
However, there are certain limitations to the proposed approach. First, it requires access to periapical radiographs, which might not be readily available due to privacy concerns. Second, the accuracy of the proposed system relies on the accuracy of the IoT-enabled device used for data collection. This can be influenced by different factors, such as device calibration and lighting conditions. Third, the proposed model may not be as effective in detecting other types of periapical lesions, as the model was specifically trained on periapical radiographs. Lastly, there may be scalability issues in resource-constrained environments, which may have a negative impact on the performance of the proposed model.
Keeping in view these limitations, several factors could influence the real-world deployment of the proposed system. These include the availability of periapical radiographs in dental clinics where the system is to be deployed, the cost and accessibility of IoT-enabled devices, the availability of computational resources, and regulatory ethical considerations related to the use of patient data for model training and testing.

5. Conclusions and Future Work

This work has presented an endo perio lesion detection system using the transfer learning and data augmentation approach involving multiple processes without involving an expert dentist. The proposed technique utilizes data acquisition using an IoT-enabled mouth guard; the data are preprocessed and used to extract suitable features with a pretrained AlexNet CNN through an augmentation approach. We additionally investigated the utilization of transfer learning in dental lesion recognition based on intraoral radiographs. CNNS trained using transfer learning require a lower amount of data compared to CNNs trained from scratch. The proposed model utilizes a pretrained AlexNet, and fine-tuned classification was performed using the SVM classifier to recognize five classes using 453 images, then tested on 81 images. The performance of the proposed method was evaluated against other conventional classifiers and achieved 98% accuracy, demonstrating the effectiveness of our approach. The proposed model can be applied in daily clinical diagnosis to help dentists make decisions to improve patient care.
Even though the model provides encouraging results, it could be further improved by utilizing more robust data augmentation techniques and more the resource-hungry deep learning methods at the cost of computing resources. Additionally, the use of dental cone beam computed tomography images, bitewing radiographs, and panoramic X-ray images could be considered in future work. Moreover, the design and development of a dashboard in the form of a smartphone application to allow the user to obtain intelligent feedback and track dental health issues could be undertaken as well.

Author Contributions

Conceptualization, I.S. and M.S.; Data curation, M.S. and A.F.; Formal analysis, I.S. and A.F.; Funding acquisition, I.d.l.T.D.; Investigation, I.d.l.T.D.; Methodology, A.F. and D.G.A.; Project administration, D.G.A. and V.L.; Resources, I.d.l.T.D.; Software, V.L.; Supervision, I.A.; Validation, I.S.; Visualization, D.G.A. and V.L.; Writing—original draft, I.S. and M.S.; Writing—review and editing, I.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the European University of the Atlantic.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tagra, H.; Batra, P. Dentistry 4.0: A Whole New Paradigm. Dicoveries Rep. 2021, 4, 1–7. [Google Scholar]
  2. Lee, J.H.; Kim, D.H.; Jeong, S.N.; Choi, S.H. Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm. J. Dent. 2018, 77, 106–111. [Google Scholar] [CrossRef] [PubMed]
  3. Torres-Lagares, D.; Segura Egea, J.J.; Rodríguez-Caballero, A.; Llamas Carreras, J.M.; Gutiérrez Pérez, J.L. Treatment of a Large Maxillary Cyst with Marsupialisation, Decompression, Surgical Endodontic Therapy and Enucleation. J. Can. Dent. Assoc. 2011, 77, b87. [Google Scholar]
  4. Ørstavik, D. Time-course and risk analyses of the development and healing of chronic apical periodontitis in man. Int. Endod. J. 1996, 29, 150–155. [Google Scholar] [CrossRef]
  5. Karamifar, K.; Tondari, A.; Saghiri, M.A. Endodontic periapical lesion: An overview on the etiology, diagnosis and current treatment Modalities. Eur. Endod. J. 2020, 5, 54. [Google Scholar] [CrossRef] [PubMed]
  6. Almanei, K.; Alsulaimani, R.; Alfadda, S.; Albabtain, S.; Alsulaimani, R. Digitally Scanned Radiographs versus Conventional Films for Determining Clarity of Periapical Lesions and Quality of Root Canal Treatment. Sci. World J. 2017, 2017, 2427060. [Google Scholar] [CrossRef]
  7. Kan, C.; Chen, Y.; Leonelli, F.; Yang, H. Mobile sensing and network analytics for realizing smart automated systems towards health internet of things. In Proceedings of the 2015 IEEE International Conference on Automation Science and Engineering (CASE), Gothenburg, Sweden, 24–28 August 2015; IEEE: Manhattan, NY, USA, 2015; pp. 1072–1077. [Google Scholar]
  8. Ma, Y.; Wang, Y.; Yang, J.; Miao, Y.; Li, W. Big health application system based on health internet of things and big data. IEEE Access 2016, 5, 7885–7897. [Google Scholar] [CrossRef]
  9. Liu, L.; Xu, J.; Huan, Y.; Zou, Z.; Yeh, S.C.; Zheng, L.R. A smart dental health-IoT platform based on intelligent hardware, deep learning, and mobile terminal. IEEE J. Biomed. Health Inform. 2019, 24, 898–906. [Google Scholar] [CrossRef] [PubMed]
  10. Düking, P.; Achtzehn, S.; Holmberg, H.C.; Sperlich, B. Integrated framework of load monitoring by a combination of smartphone applications, wearables and point-of-care testing provides feedback that allows individual responsive adjustments to activities of daily living. Sensors 2018, 18, 1632. [Google Scholar] [CrossRef]
  11. Berdouses, E.D.; Koutsouri, G.D.; Tripoliti, E.E.; Matsopoulos, G.K.; Oulis, C.J.; Fotiadis, D.I. A computer-aided automated methodology for the detection and classification of occlusal caries from photographic color images. Comput. Biol. Med. 2015, 62, 119–135. [Google Scholar] [CrossRef]
  12. O’donoghue, J.; Herbert, J. Data management within mHealth environments: Patient sensors, mobile devices, and databases. J. Data Inf. Qual. (JDIQ) 2012, 4, 1–20. [Google Scholar] [CrossRef]
  13. Liu, Y.; Yin, L.; Zhang, W.; Wang, J.; Hou, S.; Wu, Z.; Zhang, Z.; Chen, C.; Li, X.; Ji, H.; et al. A wearable real-time power supply with a Mg3Bi2-based thermoelectric module. Cell Rep. Phys. Sci. 2021, 2, 100412. [Google Scholar] [CrossRef]
  14. Khan, H.A.; Haider, M.A.; Ansari, H.A.; Ishaq, H.; Kiyani, A.; Sohail, K.; Muhammad, M.; Khurram, S.A. Automated feature detection in dental periapical radiographs by using deep learning. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2021, 131, 711–720. [Google Scholar] [CrossRef] [PubMed]
  15. Lee, J.H.; Kim, D.H.; Jeong, S.N. Diagnosis of cystic lesions using panoramic and cone beam computed tomographic images based on deep learning neural network. Oral Dis. 2020, 26, 152–158. [Google Scholar] [CrossRef]
  16. Nahata, H.; Singh, S.P. Deep learning solutions for skin cancer detection and diagnosis. In Machine Learning with Health Care Perspective; Springer: Berlin/Heidelberg, Germany, 2020; pp. 159–182. [Google Scholar]
  17. Farooq, M.; Hafeez, A. Covid-resnet: A deep learning framework for screening of covid19 from radiographs. arXiv 2020, arXiv:2003.14395. [Google Scholar]
  18. Wang, H.; Xia, Y. Chestnet: A deep neural network for classification of thoracic diseases on chest radiography. arXiv 2018, arXiv:1807.03058. [Google Scholar]
  19. Kühnisch, J.; Meyer, O.; Hesenius, M.; Hickel, R.; Gruhn, V. Caries Detection on Intraoral Images Using Artificial Intelligence. J. Dent. Res. 2021, 101, 00220345211032524. [Google Scholar] [CrossRef]
  20. Duong, D.L.; Kabir, M.H.; Kuo, R.F. Automated caries detection with smartphone color photography using machine learning. Health Inform. J. 2021, 27, 14604582211007530. [Google Scholar] [CrossRef] [PubMed]
  21. Lee, S.; Oh, S.i.; Jo, J.; Kang, S.; Shin, Y.; Park, J.W. Deep Learning for Early Dental Caries Detection in Bitewing Radiographs. Sci. Rep. 2021, 11, 16807. [Google Scholar] [CrossRef]
  22. Prajapati, S.A.; Nagaraj, R.; Mitra, S. Classification of dental diseases using CNN and transfer learning. In Proceedings of the 2017 5th International Symposium on Computational and Business Intelligence (ISCBI), Dubai, United Arab Emirates, 11–14 August 2017; IEEE: Manhattan, NY, USA, 2017; pp. 70–74. [Google Scholar]
  23. Mao, Y.C.; Chen, T.Y.; Chou, H.S.; Lin, S.Y.; Liu, S.Y.; Chen, Y.A.; Liu, Y.L.; Chen, C.A.; Huang, Y.C.; Chen, S.L.; et al. Caries and restoration detection using bitewing film based on transfer learning with CNNs. Sensors 2021, 21, 4613. [Google Scholar] [CrossRef]
  24. Li, S.; Fevens, T.; Krzyżak, A.; Jin, C.; Li, S. Semi-automatic computer aided lesion detection in dental X-rays using variational level set. Pattern Recognit. 2007, 40, 2861–2873. [Google Scholar] [CrossRef]
  25. Lin, P.L.; Huang, P.Y.; Huang, P.W. An automatic lesion detection method for dental X-ray images by segmentation using variational level set. In Proceedings of the 2012 International Conference on Machine Learning and Cybernetics, Xian, China, 15–17 July 2012; IEEE: Manhattan, NY, USA, 2012; Volume 5, pp. 1821–1825. [Google Scholar]
  26. Javaid, M.; Haleem, A.; Singh, R.P.; Suman, R. Dentistry 4.0 technologies applications for dentistry during COVID-19 pandemic. Sustain. Oper. Comput. 2021, 2, 87–96. [Google Scholar] [CrossRef]
  27. Tyas, M.J.; Anusavice, K.J.; Frencken, J.E.; Mount, G.J. Minimal intervention dentistry—A review* FDI Commission Project 1–97. Int. Dent. J. 2000, 50, 1–12. [Google Scholar] [CrossRef]
  28. Cotti, E.; Campisi, G.; Ambu, R.; Dettori, C. Ultrasound real-time imaging in the differential diagnosis of periapical lesions. Int. Endod. J. 2003, 36, 556–563. [Google Scholar] [CrossRef]
  29. Pitcher, B.; Alaqla, A.; Noujeim, M.; Wealleans, J.A.; Kotsakis, G.; Chrepa, V. Binary decision trees for preoperative periapical cyst screening using cone-beam computed tomography. J. Endod. 2017, 43, 383–388. [Google Scholar] [CrossRef]
  30. Patil, S.; Kulkarni, V.; Bhise, A. Algorithmic analysis for dental caries detection using an adaptive neural network architecture. Heliyon 2019, 5, e01579. [Google Scholar] [CrossRef] [PubMed]
  31. Prados-Privado, M.; García Villalón, J.; Martínez-Martínez, C.H.; Ivorra, C.; Prados-Frutos, J.C. Dental Caries Diagnosis and Detection Using Neural Networks: A Systematic Review. J. Clin. Med. 2020, 9, 3579. [Google Scholar] [CrossRef]
  32. Casalegno, F.; Newton, T.; Daher, R.; Abdelaziz, M.; Lodi-Rizzini, A.; Schürmann, F.; Krejci, I.; Markram, H. Caries detection with near-infrared transillumination using deep learning. J. Dent. Res. 2019, 98, 1227–1233. [Google Scholar] [CrossRef] [PubMed]
  33. Geetha, V.; Aprameya, K.; Hinduja, D.M. Dental caries diagnosis in digital radiographs using back-propagation neural network. Health Inf. Sci. Syst. 2020, 8, 1–14. [Google Scholar] [CrossRef]
  34. Schwendicke, F.; Elhennawy, K.; Paris, S.; Friebertshäuser, P.; Krois, J. Deep learning for caries lesion detection in near-infrared light transillumination images: A pilot study. J. Dent. 2020, 92, 103260. [Google Scholar] [CrossRef]
  35. Karimian, N.; Salehi, H.S.; Mahdian, M.; Alnajjar, H.; Tadinada, A. Deep learning classifier with optical coherence tomography images for early dental caries detection. In Proceedings of the Lasers in Dentistry XXIV. International Society for Optics and Photonics, 28 January, San Fransicso, CA, USA, 28 January 2018; Volume 10473, p. 1047304. [Google Scholar]
  36. Birdal, R.G.; Gumus, E.; Sertbas, A.; Birdal, I.S. Automated lesion detection in panoramic dental radiographs. Oral Radiol. 2016, 32, 111–118. [Google Scholar] [CrossRef]
  37. Ghaedi, L.; Gottlieb, R.; Sarrett, D.C.; Ismail, A.; Belle, A.; Najarian, K.; Hargraves, R.H. An automated dental caries detection and scoring system for optical images of tooth occlusal surface. In Proceedings of the 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Chicago, IL, USA, 26–30 August 2014; IEEE: Manhattan, NY, USA, 2014; pp. 1925–1928. [Google Scholar]
  38. Gawad, A.L.A.; El-Sharkawy, Y.H.; El-Sherif, A.F. Classification of human teeth caries using custom non-invasive optical imaging system. Lasers Dent. Sci. 2017, 1, 73–81. [Google Scholar] [CrossRef]
  39. Schwendicke, F.; Rossi, J.; Göstemeyer, G.; Elhennawy, K.; Cantu, A.; Gaudin, R.; Chaurasia, A.; Gehrung, S.; Krois, J. Cost-effectiveness of artificial intelligence for proximal caries detection. J. Dent. Res. 2021, 100, 369–376. [Google Scholar] [CrossRef]
  40. Kim, J.; Shin, T.; Kong, H.; Hwang, J.Y.; Hyun, H. High-frequency ultrasound imaging for examination of early dental caries. J. Dent. Res. 2019, 98, 363–367. [Google Scholar] [CrossRef]
  41. Srivastava, M.M.; Kumar, P.; Pradhan, L.; Varadarajan, S. Detection of tooth caries in bitewing radiographs using deep learning. arXiv 2017, arXiv:1711.07312. [Google Scholar]
  42. Bokefode, J.D.; Ubale, S.A.; Gaikwad, R.M. Retrieving Real Time Data Through IOT Devices and Storing Securely on Cloud Using Li-Fi. In Proceedings of the 2018 3rd International Conference for Convergence in Technology (I2CT), Pune, India, 6–8 April 2018; IEEE: Manhattan, NY, USA, 2018; pp. 1–5. [Google Scholar]
  43. Mamoun, J.S. Assembly and clinical use of the XCP dental X-ray film holder and orientation devices in dentistry. Dent. Assist. 2011, 80, 8. [Google Scholar] [PubMed]
  44. Erin, N. 2008. Available online: https://www.dentalcompare.com/Featured-Articles/1994-Digital-X-ray-Sensors-Placement-for-Optimum-Images/#:~:text=The%20principle%20behind%20this%20technique,true%20parallelism%20and%20greater%20patient (accessed on 21 March 2023).
  45. SparkFun. 2021. Available online: https://www.sparkfun.com/products/17146 (accessed on 23 March 2023).
  46. Mauriello, S.M.; Broome, A.M.; Platin, E.; Mol, A.; Inscoe, C.; Lu, J.; Zhou, O.; Moss, K. The role of stationary intraoral tomosynthesis in reducing proximal overlap in bitewing radiography. Dentomaxillofacial Radiol. 2020, 49, 20190504. [Google Scholar] [CrossRef]
  47. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  48. Li, S.; Fevens, T.; Krzyżak, A.; Li, S. An automatic variational level set segmentation framework for computer aided dental X-rays analysis in clinical environments. Comput. Med. Imaging Graph. 2006, 30, 65–74. [Google Scholar] [CrossRef] [PubMed]
  49. Mahmoud, Y.E.; Labib, S.S.; Mokhtar, H.M. Teeth periapical lesion prediction using machine learning techniques. In Proceedings of the 2016 SAI Computing Conference (SAI), London, UK, 13–15 July 2016; IEEE: Manhattan, NY, USA, 2016; pp. 129–134. [Google Scholar]
  50. Balasundaram, S.; Prasad, S.C. Robust twin support vector regression based on Huber loss function. Neural Comput. Appl. 2020, 32, 11285–11309. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.