Next Article in Journal
Ensuring the QoE-Related Fairness to Reduce the User Abandonment Ratio
Next Article in Special Issue
The Accuracy and Reliability of Tooth Shade Selection Using Different Instrumental Techniques: An In Vitro Study
Previous Article in Journal
Photometric Calibration for Stereo Camera with Gamma-like Response Function in Direct Visual Odometry
Previous Article in Special Issue
The Use and Performance of Artificial Intelligence in Prosthodontics: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Dental Apical Lesions Using CNNs on Periapical Radiograph

1
Department of General Dentistry, Chang Gung Memorial Hospital, Taoyuan City 33305, Taiwan
2
Department of Computer Science and Information Engineering, National Ilan University, Yilan City 260, Taiwan
3
Department of Electronic Engineering, Chung Yuan Christian University, Taoyuan City 32023, Taiwan
4
Department of Electrical Engineering, Ming Chi University of Technology, New Taipei City 243303, Taiwan
5
Center for Internet of Things and Intelligent Cloud, Chung Yuan Christian University, Taoyuan City 32023, Taiwan
6
Department of Information Systems and Computer Science, Ateneo de Manila University, Quezon City 1108, Philippines
7
National Synchrotron Radiation Research Center, Hsinchu City 30076, Taiwan
*
Authors to whom correspondence should be addressed.
Equal contribution to the first author.
Sensors 2021, 21(21), 7049; https://doi.org/10.3390/s21217049
Submission received: 17 September 2021 / Revised: 19 October 2021 / Accepted: 22 October 2021 / Published: 24 October 2021
(This article belongs to the Special Issue Artificial Intelligence & Robotics in Dental Medicine)

Abstract

:
Apical lesions, the general term for chronic infectious diseases, are very common dental diseases in modern life, and are caused by various factors. The current prevailing endodontic treatment makes use of X-ray photography taken from patients where the lesion area is marked manually, which is therefore time consuming. Additionally, for some images the significant details might not be recognizable due to the different shooting angles or doses. To make the diagnosis process shorter and efficient, repetitive tasks should be performed automatically to allow the dentists to focus more on the technical and medical diagnosis, such as treatment, tooth cleaning, or medical communication. To realize the automatic diagnosis, this article proposes and establishes a lesion area analysis model based on convolutional neural networks (CNN). For establishing a standardized database for clinical application, the Institutional Review Board (IRB) with application number 202002030B0 has been approved with the database established by dentists who provided the practical clinical data. In this study, the image data is preprocessed by a Gaussian high-pass filter. Then, an iterative thresholding is applied to slice the X-ray image into several individual tooth sample images. The collection of individual tooth images that comprises the image database are used as input into the CNN migration learning model for training. Seventy percent (70%) of the image database is used for training and validating the model while the remaining 30% is used for testing and estimating the accuracy of the model. The practical diagnosis accuracy of the proposed CNN model is 92.5%. The proposed model successfully facilitated the automatic diagnosis of the apical lesion.

1. Introduction

The radiographic finding of a potential endodontic pathosis is an essential part of endodontic assessment in daily dental practice. Apical periodontitis (AP) is an inflammatory response in the apical periodontium that often results from tooth root canal microorganisms, which may cause apical bone resorption, resulting in apical lesion [1]. Thus, diagnosis of periapical pathology has always been a topic of concern in endodontics. Identifying the periapical pathology of nonvital teeth is a key factor of successful treatment planning and prognosis evaluation [2]. Due to the fact that tooth with periapical pathosis is sometimes asymptomatic, it is often only detected via routine radiographic examination [3]. Few diagnostic tests provide as much useful information as dental radiography. In general, apical endodontic pathosis appears radiographically as bone loss in the area of the periapex and presenting radiolucent area at the apex of the root. Among all types of dental radiography, the periapical film which like Figure 1 is most commonly used in apical lesion diagnosis clinically [4].
Although cone-beam computed tomography(CBCT) is more sensitive in detecting apical lesions [5,6], the periapical film is still important, owing to the fact that periapical film requires lower radiation dosage exposure and is more commonly used in the daily dental radiographic examination [7]. Assessment of the location and extent of apical periodontitis (AP) will determine the treatment plan and subsequent treatment modality [8]. Therefore, it is necessary to be very cautious about the accuracy of the differential diagnosis of AP and normal periodontium, because each judgment has a huge impact on the treatment course of the patient [9]. However, the radiographic appearance of endodontic pathosis in a periapical film is sometimes highly subjective. In a study by Goldman et al., there was only 50% agreement among interpreters for the radiographic presence of pathosis. When the cases were reevaluated several months later, the same evaluators agreed with their own original diagnosis less than 85% of the time [10]. This also shows that in image interpretation, dentists make different judgments on the same image from time to time. Additionally, diagnosis of potential endodontic pathosis may be time-consuming for dentists in daily dental practice. With the rapid growth of technology, big data analytics, and machine learning techniques have been widely used in medicine, such as in autism [11], detection of pneumonia [12], risk of falls [13], classification of cancers [14], application of disease detection [15], predicting severe retinopathy of prematurity [16], dynamic modeling of medical imaging [17], and lung and pancreatic tumor characterization [18]. In response to this problem, this article combines big data analysis and machine learning technology to improve. Apical lesion detection by using periapical radiograph based on transfer learning with convolutional neural networks (CNN) can provide objective radiography interpretation and could potentially save more clinical time for dentists allowing them to focus on treatment modality and clinical operation.
Krajnc et al. [19] provided one of breast cancer detection by using machine learning the improvement took advantage of the advanced data pre-processing techniques in radioactive models. The differentiation was added for benign and malignant tumors in patients’ patterns that cannot be assessed sufficiently with conventional breast imaging and who are not candidates for MRI. Thus, the concept of the article is to develop a new technology to improve detection accuracy. The literature in [20] applied image preprocessing methods which included image resizing, center cropping, and intensity normalization solving the issues of dominant prostate pixels by inter-patient, inter-scanner variability, and background pixels. In order, to enrich the network with more data, incrementing variable data, accuracy improvement, patch extraction, and data enhancement are applied before network training.
Most of the studies are based on the analysis of oral X-rays for finding the characteristics of the disease. The methods for achieving the purpose of identifying the medical condition are directed toward the use of artificial intelligence. One is the example of the usage of artificial intelligence [21] that develop a new CNN model to determine the condition of the three dental caries of the panoramic image. The overall accuracy [22] is up to 86% using CNN for detecting teeth and classifying problems. Semantic segmentation was first used to distinguish between targets of interest with an accuracy of up to 89%. Other technologies do not use artificial intelligence [23] in determining the feature of the tooth decay of the bitewing image, but use edge identification and Otsu thresholding for the image preprocessing. The affected areas were highlighted by the connected component and mask to find out where dental caries are [24] and analyze panoramic images on the mandibular bone for osteoporosis detection. The GLCM matrix is used to find the features. Then, the SVM algorithm is used to classify normal and osteoporosis classes. The classification results are validated by using data prepared by the Dental Radiographs Department. In this study, the method for analyzing the image of the periapical and prejudging whether the apical is inflamed and developed has improved to an accuracy of 85.71%. A semi-automatic method is presented in the study in [25]; where the lesions were manually selected using the TopHat technology in calculating and differentiating between the features of the symptoms and the normal conditions. The deep-learning techniques are currently being wildly used. The article in [26] provided a method to enhance feature extraction including shape, color, and texture by applying a convolutional neural networks (CNN) architecture to extract deep-learning functions. Moreover, the function is a pre-trained ImageNet (ILSVRC ImageNet task). The proposed method in the literature is different from other works that choose the CNN as a technique to detect the apical lesion of the periapical images. The CNN automatically extracts the significant features and improved the accuracy by classifying the images for the proposed model. The semi-automatic method utilized logic regression analysis and provided an improved accuracy of 80.7% [25] for detecting the tooth decay and lesion of the periapical image. The active contour method and level set method were used to cut out the targets of interest while FVM was used to extract the features. It is also suggested that the accuracy can be improved by using fuzzy clustering [27] for detecting the granuloma of the periapical image. Gabor wavelengths were used to extract features from a periapical image, and SVM was adopted as a classifier. The reported accuracy reached as high as 91.67%. Based on the above information, this article proposes apical lesion detection by using a periapical radiograph based on transfer learning with convolutional neural networks (CNN), using the literature in [23,27] as important references for comparison. This provides objective radiography interpretation and potentially saves more clinical time for dentists, allowing them to focus on treatment modality and clinical operation.

2. Materials and Methods

The purpose of this study is to identify the apical lesions of the periapical image through convolutional neural networks. The proposed approach presented in this study shows in Figure 2, can be divided into four steps: image processing, image cropping, retouching the setup database, and CNN image identification.

2.1. Image Preprocessing

The purpose of the image preprocessing step is to perform the binary processing [28] on the root tip. The advantage of binary processing is to effectively emphasize details that are difficult to find in the image or to separate the region of interest from non-targets. The pen tip is binary, and its background is changed. It is a very small gray value (i.e., 0). The bones of the teeth and grooves become the maximum ash value (i.e., 255). This step is performed by the subsequent imaging and cutting techniques [29]. According to the above references, if the original dental X-ray image is directly selected, such as using a bitewing image, and directly perform iterative threshold and binarization (two-value), the result will not be as expected. However, in the original clinical images (background, teeth, and grooved bones), the grayscale values of each segment may not be significantly different. It is not necessarily as distinct as described above, each segment is obvious and does not interfere with each other. The root tip is no exception, and there may be uneven grayscale values. In this case, it cannot be considered as the target and the background is indeed separated. With that, there will be incomplete dental images. If this problem occurs, it will seriously affect the subsequent image cutting. Figure 3 shows the results of image preprocessing in this paper.

2.1.1. Gaussian High Pass Filter

Due to the problems mentioned in the other methods proposed in the literature, this article takes advantage of the Gauss Qualcomm filter first, followed by using the iterative algorithm to select the threshold to avoid the problems encountered. The Gauss high pass filter is explored, which enables the sharpening of the image and better extracts the edge information in the image. The formula in Equation (1) is used, which makes use of an image with size u × v :
H ( u , v ) = 1 e D ( u , v ) 2 2 × D 0
In this study, where the root tip by filter sharpening the image is performed, its filtered image shows the highlighted impurities and edges. The filtered image [29] is then subtracted to the original image. By doing so, the original image with impurities is reduced and the gray-scale value difference is more evident.

2.1.2. Iterative Thresholding

There are many ways to choose the threshold value. The threshold value can either be a fixed threshold value or a variable one where an adaptive threshold value is implemented through many different ways of selection. The adaptive threshold is the main technology because it is based on a given image to calculate the best threshold so as to obtain a better two-value results. On the other hand, as using a fixed threshold is sensitive to noise, the differences among the gray-scale values become less obvious. The image segmentation with highly overlapping area among different target gray-scale values make the region of interests less evident. For that reason, the said method is seldom used.
In the case of using an adaptive threshold, the principle of the iterative threshold is to define the initial threshold, T0, as the average grayscale value. The practice is as follows (a)~(c):
(a)
Finding the maximum (Z_max0) and minimum(Z_min0) grayscale value of the image to get the initial threshold value(T0).
(b)
Using T0 as the cutting condition divide the image into two grayscale group and find the average grayscale value of two group (Zout and Zback).
(c)
Calculating the average of the Zout and Zback to find the new the threshold value (T1).
Lastly is to compare the gap conditions for the computed T1 and T0, like Equations (2) and (3). If the conditions are all met, T1 is the best threshold value as determined by the iterative algorithm [30].
T 0 = Z m a x 0 + Z _ m i n 0 2
T 1 = Z _ o u t + Z _ b a c k 2

2.2. Image Cropping and Retouching

The introduction of image-based diagnosis, machine learning methods in disease prognosis, and risk assessment in the biomedical field paved the way to the CNN model for symptom classification established in this paper. This is performed by inputting a photo of a tooth to the network, and then the model tries to assess whether the tooth being examined has any symptoms or none. To improve the accuracy of judging the symptoms for each tooth in the periapical film (PA), it is separated into individual tooth images and the photos are preprocessed for noise removal. To avoid miscalculations of the CNN model, it is important to eliminate the noise content of the teeth in the non-primary judgment and the unnecessary portion of the trimmed image.
The previous papers related to image segmentation focused on atlas-based segmentation, model-based segmentation, and deep learning-based segmentation. Moreover, automatic image [27] focuses on prostate segmentation based on deformable models, region-based, and patch-based CNNs. The center cropping and intensity normalization [19] were used to separate each tooth by finding a cutting line with its adjacent teeth in a two-valued PA. The identified cutting line is separated from each tooth in the PA into individual tooth photos. However, in this technology [22] cut photos with incomplete pieces results to noisy images. This problem can be resolved by cropping the photos, segmenting the target image, and trimming the finished cropped output photo. The goal of the trimming technology is to cover the non-target image. The image is compared with the original image, and the non-target image is completely covered, but leaving the target image only. Finally, the non-target noise image masking technology is realized, which simplifies the image.

2.2.1. Introduction to Vertical Cutting

Vertical cutting requires the use of a vertical projection in the photo to find a cutting line that separates adjacent teeth, and the step is to add up the pixel values of each column in a binary image to find the column with the smallest gross value. The principle behind this is that a cutting line that separates adjacent teeth will inevitably be on the tooth seam, and the tooth seam will be black after the two-value, with a pixel value of 0. Therefore, the location of the cutting line will be in the column with the smallest sum of pixel values. However, the teeth in the original image are not necessarily at the same angle. Some images perform the cutting line searching, which will cause the error of cropping to the target image. Thus, before using the vertical projection to search for the cutting line, the photo rotation must be corrected first. With the cutting line, the input photo is cut into slices of specific sizes. The two output photos are called the left photo after cutting and the right photo after cutting, which are respectively the area on the left of the cutting line and the area on the right of the cutting line. In order for the two output photos to present complete targets and reduce noise, it is necessary to determine which areas of the input images the output image is in. Here, the target of the left photo after cutting is the area to the left of the cutting line in the input photo, and the target of the right photo after cutting is the area to the right of the cutting line in the input photo. Therefore, the area of the left photo after cutting must be from the left edge of the input photo to the right end of the cutting line, and the area of the right photo after cutting must be from the left end of the cutting line to the right edge of the input photo.
Take Figure 4b as an example: the orange line in the figure is the cutting line found by vertical projection [31]. The blue box is the area of the left photo after cutting in the input photo, and the green box is the area of the right photo after cutting in the input photo. In order to separate each tooth in the input photo into photos of a single tooth after vertical cutting, it is necessary to judge whether the width of the output photo is less than 1/4 of the width of the input photo. If it does, the vertical cutting is put on halt. Otherwise, the vertical cutting step is repeated.

2.2.2. Description of Image Retouching Execution Method

In the vertical cutting, there will inevitably be noise images, that is, non-target images. In order to obtain more effective image data for improving accuracy, the output photos after image cropping must proceed to image retouching. The method is to set the pixel value of the non-target object to 0, and the area of the non-target object is outside of the two cutting lines, as shown in Figure 5. This part is done automatically.

2.3. Setup Database

Clinical images were annotated by three professional dentists, having at least 4 years of clinical experience. The dentists guided the researchers, providing knowledge of the periapical radiographic finding and using actual cases to teach the researchers (describing the characteristics of apical lesions). They also provided the researchers clinical data to calibrate the CNN model (eliminating other nontarget symptoms). For the cut image, the first classification, a total of 191 root tip pieces were cut, afterward obtained 476 separate tooth images. According to the clinical database provided by the dentist, this paper distinguishes the images into normal and lesion. Due to the limited information provided, it is found that the samples in the database are unbalanced with a wide disparity between the number of normal and the number of the lesion, as shown in Table 1. If only a small number of images with apical lesions are used in CNN training, the CNN model will not be able to exert its advantages, so the learning effect is not good enough and the result cannot be judged correctly. To solve this problem, this article adds image augmentation technology during CNN training. According to the method of data enhancement [32], the following transformations are introduced: flip, zoom, rotation, translation, contrast, brightness, vertical flip, and horizontal flip, as shown in Figure 6. This technique increases the number of apical lesions to 131. At the same time, the input of normal images will be reduced during training. This converges the ratio between the two sets of samples, thereby reducing imbalance. The data enhancement step is only used to train the CNN model. Therefore, this did not cause confusion when verifying the CNN model.
The image after the normalizations step is shown in Figure 6a. Table 1 shows a six-time difference between the number of lesions and normal. With this amount of data put directly into CNN training, CNN will not be able to fully use its advantage and will not be effective in learning. Therefore, the usage of vertical flip and horizontal flip is increased, the number of lesion images is increased and made equal to the number of normal images, in this case 230 for both lesion images and normal images, giving a total of 460 images were built into the database. Allowing the proportions of the samples to be closer to equilibrium.
If the image cutting size is too large, the training process will take a longer time, causing network problems and resulting in low accuracy. So in this 460 image size standardization, the option is to use a resolution of 200 × 100. The time and network problems are resolved but resolution is traded off.

2.4. CNN Image Identification

Deep learning is a type of machine learning based on artificial neural networks, whose purpose is to train computers to perform human-like tasks. To simulate the way the human brain works, the system is expected to achieve the same learning ability as humans, including identification of objects, speech recognition, make decisions, or make predictions. In short, a large amount of regular information is provided to the computer, automatically finding the best function after the training process.
There are many deep learning networks in the scientific community such as convolutional neural networks (CNN), recurrent neural networks (RNN), and deep neural networks (DNN), with each of this network used in different specific applications. Considering this study, with the aim to ascertain the symptoms in the periapical image, CNN is the best method to apply. CNN is the most powerful method for image recognition, in dealing with convolutional layers to automatically fetch features and carry out feature integration and analysis.

2.4.1. Adjust Model

Before CNN training, the model and database are needed. The database will be used in the step of image cropping and retouching, then the model is developed. Deep learning and neural network-like functions are used to describe data. With the objective function parameters well-defined, the input data is transformed into prediction results. In building-up the network architecture, a select group of possible features for the subsequent deep learning training process is first established. Proper network architecture is defined to generate an effective deep learning model through the training process. A pre-trained model has learned how to recognize the basic features of an image. For example, color, edge, curve, etc. On this basis, training time is reduced and problems that may be encountered while creating a new model from the beginning are avoided thereby improving the training efficiency. According to Yang et al. [31], its technical description of oral cancer in deep learning lies in the selection of adaptive modules as the goal of deep learning. This approach will be adapted in this study for the unique characteristics of the disease in the image to build automated module training. Overcoming the technical difficulties such as identification and marking results to a reduction of cost and improvement in the accuracy of deep learning. The concept of adaptive modules is used as a basis for adjustment and allow modules to have convenience without data restrictions. The adjustment module is no longer a challenge and allows for more flexible use of current data. This article places the input layer indented 200 × 100 × 3 and finds the association positively.
Within CNN, each layer is closely related. In the case of the model stride and convolutional kernel, the size does not adjust. The full-connection layer will need to have adjustments. Making changes in the input of the full connection layer will results to changes on the number of neurons for the entire model. Thus, it is necessary to carefully choose the size to be changed. Too many neurons do not only increase the model complexity and easy for it to overfit, it also makes the calculation time increase and less efficient. The class output layer was changed to 2 in order to see if there is a lesion or not. Table 2 lists the input and output between the layers after the modification.

2.4.2. Adjust Hyperparameter

This paper adopts the SGD gradient drop method to find the best result. This is based on the principle of an iterative algorithm to find the minimum of the loss function. The calculated error of the result and updating the model to obtain high accuracy as well as the impact of learning rates is very large, as detailed in Table 3. In the training data, if a large amount of data is applied to the network, it will lead to a longer training time and memory limitations. In this case, the use of a non-convex function coupled with the neural network will provide a locally optimal solution. Therefore, the concept of mini-batch is used that treats the data by parts when training and does the training just once. This allows a speed up in the model convergence and improves accuracy. There are three hyperparameters that require special attention:
(1)
Learning Rate: The size of the learning rate will determine whether the neural network could converge to the global minimum value and to get a higher accuracy. A too high learning rate causes the loss function to not converge or miss the global extreme of the gradient. On the other hand, a too low learning rate will lead to a slow network convergence or make the neural network converge to the local extreme value which is not the best solution.
(2)
Max Epoch: An Epoch means that all training data is fully passed through the neural network once. One Epoch makes the training less accurate. An appropriate increase in the number of Epoch will result in a better accuracy but will in turn increase the training time.
(3)
Mini BatchSize: BatchSize is the number of samples for a training session. A fewer number for the BatchSize causes the loss functions to be difficult to converge and harder to improve on the accuracy. Too large of a BatchSize will increase memory capacity and the performance of the model will also decline.

2.4.3. Training

When training CNN, the overall data is divided into three general sets in order to conform to the actual situation. They are the training set, the validation set, and the test set that is not put in the learning training. Seventy percent (70%) randomly selected images from the database (322 images) are used as the training and validation set. The remaining 30% is used as the test set (138 images). The training set and the validation set are trained into the network at the same time. After waiting for the network to be trained, about 125 remaining images are put into it as a test set to verify the accuracy of the network and generate a confusion matrix and its truth table.

3. Results

The verification set is used and fed into the network to evaluate the accuracy of the model. The resulting prediction of the model versus the actual labels of the images are monitored in order to calculate the accuracy of the CNN network. Using Equation (4), the computed accuracy of the model is 92.75%. Using Equation (4), as derived from the results, the accuracy of detection of apical lesions based on the technique presented on this paper has been success-fully increased to 92.75%. This is in contrast to the study in [23] which reported an accuracy of 80.7%. In [23], TOPHAT technology was used to crop the periapical, and the model was determined by logistic regression analysis. Although this study showed a significant improvement. However, the possibility of clinical application is unlikely, because of its technical accuracy rate. To further improve on the performance of the model in terms of accuracy, recommendations for future improvements are presented in the next chapter in order to eventually meet the standards for actual use in clinical medicine.
A c c u r a c y = C o r r e c t   p r e d i c t   i m a g e s t o t a l   i m a g e s × 100 %
Table 4 shows that the larger the indicators the better is the accuracy. Table 4 also shows that the indicators in the proposed model are larger than that of the previous work in [23]. Consequently, the CNN image identification ability should be excellent especially in clinical medicine thus the need for high-precision judgment to provide better medical quality. The actual application of this technology uses the clinical images of Figure 7 and Figure 8 as the target image for judging the symptoms. After implementing this technology, the results are shown in Table 5 and Table 6.
In this study, there are two main points to consider for accuracy improvement. For the first point, if a larger sized training input is fed into the network and the training result is not ideal, it implies that the loss function did not converge and that it took a long time for the training resulting in a low accuracy. With that, this article aimed to reduce the image size to solve the problem stated above. This is to remove the uninteresting and insignificant objects in the cut image to allow the CNN image recognition to focus more on the target of this article. Figure 9 shows that the proposed model and its accuracy rate contributed significantly, and its accuracy rate has grown significantly from 54.31% to 92.75% as listed in Table 7.
It is shown from Figure 10 that the smaller the loss function, the higher is the accuracy of the reflection.
Table 8 and Table 9 present the resulting confusion matrix and accuracy for classifying apical lesion, respectively. From Table 9, the probability of misjudgment of being sick or not being judged to be sick has been controlled to 7%. The possible cause of this is an extremely mild disease and possible lesions of the apical tooth film image. The process of CNN image recognition may cause training confusion. It is expected that in the future, these subtle changes are learned and trained in a semi-supervised manner to reduce the probability of error judgment. Through the current image masking technology, the word order is improved to control the error rate. In Table 9, within 7%, in the future, it is also planned to optimize the characteristics of the symptoms by means of image processing. With that, the CNN can perform more accurately which determines the presence or absence of the disease with a view to higher accuracy.
To sum up, the accuracy of correct detection of apical lesions in this study has improved to 92.5%percent. For comparison, the paper by Orhan et al. with a fairly close topic is considered [33]. Orhan et al. (2020) investigated an evaluation of artificial intelligence for detecting periapical pathosis on cone-beam computed tomography scan. Its database consists of 153 images of root-tip-week lesions, and the CNN architecture was conducted using the U-net-like architecture. From the comparison table shown in Table 10, the accuracy of our model is quite close, and the accuracy compared with other papers is shown in Table 11. It can be seen that this model has a certain research value. In the future, we will continue to expand the database. Combining Fast R-CNN technology [34] to make the diagnosis of symptoms technology more accurate. And continue to strengthen the training volume of famous banks to achieve the goal of enhancing accuracy.

4. Discussion

The results of the current model in this work reached an accuracy of 92.75%. The outcomes indicated that there was a significant improvement compared with the literature [34], but there was still an improvement space in clinical medicine and the possibility of its future development. This scheme is designed to improve accuracy continuously and develop the practical application gradually of the medical system. For example, the PICCOLO data set [35] is used to build a powerful database to make the current overall performance more eye-catching, and the image reconstruction technology [36] is provided a good image reconstruction technology that uses image parameters adjustment to obtain a high SNR. The output of model results is optimized and automatically compared with the data provided by dentists. Convolutional neural networks can provide objective radiography interpretation and could save more clinical time for dentists to focus on treatment modality and clinical operation. There is considerable confidence toward future application development, which will be patented for these proposed results. Because this program is valuable and is a long-term continuous study, both the technology research and development results is to be protected and granted patents to enable more people to study and maintain the field of medical diseases.
Regarding the research limitations in this study, as it is impossible to collect all different dental apical lesions image data, the periapical radiograph image data sources collected in this study only comes from patients in Chang Gung Memorial Hospital. In addition, in terms of detection model selection, the CNN models (AlexNet, GoogleNet, Vgg, ResNet) were used in this study in consideration of both machine performance and accuracy. In the future, more deep learning models and more various periapical radiograph image data can be used. Although the CNNs can achieve excellent image recognition and detection results, it requires many data from various sources behind it and must be labeled as learning features.

5. Conclusions

Along with the image cropping process, it was found out in this study that cropping each tooth individually from a given X-ray image and trying to avoid ruining adjacent tooth while cropping them individually relies heavily on the ability to pre-process the images. Due to adopting the method of using the vertical sum of a pixel for cropping images, thereby binarization process, took an important role in the pre-processing step. Thus, taking advantage of the Gaussian high pass filter which has a critical capability to edge crispening led to improving the method of cropping that was implemented in this study.
On the other hand, to further improve on the success rate, two hypotheses are proposed. The first one is more intuitive, when the image data is increased it will be helpful to the improvement of the success rate. Secondly, is to enhance the characteristics of the Periapical lesions. Before using the image for CNN training, the area of the lesion is to be processed. As it is difficult to identify special cases for the CNN, this will allow the CNN to obtain new judgment data. The accuracy of this model represents the objective data for CNN to determine the lesions of Periapical lesions. The results presented in this study show the possibility of automatically identifying and judging the periapical lesions with a success rate of as high as 92.75%. This study paved the way in presenting an alternative way of automating recognition of apical lesions on periapical radiograph that are still done manually until the present.

Author Contributions

Conceptualization, C.-W.L., Y.-C.H. and Y.-C.M.; Data Curation, C.-W.L., Y.-C.H. and Y.-C.M.; Formal Analysis, H.-S.C., Y.-L.L., C.-A.C. and S.-L.C.; Funding Acquisition, S.-L.C.; Methodology, S.-Y.L. (Szu-Yin Lin), H.-S.C. and T.-Y.C.; Resources, S.-Y.L. (Szu-Yin Lin), C.-A.C., S.-L.C., W.-S.L. and W.-Y.C.; Software, S.-Y.L., Y.-A.C., S.-Y.L. (Sheng-Yu Liu) and Y.-L.L.; Supervision, C.-A.C. and S.-L.C.; Validation, Y.-A.C.; Visual-ization, S.-Y.L. (Sheng-Yu Liu), P.A.R.A. and W.-Y.C.; Writing—Original Draft, C.-W.L., S.-Y.L. (Szu-Yin Lin) and T.-Y.C.; Writing—Review and Editing, C.-A.C., P.A.R.A. and W.-S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of Science and Technology (MOST), Taiwan, under grant numbers of MOST-108-2628-E-033-001-MY3, MOST-110-2823-8-033-002, MOST-110-2622-E-131-002, MOST-109-2622-E-131-001-CC3, MOST-108-2622-E-033-012-CC2, 109-2410-H-197-002-MY3, 109-2221-E-131-025, 109-2622-E-131-001-CC3, 110-2622-E-131-002, 108-2112-M-213-014, 109-2112-M-213-013, and the National Chip Implementation Center, Taiwan.

Institutional Review Board Statement

Chang Gung Medical Foundation Institutional Review Board; IRB number: 202002030B0; Date of Approval: 1 December 2020; Protocol Title: A Convolutional Neural Network Approach for Dental Bite-Wing, Panoramic and Periapical Radiographs Classification; Executing Institution: Chang Gung Medical Foundation Taoyuan Chang Gung Memorial Hospital of Taoyuan; Duration of Approval: From 1 December 2020 to 30 November 2021; The IRB reviewed and determined that it is expedited review according to Case research or cases treated or diagnosed by clinical routines. However, this does not include HIV-positive cases.

Informed Consent Statement

The IRB approves the waiver of the participants’consent.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Glossary of Endodontic Terms. American Association of Endodontists. Available online: https://www.aae.org/specialty/clinical-resources/glossary-endodontic-terms/ (accessed on 9 September 2021).
  2. Wallace, J.A.; Nair, M.K.; Colaco, M.F.; Kapa, S.F. A comparative evaluation of the diagnostic efficacy of film and digital sensors for detection of simulated periapical lesions. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. Endod. 2001, 92, 93–97. [Google Scholar] [CrossRef]
  3. Bender, B. Factors influencing the radiographic appearance of bony lesions. J. Endod. 1982, 8, 161–170. [Google Scholar] [CrossRef]
  4. Estrela, C.; Bueno, M.R.; Leles, C.R.; Azevedo, B.; Azevedo, J.R. Accuracy of cone beam computed tomography and panoramic and periapical radiography for detection of apical periodontitis. J. Endod. 2008, 34, 273–279. [Google Scholar] [CrossRef] [PubMed]
  5. Dutra, K.L.; Haas, L.; Porporatti, A.L.; Flores-Mir, C.; Santos, J.N.; Mezzomo, L.A.; Correa, M.; Canto, G.D. Diagnostic Accuracy of Cone-beam Computed Tomography and Conventional Radiography on Apical Periodontitis: A Systematic Review and Meta-analysis. J. Endod. 2016, 42, 356–364. [Google Scholar] [CrossRef]
  6. Saidi, A.; Naaman, A.; Zogheib, C. Accuracy of Cone-beam Computed Tomography and Periapical Radiography in Endodontically Treated Teeth Evaluation: A Five-Year Retrospective Study. J. Int. Oral Health 2015, 7, 15–19. [Google Scholar]
  7. Cone Beam Computed Tomography in Endodontics. American Association of Endodontists (Summer 2011 Endodontics: Colleagues for Excellence). Available online: https://f3f142zs0k2w1kg84k5p9i1o-wpengine.netdna-ssl.com/specialty/wp-content/uploads/sites/2/2017/06/ecfe-summer-11-final.pdf (accessed on 21 October 2021.).
  8. Restrepo-Restrepo, F.A.; Cañas-Jiménez, S.J.; Romero-Albarracín, R.D.; Villa-Machado, P.A.; Pérez-Cano, M.I.; Tobón-Arroyave, S.I. Prognosis of root canal treatment in teeth with preoperative apical periodontitis: A study with cone-beam computed tomography and digital periapical radiography. Int. Endod. J. 2019, 52, 1533–1546. [Google Scholar] [CrossRef]
  9. Goldman, M.; Pearson, A.H.; Darzenta, N. Endodontic success—Who’s reading the radiograph? Oral Surg. Oral Med. Oral Pathol. 1972, 33, 432–437. [Google Scholar] [CrossRef]
  10. Kruse, C.; Spin-Neto, R.; Reibel, J.; Wenzel, A.; Kirkevang, L.-L. Diagnostic validity of periapical radiography and CBCT for assessing periapical lesions that persist after endodontic surgery. Dentomaxillofac. Radiol. 2017, 46, 20170210. [Google Scholar] [CrossRef] [PubMed]
  11. Kido, S.; Hirano, Y.; Hashimoto, N. Detection and classification of lung abnormalities by use of convolutional neural network (CNN) and regions with CNN features (R-CNN). In Proceedings of the 2018 International Workshop on Advanced Image Technology (IWAIT), Chiang Mai, Thailand, 7–9 January 2018; pp. 1–4. [Google Scholar] [CrossRef]
  12. Lie, W.-N.; Le, A.T.; Lin, G.-H. Human fall-down event detection based on 2D skeletons and deep learning approach. In Proceedings of the 2018 International Workshop on Advanced Image Technology (IWAIT), Chiang Mai, Thailand, 7–9 January 2018; pp. 1–4. [Google Scholar] [CrossRef]
  13. Zeinalzadeh, A.; Wenska, T.; Okimoto, G. A neural network model to classify liver cancer patients using data expansion and compression. In Proceedings of the 2017 American Control Conference (ACC), Seattle, WA, USA, 24–26 May 2017; pp. 2135–2139. [Google Scholar] [CrossRef] [Green Version]
  14. Roy, S.; Menapace, W.; Oei, S.; Luijten, B.; Fini, E.; Saltori, C.; Huijben, I.; Chennakeshava, N.; Mento, F.; Sentelli, A.; et al. Deep Learning for Classification and Localization of COVID-19 Markers in Point-of-Care Lung Ultrasound. IEEE Trans. Med Imaging 2020, 39, 2676–2687. [Google Scholar] [CrossRef]
  15. Karatekin, T.; Sancak, S.; Celik, G.; Topcuoglu, S.; Karatekin, G.; Kirci, P.; Okatan, A. Interpretable Machine Learning in Healthcare through Generalized Additive Model with Pairwise Interactions (GA2M): Predicting Severe Retinopathy of Prematurity. In Proceedings of the 2019 International Conference on Deep Learning and Machine Learning in Emerging Applications (Deep-ML), Istanbul, Turkey, 26–28 August 2019; pp. 61–66. [Google Scholar] [CrossRef]
  16. Swarna, S.R.; Boyapati, S.; Dutt, V.; Bajaj, K. Deep Learning in Dynamic Modeling of Medical Imaging: A Review Study. In Proceedings of the 2020 3rd International Conference on Intelligent Sustainable Systems (ICISS), Thoothukudi, India, 3–5 December 2020; pp. 745–749. [Google Scholar] [CrossRef]
  17. Hussein, S.; Kandel, P.; Bolan, C.W.; Wallace, M.B.; Bagci, U. Lung and Pancreatic Tumor Characterization in the Deep Learning Era: Novel Supervised and Unsupervised Learning Approaches. IEEE Trans. Med. Imaging 2019, 38, 1777–1787. [Google Scholar] [CrossRef] [Green Version]
  18. Krajnc, D.; Papp, L.; Nakuz, T.S.; Magometschnigg, H.F.; Grahovac, M.; Spielvogel, C.P.; Ecsedi, B.; Bago-Horvath, Z.; Haug, A.; Karanikas, G.; et al. Breast Tumor Characterization Using [18F] FDG-PET/CT Imaging Combined with Data Preprocessing and Radiomics. Cancers 2021, 13, 1249. [Google Scholar] [CrossRef]
  19. Khan, Z.; Yahya, N.; Alsaih, K.; Ali, S.S.A.; Meriaudeau, F. Evaluation of Deep Neural Networks for Semantic Segmentation of Prostate in T2W MRI. Sensors 2020, 20, 3183. [Google Scholar] [CrossRef]
  20. Bouchahma, M.; Hammouda, S.B.; Kouki, S.; Alshemaili, M.; Samara, K. An Automatic Dental Decay Treatment Prediction using a Deep Convolutional Neural Network on X-ray Images. In Proceedings of the 2019 IEEE/ACS 16th International Conference on Computer Systems and Applications (AICCSA), Abu Dhabi, United Arab Emirates, 3–7 November 2019; pp. 1–4. [Google Scholar] [CrossRef]
  21. Muresan, M.P.; Barbura, A.R.; Nedevschi, S. Teeth Detection and Dental Problem Classification in Panoramic X-ray Images using Deep Learning and Image Processing Techniques. In Proceedings of the 2020 IEEE 16th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 3–5 September 2020; pp. 457–463. [Google Scholar] [CrossRef]
  22. Bhan, A.; Harsh, G.A.; Chauhan, N.; Wang, C.-W. “Feature Line Profile Based Automatic Detection of Dental Caries in Bitewing Radiography. In Proceedings of the 2016 International Conference on Micro-Electronics and Telecommunication Engineering (ICMETE), Uttarpradesh, India, 22–23 September 2016; pp. 635–640. [Google Scholar] [CrossRef]
  23. Lee, T.; Jeong, H.; Kim, D.; Lee, B.-D. Quantitative Analysis of Endodontic Treatment for Periapical Lesions in Intraoral Radiographs. In Proceedings of the 2009 Ninth IEEE International Conference on Bioinformatics and Bioengineering, Taichung, Taiwan, 22–24 June 2009; pp. 352–355. [Google Scholar] [CrossRef]
  24. Patel, S.; Shah, H.; Jain, K. Inquisition on lesion identification based on clustering in periapical dental radiographs. In Proceedings of the 2017 International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET), Chennai, India, 22–24 March 2017; pp. 449–453. [Google Scholar] [CrossRef]
  25. Almaraz-Damian, J.-A.; Ponomaryov, V.; Sadovnychiy, S.; Castillejos-Fernandez, H. Melanoma and Nevus Skin Lesion Classification Using Handcraft and Deep Learning Feature Fusion via Mutual Information Measures. Entropy 2020, 22, 484. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Zuandi, M.F.; Hidayat, B.; Sitam, S. Granuloma image detection through periapical radiograph by using Gabor wavelet method and support vector machine classification. In Proceedings of the 2018 International Conference on Information and Communications Technology (ICOIACT), Yogyakarta, Indonesia, 6–7 March 2018; pp. 617–622. [Google Scholar] [CrossRef]
  27. Su, B.; Lu, S.; Tan, C.L. Combination of Document Image Binarization Techniques. In Proceedings of the 2011 International Conference on Document Analysis and Recognition, Beijing, China, 18–21 September 2011; pp. 22–26. [Google Scholar] [CrossRef]
  28. Dogra, A.; Bhalla, P. Image Sharpening By Gaussian and Butterworth High Pass Filter. Biomed. Pharmacol. J. 2015, 7, 707–713. [Google Scholar]
  29. Ridler, T.W.; Calvard, S. Picture Thresholding Using an Iterative Selection Method. IEEE Trans. Syst. Man Cybern. 1978, 8, 630–632. [Google Scholar] [CrossRef]
  30. Nomir, O.; Abdel-Mottaleb, M. A system for human identification from X-ray dental radiographs. Pattern Recognit. 2005, 38, 1295–1305. [Google Scholar] [CrossRef]
  31. Yang, J.; Xie, Y.; Liu, L.; Xia, B.; Cao, Z.; Guo, C. Automated Dental Image Analysis by Deep Learning on Small Dataset. In Proceedings of the 2018 IEEE 42nd Annual Computer Software and Applications Conference (COMPSAC), Tokyo, Japan, 23–27 July 2018; Volume 1, pp. 492–497. [Google Scholar] [CrossRef]
  32. Mao, Y.C.; Chen, T.Y.; Chou, H.S.; Lin, S.Y.; Liu, S.Y.; Chen, Y.A.; Liu, Y.L.; Chen, C.A.; Huang, Y.C.; Chen, S.L.; et al. Caries and Restoration Detection Using Bitewing Film Based on Transfer Learning with CNNs. Sensors 2021, 21, 4613. [Google Scholar] [CrossRef]
  33. Welikala, R.A.; Remagnino, P.; Lim, J.H.; Chan, C.S.; Rajendran, S.; Kallarakkal, T.G.; Zain, R.B.; Jayasinghe, R.D.; Rimal, J.; Kerr, A.R.; et al. Automated Detection and Classification of Oral Lesions Using Deep Learning for Early Detection of Oral Cancer. IEEE Access 2020, 8, 132677–132693. [Google Scholar] [CrossRef]
  34. Orhan, K.; Bayrakdar, I.S.; Ezhov, M.; Kravtsov, A.; Özyürek, T. Evaluation of artificial intelligence for detecting periapical pathosis on cone-beam computed tomography scans. Int. Endod. J. 2020, 53, 680–689. [Google Scholar] [CrossRef] [PubMed]
  35. Sánchez-Peralta, L.F.; Pagador, J.B.; Picón, A.; Calderón, Á.J.; Polo, F.; Andraka, N.; Bilbao, R.; Glover, B.; Saratxaga, C.L.; Sánchez-Margallo, F.M. PICCOLO White-Light and Narrow-Band Imaging Colonoscopic Dataset: A Performance Comparative of Models and Datasets. Appl. Sci. 2020, 10, 8501. [Google Scholar] [CrossRef]
  36. Dachena, C.; Casu, S.; Fanti, A.; Lodi, M.B.; Mazzarella, G. Combined Use of MRI, fMRIand Cognitive Data for Alzheimer’s Disease: Preliminary Results. Appl. Sci. 2019, 9, 3156. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Periapical film database.
Figure 1. Periapical film database.
Sensors 21 07049 g001
Figure 2. Proposal flowchart.
Figure 2. Proposal flowchart.
Sensors 21 07049 g002
Figure 3. The preprocessed image results of this article. (a) Periapical film; (b) filtered result of the Gaussian high pass filter; (c) difference of (a) and (b); (d) result after the binarization.
Figure 3. The preprocessed image results of this article. (a) Periapical film; (b) filtered result of the Gaussian high pass filter; (c) difference of (a) and (b); (d) result after the binarization.
Sensors 21 07049 g003
Figure 4. (a) Schematic diagram of the cutting line found by vertical projection; (b) schematic diagram of the cutting range; (c) and (d) are the output photos after vertical cutting the photo in (b).
Figure 4. (a) Schematic diagram of the cutting line found by vertical projection; (b) schematic diagram of the cutting range; (c) and (d) are the output photos after vertical cutting the photo in (b).
Sensors 21 07049 g004
Figure 5. (a) The photo output after Image cropping, (b) the schematic diagram of Image retouching (the orange line is the cutting line, between the orange line is the target, and the rest of the area is the non-target), (c) the image after retouching.
Figure 5. (a) The photo output after Image cropping, (b) the schematic diagram of Image retouching (the orange line is the cutting line, between the orange line is the target, and the rest of the area is the non-target), (c) the image after retouching.
Sensors 21 07049 g005
Figure 6. Image after standardization. (a) Original image, (b) horizontal flip, (c) vertical flip, (d) vertical and horizontal flip.
Figure 6. Image after standardization. (a) Original image, (b) horizontal flip, (c) vertical flip, (d) vertical and horizontal flip.
Sensors 21 07049 g006
Figure 7. Outer example for validation (number 1 to 5 from left to right).
Figure 7. Outer example for validation (number 1 to 5 from left to right).
Sensors 21 07049 g007
Figure 8. Outer example for validation (number 1 to 3 from left to right).
Figure 8. Outer example for validation (number 1 to 3 from left to right).
Sensors 21 07049 g008
Figure 9. The training process of the model with the accuracy in the test set (black line) and the training set (blue line).
Figure 9. The training process of the model with the accuracy in the test set (black line) and the training set (blue line).
Sensors 21 07049 g009
Figure 10. Loss process of the model with the accuracy in the test set (black line) and the training set (blue line).
Figure 10. Loss process of the model with the accuracy in the test set (black line) and the training set (blue line).
Sensors 21 07049 g010
Table 1. Clinical application of the periapical image of the number of apical lesions.
Table 1. Clinical application of the periapical image of the number of apical lesions.
Periapical Image of the Number of Apical Lesions
NormalLesionTotal
Quantity41165476
Table 2. Input and output of each layer for the proposed CNN model.
Table 2. Input and output of each layer for the proposed CNN model.
Input and Output
NameFeature Size
1Input200 × 100 × 3
2Convolution50 × 25 × 48
3Relu50 × 25 × 48
4Normalization50 × 25 × 48
5Maxpooling25 × 13 × 48
6Convolution13 × 7 × 48
7Relu13 × 7 × 48
8Normalization13 × 7 × 48
9Maxpooling13 × 7 × 48
10Fully-Connected1 × 1 × 2184
11Relu1 × 1 × 2184
12Dropout1 × 1 × 2184
13Fully-Connected1 × 1 × 2184
14Relu1 × 1 × 2184
15Dropout1 × 1 × 2184
16Fully-Connected1 × 1 × 2
17Softmax1 × 1 × 2
18Class-output2
Table 3. The hyperparameter values used in the proposed algorithm.
Table 3. The hyperparameter values used in the proposed algorithm.
Hyperparameters
Momentum0.9
Initial Learning Rate6 × 10−5
L2 Regularization1 × 10−4
Gradient Threshold Methodl2norm
Gradient ThresholdInf
Max Epochs100
Mini BatchSize128
Table 4. Accuracy (%) comparison.
Table 4. Accuracy (%) comparison.
Various Indicators Compared with [23]
Use Semi-Automatic Method Utilizing Logic Regression Analysis [23]Test Accuracy
Accuracy80.70%92.75%
Sensitivity80.00%94.87%
Specificity81.39%90.00%
Precision80.00%92.50%
Recall80.00%94.87%
Table 5. Result after judgement for Figure 7 sample image.
Table 5. Result after judgement for Figure 7 sample image.
Figure 7
Number
Clinical DataThis Study
1Normal99.6% to be normal
2Normal99.8% to be normal
3Normal98.6% to be normal
4Normal99.9% to be normal
5Normal99.9% to be normal
Table 6. Result after judgement for Figure 8 sample image.
Table 6. Result after judgement for Figure 8 sample image.
Figure 8
Number
Clinical DataThis Study
1Normal98.5% to be normal
2Apical lesion99.5% to be apical lesion
3Normal97.3% to be normal
Table 7. Verbose.
Table 7. Verbose.
Training Process
EpochIterationTime ElapsedMini-Batch AccuracyTest AccuracyMini-Batch LossValidation Loss
1100:00:0151.56%54.31%1.06150.7952
102000:00:0657.03%71.55%0.82950.5364
204000:00:1065.63%81.03%0.66420.4165
306000:00:1671.09%87.93%0.46520.3354
408000:00:2177.34%91.38%0.43050.2998
5010000:00:2681.25%92.24%0.40620.2727
6012000:00:3179.69%92.24%0.40790.2558
7014000:00:3788.28%92.75%0.31600.2413
8016000:00:4279.69%92.75%0.44390.2370
9018000:00:4785.94%92.24%0.28810.2248
10020000:00:5389.06%92.75%0.26720.2200
Table 8. The confusion matrix of accuracy.
Table 8. The confusion matrix of accuracy.
The Confusion Matrix of Accuracy
True ClassPredict Class
HealthyUnhealthy
Healthy74
53.62%
4
2.89%
Unhealthy6
4.36%
54
39.13%
Table 9. Truth table of accuracy for classifying apical lesion.
Table 9. Truth table of accuracy for classifying apical lesion.
Accuracy of Classifying Apical Lesion
Actual PredictedTRUEFALSE
TRUE92.50%7.50%
FALSE6.90%93.10%
Table 10. Network comparison for apical lesion.
Table 10. Network comparison for apical lesion.
Network Comparison for Apical Lesion
AlexNetGoogleNetVgg19ResNet50
Accuracy92.91%89.36%87.94%88.65%
MaxEpoch100100100100
MiniBatchSize64646464
Iterations per epoch5555
Max iterations500500500500
Validation patience10101010
Learning rate0.000060.000060.000060.00006
Elapsed time1 min 15 s3 min 53 s81 min 8 s167 min 32 s
Table 11. Comparison of accuracy with other papers.
Table 11. Comparison of accuracy with other papers.
Comparison of Accuracy with Other Papers
Our Method with Four Different Models of Transfer LearningMethod in [23]Method in [27]
AlexNetGoogleNetVgg19ResNet50CNNFuzzy clustering and SVM
Accuracy92.91%89.36%87.94%88.65%80.70%91.67%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, C.-W.; Lin, S.-Y.; Chou, H.-S.; Chen, T.-Y.; Chen, Y.-A.; Liu, S.-Y.; Liu, Y.-L.; Chen, C.-A.; Huang, Y.-C.; Chen, S.-L.; et al. Detection of Dental Apical Lesions Using CNNs on Periapical Radiograph. Sensors 2021, 21, 7049. https://doi.org/10.3390/s21217049

AMA Style

Li C-W, Lin S-Y, Chou H-S, Chen T-Y, Chen Y-A, Liu S-Y, Liu Y-L, Chen C-A, Huang Y-C, Chen S-L, et al. Detection of Dental Apical Lesions Using CNNs on Periapical Radiograph. Sensors. 2021; 21(21):7049. https://doi.org/10.3390/s21217049

Chicago/Turabian Style

Li, Chun-Wei, Szu-Yin Lin, He-Sheng Chou, Tsung-Yi Chen, Yu-An Chen, Sheng-Yu Liu, Yu-Lin Liu, Chiung-An Chen, Yen-Cheng Huang, Shih-Lun Chen, and et al. 2021. "Detection of Dental Apical Lesions Using CNNs on Periapical Radiograph" Sensors 21, no. 21: 7049. https://doi.org/10.3390/s21217049

APA Style

Li, C. -W., Lin, S. -Y., Chou, H. -S., Chen, T. -Y., Chen, Y. -A., Liu, S. -Y., Liu, Y. -L., Chen, C. -A., Huang, Y. -C., Chen, S. -L., Mao, Y. -C., Abu, P. A. R., Chiang, W. -Y., & Lo, W. -S. (2021). Detection of Dental Apical Lesions Using CNNs on Periapical Radiograph. Sensors, 21(21), 7049. https://doi.org/10.3390/s21217049

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop