Next Article in Journal
The Effect of Vitamin D Level on Parathyroid Hormone and Alkaline Phosphatase
Previous Article in Journal
Over-the-Counter Breast Cancer Classification Using Machine Learning and Patient Registration Records
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Endoscopic Automatic Classification Model for Gastroesophageal Reflux Disease Using Deep Learning Integrated Machine Learning

1
Division of Gastroenterology, Changhua Christian Hospital, Changhua 500, Taiwan
2
Artificial Intelligence Development Center, Changhua Christian Hospital, Changhua 500, Taiwan
3
Department of Post-Baccalaureate Medicine, College of Medicine, National Chung Hsing University, Taichung 400, Taiwan
4
Department of Medical Informatics, Chung Shan Medical University, Taichung 402, Taiwan
5
Institute of Medicine, Chung Shan Medical University, Taichung 402, Taiwan
6
Division of Gastroenterology and Hepatology, Department of Internal Medicine, Chung Shan Medical University Hospital, Taichung 402, Taiwan
7
Information Technology Office, Chung Shan Medical University Hospital, Taichung 402, Taiwan
*
Authors to whom correspondence should be addressed.
Diagnostics 2022, 12(11), 2827; https://doi.org/10.3390/diagnostics12112827
Submission received: 30 September 2022 / Revised: 5 November 2022 / Accepted: 15 November 2022 / Published: 17 November 2022
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Gastroesophageal reflux disease (GERD) is a common digestive tract disease, and most physicians use the Los Angeles classification and diagnose the severity of the disease to provide appropriate treatment. With the advancement of artificial intelligence, deep learning models have been used successfully to help physicians with clinical diagnosis. This study combines deep learning and machine learning techniques and proposes a two-stage process for endoscopic classification in GERD, including transfer learning techniques applied to the target dataset to extract more precise image features and machine learning algorithms to build the best classification model. The experimental results demonstrate that the performance of the GerdNet-RF model proposed in this work is better than that of previous studies. Test accuracy can be improved from 78.8% ± 8.5% to 92.5% ± 2.1%. By enhancing the automated diagnostic capabilities of AI models, patient health care will be more assured.

1. Introduction

Gastroesophageal reflux is a common physiological phenomenon, and most people may have experienced different severity of reflux [1]. However, if the frequency of gastroesophageal reflux is too high and continues, frequent symptoms or related complications, such as changes in the esophageal mucosa, are called gastroesophageal reflux disease (GERD) [2,3,4,5]. GERD is a relatively common digestive tract disease; it can cause discomfort for the patient and damage the esophageal mucosal tissue. Although GERD is further divided into erosive esophagitis and non-erosive esophagitis, most clinicians use the Los Angeles (LA) classification of esophagitis and diagnose the severity of the disease to provide appropriate treatment.
With the advancement of artificial intelligence (AI), the computer-aided diagnosis technique has become increasingly mature [6]. There is much literature in medicine that uses deep learning models to help physicians diagnose [7,8,9,10,11,12,13,14,15,16,17]. In the classification of GERD, in 2010, Pace et al. proposed a QUID questionnaire (Questionario Italiano Diagnostico) [18] and used an artificial neural network (ANN) to predict. The results of the study showed that the combination of ANN and QUID questionnaires could help distinguish whether a patient suffers from GERD or not. Huang et al. [19] proposed a hierarchical heterogeneous descriptor fusion support vector machine (HHDF-SVM) method for the diagnosis of GERD from conventional endoscopic images. In 2021, Wang et al. proposed the Gerd-VGGNet architecture [20]. This model is a deep convolutional neural network architecture with high generalization. It uses traditional RGB endoscopic images and narrow band imaging (NBI) images for analysis and uses the LA classification to predict the classification of diseases. The experiment compares four pre-trained models. The results show that selecting VGG16 as the pre-trained model has better model accuracy.
In recent years, automatic image feature extraction using pre-trained models of deep learning methods has become a modern technique [21]. In contrast to the pre-training model based on Imagenet [22], studies by Fan et al. [23] and Chang et al. [24] both confirmed that more precise image features can be extracted after fine-tuning the pre-trained network parameters on the target dataset. Yadav et al. [25] compared performance differences between machine learning and deep learning methods for facial acne binary classification using only 120 images and found that a CNN model with the LeakyReLU activation function is better than the SVM model in accuracy. It is important to note that they did not extract image features using a pre-trained model before performing classification using the SVM model.
The purpose of this study is to facilitate the endoscopist in distinguishing the LA classification in the endoscopic mucosa characters at the esophageocardiac junction during the esophagogastroduodenoscopy exam. This relevant AI assistant mode can improve the diagnosis of endoscopic classification of GERD in inexperienced endoscopists in our previous study [20]. This study attempts to improve the accuracy of the prediction system through different deep learning and machine learning techniques to assist in endoscopy of gastroesophageal reflux disease in clinical practice.
The research contributions of this paper are as follows. In this work, a two-stage process for the classification of three-class endoscopic in GERD is proposed that combines deep learning and machine learning techniques. This process includes applying transfer learning to the target data set to extract more precise image features and using machine learning algorithms to create the best classification model. As an image feature extractor, we suggest using the Gerd-VGGNet model. The extracted feature set is then classified using a random forest (RF) classifier with parameter optimization. The experimental results confirm that the prediction performance of this study is superior to that of previous studies.

2. Methods

2.1. GERD Dataset

This study aims to improve the predictive performance of Gerd-VGGNet [20] for GERD, so the same dataset is used for subsequent analysis and comparison. In this study, the NBI endoscopic image data set collected by Wang et al. [20] is used to perform a three-class classification of GERD due to previous references that showed better reproducibility in the classification of esophagitis [26,27] and better differentiation of erosive esophagitis from non-erosive reflux disease [28,29] comparing NBI with conventional images.
Images were obtained from the Chung Shan Medical University Hospital endoscopy system and image data from 496 patients were collected [20]. After processing, a total of 671 images were obtained as the development set, and the other 32 images were used as the test set. In the development set, 244 NBI endoscopic images of grades A–B, 229 images of grades C–D, and 198 normal endoscopic images were used. Furthermore, the distribution of the test set was 12 images from grades A–B, 10 images from grades C–D, and 10 normal endoscopic images.

2.2. Feature Extraction

Compared to traditional machine learning methods that require manual feature extraction, deep learning pre-trained models can automatically extract features from many images. To compare the effects of different deep learning pre-trained models on the classification of GERD, this study compared five image feature extraction techniques described below, where the input size of the image is 96 × 96 × 3, same as Wang et al. [20].
(1)
EfficientNetB7: The deep network architecture proposed by Tan et al. in 2019 [30]. Seven layers were used as the backbone, including the input layer, rescaling, normalization, zero padding, convolution layer, batch normalization, and activation, and then the seven blocks were connected. Each block was connected to a different sub-block. There were 813 layers in total. Our study uses this deep learning model to automatically extract 2560 image features from a global maximum pooling (GMP) layer.
(2)
ResNet50: Residual learning framework proposed by He et al. in 2016 [31] to simplify the training of deeper networks than previously used. On the ImageNet dataset, residual networks with a depth of up to 152 layers were evaluated, which was eight times deeper than the VGG network, but still had lower complexity. This study uses this deep learning model to automatically extract 2048 image features from a GMP layer.
(3)
InceptionResNetV2: Proposed by Szegedy et al. in 2017 [32], InceptionResNetV2 combined the Inception Module and the residual network. Better model accuracy was obtained by deepening the network and using asymmetric convolutional layers and 1 × 1 convolutional layers to implement the model. This study uses this deep learning model to automatically extract 1536 image features from a GMP layer.
(4)
VGG16: VGG16 is a convolutional neural network model proposed by Simonyan et al. at ILSVRC-2014 [33], which includes 13 convolutional layers, five pooling layers, three fully connected (FC) layers, and finally the softmax layer. This model was used in the classification of the ImageNet dataset. There are a total of 1000 classes, as shown in Figure 1. This study uses this deep learning model to automatically extract 512 image features from a GMP layer automatically.
(5)
Gerd-VGGNet: Taking VGG16 as the feature extractor, the classification architecture was designed and fine-tuned for the classification of GERD [20]. In the classification head composed with a GMP layer, a fully connected layer of 256, 128, 64, and 32 nodes and 4 batch normalization layers were added in sequence. Finally, a softmax layer was introduced. The detailed architecture is shown in Figure 1. This study uses this deep learning model to automatically extract 512, 256, 128, 64, and 32 features from the GMP layer and 4 different FC layers, respectively, as shown in Figure 1.
Figure 1. Gerd-VGGNet architecture.
Figure 1. Gerd-VGGNet architecture.
Diagnostics 12 02827 g001

2.3. Random Forest Classifier

The random forest algorithm is a classifier that contains multiple decision trees, proposed by Breiman [34]. An ensemble learning technique uses the bagging method and random feature sampling. The random forest algorithm uses replacement sampling to generate multiple training sets, randomly selects independent variables, and generates multiple decision trees. Finally, the classification results of multiple trees are combined using majority voting.
Compared with the decision tree, the random forest algorithm is based on the ensemble learning technique and uses the majority decision to improve model performance. In addition, it is made up of multiple different trees, each of which is independent. Therefore, a random forest model is less likely to overfit and have high generalization, high accuracy for most datasets, and fast training speed.
Scikit-learn (Sklearn) [35] is a very robust and useful machine learning library in Python. After completing image feature extraction, this study compares 11 Sklearn machine learning classifiers and finally selects the random forest algorithm as the best classification model. The optimization of the configuration of a random forest classifier that includes the number of estimators, maximum depth, random state, and maximum features in this study.

2.4. Performance Evaluation

This study is a 3-class classification task rather than a binary classification task, and all experiments in this paper applied the accuracy value (ACC) and the Kappa value as model performance evaluation indicators to compare the differences in multiclass classification performance. The higher the ACC and Kappa values, the better the model’s classification performance. The detailed calculation formula is as follows:
ACC: For a given test set, the proportion of the total number of samples that the classifier correctly classified.
ACC = T P + T N T P + T N + F P + F N
Kappa: This indicator is used to analyze the consistency of the classification model for each prediction of categories in the multiclass classification task. P 0 is the observed probability or the proportion of agreement and P c is the expected probability.
Kappa = P 0 P c 1 P c

2.5. Cross Validation

In this study, a 10-fold cross-validation is used for training and validation of the classification model. The development set was randomly and equally divided into ten subsets. A subset was used as the validation set, and the remaining nine subsets were used as the training set. In each training process, 608 images were used as the training set and the remaining 63 images were used as the validation set. This was repeated ten times until each subset was used as the validation set, and then the average of the classification results of 10 experiments was calculated as the training and validation results. Lastly, the ten trained classification models were used to evaluate the performance of the test set.

2.6. Proposed Framework

This study uses NBI images of gastroesophageal endoscopy to develop an improved classification model of GERD. The overall study flowchart is shown in Figure 2 and the detailed architecture proposed in this study is demonstrated in Figure 3. First, the development and test sets apply different pre-trained models for automatic image feature extraction, including EfficientNetB7, ResNet50, InceptionResNetV2, VGG16, and Gerd-VGGNet. For the Gerd-VGGNet model, this study uses the output of five different layers (one GMP and four FC) to extract 32, 64, 128, 256, and 512 features, respectively, as shown in Figure 3. Then, the extracted feature sets from the development and test sets are used to compare the classification performance of different machine learning algorithms and perform a 10 fold cross-validation to find the best classification model for GERD.
In order to evaluate the performance of image feature extraction technology, machine learning algorithm, parameter optimization of random forest classifier, and image feature extraction layer, this paper conducts four experiments, as detailed in Section 3.1, Section 3.2, Section 3.3 and Section 3.4.
The architecture proposed in this study can be divided into three blocks, as shown in Figure 3. The first part is the VGG16 pre-trained model, the second part describes Gerd-VGGNet [20], and the third part applies the machine learning classifier. The overall architecture of the proposed model takes VGG16 as the backbone, fine-tunes Gerd-VGGNet for image feature extraction, and finally uses a random forest classifier for GERD classification.

3. Experimental Result

This paper uses the feature data set extracted from the NBI images of GERD. It is divided into a development set and a test set, and the number of data is 671 and 32, respectively [20], in which the development set uses Stratified K-Fold to perform 10-fold cross-validation to train and verify the classification model and build 10 classification models. Finally, the performance evaluation of the target data set is carried out and the relevant experimental results are described below.

3.1. Performance Comparison of Image Feature Extraction Techniques

In this section, we first discuss the difference in the classification performance of GERD NBI images by using the traditional deep learning models with transfer learning and the random forest model with feature extraction by deep learning pre-trained models.
Four popular CNN architectures including EfficientNetB7, ResNet50, InceptionResNetV2 and VGG16 are used as a feature extractor by loading the pre-trained network that does not include the classification layers at the top. For the traditional deep learning models with transfer learning technique, a GMP layer and a FC layer with 3 nodes are added as the classification head. This network of the Gerd-VGGNet model is proposed by Wang et al. [20] as shown in Figure 3. The classification architecture of Gerd-InceptionResNetV2 is the same as that of Gerd-VGGNet, except that the InceptionResNetV2 pre-trained model is used as the backbone. It should be noticed that the number of image features is 2560, 2048, 1536, 512, and 512 using the feature extractors of EfficientNetB7, ResNet50, InceptionResNetV2, VGG16, and Gerd-VGGNet, respectively.
Using five pre-trained models for feature extraction on GERD NBI images, the first six columns in Table 1 list the classification results of different deep learning models with transfer learning, and the last five columns were the classification performance of the random forest classifier without parameter optimization on extracted feature sets from different pre-trained models.
The results in Table 1 show that among the six deep learning models with transfer learning, Gerd-VGGNet has the best classification performance. When using the random forest classifier with five deep learning pre-trained models, the classification performance of Gerd-VGGNet + RF is the best. In the study, many configurations using Gerd-VGGNet + RF have been tested. The 512 one works best. The Gerd-VGGNet + RF with the 512 configuration is named GerdNet-RF here. Furthermore, the results also reveal that the performance of using the random forest algorithm combined with GERD image feature set extracted by the pre-trained model is better than the traditional deep learning model with transfer learning. The results in Table 1 show that the best classification performance for image feature extraction is with the Gerd-VGGNet pre-trained model. The ACC and Kappa for the validation set are 0.982 ± 0.017 and 0.973 ± 0.026, respectively, while for the test set, they can increase to 0.900 ± 0.013 and 0.851 ± 0.019. Using Gerd-VGGNet as the image feature extractor for the GERD NBI images has the highest classification accuracy, as shown in Figure 4. It should be noted that the results of the GerdNet-RF model in Table 1 are without parameter optimization.

3.2. Performance Comparison of Machine Learning Algorithms

Based on the 512 features extracted from the Gerd-VGGNet pre-trained model, this section compares the differences in the classification performance of 11 different machine learning algorithms [35] for endoscopic classification in GERD: Histogram-based gradient boosting classification tree (HistGB), extreme gradient boosting classification tree (XGB), gradient boosting classification tree (GB), random forest, K-nearest neighbors (KNN), linear support vector machine (Linear SVM), support vector machine with radial basis function kernel (RBF SVM), decision tree, Gaussian naive Bayes, logistic regression, and multilayer perceptron (MLP). All of these machine learning algorithms are employed in the Sklearn library, and their parameters were set as default values of the Sklearn library. The results of 11 different classifiers for GERD classification are compared in Table 2, demonstrating that the prediction performance of the use of the random forest classifier is significantly better than that of other machine learning algorithms. The ACC and Kappa are 0.982 ± 0.017 and 0.973 ± 0.026 for the validation set, and they can reach 0.900 ± 0.013 and 0.851 ± 0.019 for the test set. Therefore, for the GERD dataset, the random forest classifier (i.e., the GerdNet-RF model without parameter optimization) has the best classification performance, as shown in Figure 5.

3.3. Performance Differences in Parameter Optimization of Random Forest Classifier

According to the above experiments, the best model is the random forest classifier. In this section, the grid search technique [35] is used to optimize the parameters of this GerdNet-RF model. Based on four important parameters of the random forest algorithm [34], including the number of estimators, maximum depth, random state, and maximum features, Table 3 lists the performance of 10 parameter optimization experiments for the GerdNet-RF classifier.
The results in Table 3 show that the performance is better when the number of estimators is equal to 200, the ACC can reach 0.982 ± 0.017 and 0.909 ± 0.009 for the validation and test sets. On the contrary, the number of estimators is set to 100, and the ACC of the test set is only 0.900 ± 0.017. When setting the maximum depth to 4, 8, and 16, it can be seen from Table 3 that when this parameter is set to 8 and 16, there is no significant difference. And the classification performance is better when the maximum depth is set to 4. The ACC can reach 0.912 ± 0.012 for the test set. Comparing the effect of max features on classification performance, when set to log2, the ACC of the validation set can be improved to 0.984 ± 0.018. By comparing the effect of the random state from 0 to 99 for dataset splitting and finding that the classification performance is the best when a random state is set to 81, the performances of the validation and test sets are both improved. The ACC can be improved to 0.984 ± 0.014 and 0.925 ± 0.021 for the validation and test sets. Finally, the optimal parameter combination for this GerdNet-RF model is the number of estimators equal to 200, the maximum depth equal to 4, the random state equal to 81, and the largest feature equal to log2, as shown in Figure 6. It should be noted that the default experiment in Table 3 is the GerdNet-RF classifier without parameter optimization, and the Tuning 9 experiment represents the GerdNet-RF model proposed in the study.

3.4. Performance Comparison of Image Feature Extraction Layers

To compare the effect of different feature extraction layers of the Gerd-VGGNet pre-trained model on GERD classification performance, this study uses five different feature extraction layers, as shown in Figure 3. Table 4 lists the classification results when the number of extracted features is 32, 64, 128, 256, and 512. When the number of extracted features is small, the classification performance is relatively low. For example, while the number of extracted features is 32, the ACC of the test set can only reach 0.894 ± 0.021. When using 512 extracted features, the ACC of the test set can be increased to 0.925 ± 0.021, and the Kappa value can be improved to 0.888 ± 0.031, as shown in Figure 7. It should be noted that the experiment of 512 features in Table 4 corresponds to the GerdNet-RF model suggested in the article.

4. Discussion

4.1. Performance Comparison of Related Literature

Table 5 lists the classification performance comparison of AI models for the prediction of GERD with three recent publications. This study uses the same dataset as Wang et al. [20], with a total of 671 images for the 3-class classification task. The other two articles used 671 questionnaire data [18] and 147 image data [19] for binary classification tasks, respectively.
Compared to Pace et al. [18] using ANN to predict whether a patient has GERD or not. The diagnosis and management of GERD-related complication is based mainly on endoscopic findings rather than symptom alone. The present study only needs to collect images without complicated questionnaire data collection and can predict the different endoscopic stages of gastroesophageal reflux disease. Such an AI model could be applied to the endoscopic electronic reporting system to help the endoscopist generate the endoscopic report in real time.
The model suggested In this work can diagnose and grade GERD without any manual selection of the region of interest and achieves superior accuracy compared to the approach of Huang et al. [19]. Our proposed method is easier to apply in future work in the field, especially when manpower for image labeling is usually limited in clinical practice.
This study combines deep learning and machine learning approaches to produce an enhanced classification model in comparison to Wang et al. [20], who used deep learning with data augmentation to build the Gerd-VGGNet model. The classification results show a better performance improvement. The overall ACC has increased from 98.9% ± 0.5% to 99.0% ± 0.1%, and the test ACC has also improved from 78.8% ± 8.5% to 92.5% ± 2.1%, as shown in Figure 8. The results of Figure 8 clearly show that the Gerd-VGGNet [20] model has an overfitting problem. A superior GERD computer-aided diagnosis model for clinical practice applications is the GerdNet-RF model proposed in this work, which achieves very good generalization performance, and all training, validation, and test ACCs are very close.

4.2. Image Automatic Classification and Interpretation

To compare the differences between the Gerd-VGGNet model presented in Wang et al. [20] and the GerdNet-RF model proposed in this study in the predicted cases, Table 6 shows the examples of endoscopic images and LA classification that both models can correctly predict for the three classes. Table 7 lists the three endoscopic images and LA grading that the two models predict incorrectly, where the incorrect grading is marked in red. For example, in the first picture, both models misjudge A–B grades as C–D grades, and the second and third images are images that Gerd-VGGNet predict incorrectly.

5. Conclusions

This study proposes a two-stage process that integrates deep learning and machine learning techniques to perform the three-class endoscopic classification in GERD. The experimental results confirm that the model using Gerd-VGGNet as the image feature extractor and the random forest classifier with parameter optimization has the best performance, and the accuracies on the training set, the validation set, and the test set can reach 0.991, 0.984, and 0.925, respectively. Compared to the Gerd-VGGNet model of Wang et al. [20], the GerdNet-RF model proposed in this work can achieve higher classification performance, with an average accuracy improvement rate of 13.7% in the test set. The accuracy of the proposed model also outperforms previous methods proposed by various researchers in the past.
This artificial intelligence model proposed in this study can be applied to an endoscope electronic reporting system to help endoscopists generate endoscopic reports in real time. Especially when the manpower for image labeling in clinical practice is usually limited, our proposed method makes it easier to achieve a computer-aided application in this field.

Author Contributions

Conceptualization, M.-H.T., H.-H.Y. and C.-C.W.; data curation, C.-C.W. and M.-C.T.; funding acquisition, H.-H.Y. and M.-H.T.; methodology, M.-H.T.; software, H.-Y.T.; visualization, H.-Y.T. and C.-C.W.; writing–original draft, H.-Y.T. and M.-H.T.; writing–review and editing, H.-H.Y., C.-C.W., M.-C.T. and M.-H.T. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by Chung Shan Medical University and Changhua Christian Hospital (CSMU-CCH-111-04).

Institutional Review Board Statement

All images were de-identified prior to their transfer to the study’s investigators, and all methods were performed in accordance with the relevant local regulations and under the surveillance of the Institutional Review Board (IRB) of Chung Shan Medical University Hospital (CSMUH No: CS1-20075) (16 December 2021).

Informed Consent Statement

Patient consent was waived due to only endoscopic images were obtained for this retrospective study without any clinical intervention to patients.

Data Availability Statement

All the data of images and analysis process were kept at the lab of M.-H.T.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

AbbreviationName
GERDgastroesophageal reflux disease
LALos Angeles
AIartificial intelligence
QUIDQUEstionario Italiano Diagnostico
ANNartificial neural network
RFRandom Forest
HHDF-SVMhierarchical heterogeneous descriptor fusion support vector machine
NBINarrow Band Imaging
FCFully Connected
GMPGlobal Maxinum Pooling
SklearnScikit-learn
ACCAccuracy
HistGBHistogram-based Gradient Boosting
XGBeXtreme Gradient Boosting
KNNK-nearest Neighbors
Linear SVMLinear Support Vector Machine
RBF SVMSupport Vector Machine with Radial Basis Function kernel
MLPMulti-layer Perceptron

References

  1. Yadlapati, R.; Gyawali, C.P.; Pandolfino, J.E.; Chang, K.; Kahrilas, P.J.; Katz, P.O.; Katzka, D.; Komaduri, S.; Lipham, J.; Menard-Katcher, P.; et al. AGA Clinical Practice Update on the Personalized Approach to the Evaluation and Management of GERD: Expert Review. Clin. Gastroenterol. Hepatol. 2022, 20, 984–994.e1. [Google Scholar] [CrossRef] [PubMed]
  2. Vakil, N.; van Zanten, S.V.; Kahrilas, P.; Dent, J.; Jones, R.; Global Consensus, G. The Montreal definition and classification of gastroesophageal reflux disease: A global evidence-based consensus. Am. J. Gastroenterol. 2006, 101, 1900–1920; quiz 1943. [Google Scholar] [CrossRef] [PubMed]
  3. Eusebi, L.H.; Ratnakumaran, R.; Yuan, Y.; Solaymani-Dodaran, M.; Bazzoli, F.; Ford, A.C. Global prevalence of, and risk factors for, gastro-oesophageal reflux symptoms: A meta-analysis. Gut 2018, 67, 430–440. [Google Scholar] [CrossRef] [PubMed]
  4. Balla, A.; Leone, G.; Ribichini, E.; Sacchi, M.C.; Genco, A.; Pronio, A.; Paganini, A.M.; Badiali, D. Gastroesophageal Reflux Disease—Health-Related Quality of Life Questionnaire: Prospective development and validation in Italian. Eur. J. Gastroenterol. Hepatol. 2021, 33, 339–345. [Google Scholar] [CrossRef]
  5. Katz, P.O.; Dunbar, K.B.; Schnoll-Sussman, F.H.; Greer, K.B.; Yadlapati, R.; Spechler, S.J. ACG Clinical Guideline for the Diagnosis and Management of Gastroesophageal Reflux Disease. Am. J. Gastroenterol. 2022, 117, 27–56. [Google Scholar] [CrossRef]
  6. Yen, H.H.; Wu, P.Y.; Chen, M.F.; Lin, W.C.; Tsai, C.L.; Lin, K.P. Current Status and Future Perspective of Artificial Intelligence in the Management of Peptic Ulcer Bleeding: A Review of Recent Literature. J. Clin. Med. 2021, 10, 3527. [Google Scholar] [CrossRef]
  7. Hosny, A.; Parmar, C.; Quackenbush, J.; Schwartz, L.H.; Aerts, H. Artificial intelligence in radiology. Nat. Rev. Cancer 2018, 18, 500–510. [Google Scholar] [CrossRef]
  8. Zhou, L.Q.; Wang, J.Y.; Yu, S.Y.; Wu, G.G.; Wei, Q.; Deng, Y.B.; Wu, X.L.; Cui, X.W.; Dietrich, C.F. Artificial intelligence in medical imaging of the liver. World J. Gastroenterol. 2019, 25, 672–682. [Google Scholar] [CrossRef]
  9. Das, N.; Topalovic, M.; Janssens, W. Artificial intelligence in diagnosis of obstructive lung disease: Current status and future potential. Curr. Opin. Pulm. Med. 2018, 24, 117–123. [Google Scholar] [CrossRef]
  10. Wong, T.Y.; Bressler, N.M. Artificial Intelligence With Deep Learning Technology Looks Into Diabetic Retinopathy Screening. JAMA 2016, 316, 2366–2367. [Google Scholar] [CrossRef]
  11. Kanagasingam, Y.; Xiao, D.; Vignarajan, J.; Preetham, A.; Tay-Kearney, M.L.; Mehrotra, A. Evaluation of Artificial Intelligence-Based Grading of Diabetic Retinopathy in Primary Care. JAMA Netw. Open 2018, 1, e182665. [Google Scholar] [CrossRef] [Green Version]
  12. Keel, S.; Lee, P.Y.; Scheetz, J.; Li, Z.; Kotowicz, M.A.; MacIsaac, R.J.; He, M. Feasibility and patient acceptability of a novel artificial intelligence-based screening model for diabetic retinopathy at endocrinology outpatient services: A pilot study. Sci. Rep. 2018, 8, 4330. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Kumagai, Y.; Takubo, K.; Kawada, K.; Aoyama, K.; Endo, Y.; Ozawa, T.; Hirasawa, T.; Yoshio, T.; Ishihara, S.; Fujishiro, M.; et al. Diagnosis using deep-learning artificial intelligence based on the endocytoscopic observation of the esophagus. Esophagus 2019, 16, 180–187. [Google Scholar] [CrossRef] [PubMed]
  14. Mori, Y.; Kudo, S.E.; Mohmed, H.E.N.; Misawa, M.; Ogata, N.; Itoh, H.; Oda, M.; Mori, K. Artificial intelligence and upper gastrointestinal endoscopy: Current status and future perspective. Dig. Endosc. 2019, 31, 378–388. [Google Scholar] [CrossRef] [Green Version]
  15. Chang, Y.Y.; Li, P.C.; Chang, R.F.; Chang, Y.Y.; Huang, S.P.; Chen, Y.Y.; Chang, W.Y.; Yen, H.H. Development and validation of a deep learning-based algorithm for colonoscopy quality assessment. Surg. Endosc. 2022, 36, 6446–6455. [Google Scholar] [CrossRef] [PubMed]
  16. Chang, Y.Y.; Yen, H.H.; Li, P.C.; Chang, R.F.; Yang, C.W.; Chen, Y.Y.; Chang, W.Y. Upper endoscopy photodocumentation quality evaluation with novel deep learning system. Dig. Endosc. 2022, 34, 994–1001. [Google Scholar] [CrossRef]
  17. Yen, H.-H.; Wu, P.-Y.; Su, P.-Y.; Yang, C.-W.; Chen, Y.-Y.; Chen, M.-F.; Lin, W.-C.; Tsai, C.-L.; Lin, K.-P. Performance Comparison of the Deep Learning and the Human Endoscopist for Bleeding Peptic Ulcer Disease. J. Med. Biol. Eng. 2021, 41, 504–513. [Google Scholar] [CrossRef]
  18. Pace, F.; Riegler, G.; de Leone, A.; Pace, M.; Cestari, R.; Dominici, P.; Grossi, E. Is it possible to clinically differentiate erosive from nonerosive reflux disease patients? A study using an artificial neural networks-assisted algorithm. Eur. J. Gastroenterol. Hepatol. 2010, 22, 1163–1168. [Google Scholar] [CrossRef] [PubMed]
  19. Huang, C.R.; Chen, Y.T.; Chen, W.Y.; Cheng, H.C.; Sheu, B.S. Gastroesophageal Reflux Disease Diagnosis Using Hierarchical Heterogeneous Descriptor Fusion Support Vector Machine. IEEE Trans. Biomed. Eng. 2016, 63, 588–599. [Google Scholar] [CrossRef]
  20. Wang, C.-C.; Chiu, Y.-C.; Chen, W.-L.; Yang, T.-W.; Tsai, M.-C.; Tseng, M.-H. A deep learning model for classification of endoscopic gastroesophageal reflux disease. Int. J. Environ. Public Health Res. 2021, 18, 2428. [Google Scholar] [CrossRef]
  21. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  22. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  23. Fan, J.; Lee, J.; Lee, Y. A transfer learning architecture based on a support vector machine for histopathology image classification. Appl. Sci. 2021, 11, 6380. [Google Scholar] [CrossRef]
  24. Chang, C.-C.; Li, Y.-Z.; Wu, H.-C.; Tseng, M.-H. Melanoma Detection Using XGB Classifier Combined with Feature Extraction and K-Means SMOTE Techniques. Diagnostics 2022, 12, 1747. [Google Scholar] [CrossRef] [PubMed]
  25. Yadav, N.; Alfayeed, S.M.; Khamparia, A.; Pandey, B.; Thanh, D.N.; Pande, S. HSV model-based segmentation driven facial acne detection using deep learning. Expert Syst. 2022, 39, e12760. [Google Scholar] [CrossRef]
  26. Sharma, P.; Wani, S.; Bansal, A.; Hall, S.; Puli, S.; Mathur, S.; Rastogi, A. A feasibility trial of narrow band imaging endoscopy in patients with gastroesophageal reflux disease. Gastroenterology 2007, 133, 454–464; quiz 674. [Google Scholar] [CrossRef]
  27. Gawron, A.J.; Hirano, I. Advances in diagnostic testing for gastroesophageal reflux disease. World J. Gastroenterol. 2010, 16, 3750–3756. [Google Scholar] [CrossRef]
  28. Fock, K.M.; Teo, E.K.; Ang, T.L.; Tan, J.Y.; Law, N.M. The utility of narrow band imaging in improving the endoscopic diagnosis of gastroesophageal reflux disease. Clin. Gastroenterol. Hepatol. 2009, 7, 54–59. [Google Scholar] [CrossRef]
  29. Dent, J. Microscopic esophageal mucosal injury in nonerosive reflux disease. Clin. Gastroenterol. Hepatol. 2007, 5, 4–16. [Google Scholar] [CrossRef]
  30. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  32. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  33. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014. [Google Scholar] [CrossRef]
  34. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  35. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
Figure 2. Study flowchart.
Figure 2. Study flowchart.
Diagnostics 12 02827 g002
Figure 3. Proposed model architecture.
Figure 3. Proposed model architecture.
Diagnostics 12 02827 g003
Figure 4. Comparison of Performance of five feature extractors.
Figure 4. Comparison of Performance of five feature extractors.
Diagnostics 12 02827 g004
Figure 5. Comparison of performance of 11 classifiers.
Figure 5. Comparison of performance of 11 classifiers.
Diagnostics 12 02827 g005
Figure 6. Comparison of performance of parameter optimization of GerdNet-RF classifier.
Figure 6. Comparison of performance of parameter optimization of GerdNet-RF classifier.
Diagnostics 12 02827 g006
Figure 7. Comparison of Performance of five feature extraction layers of Gerd-VGGNet.
Figure 7. Comparison of Performance of five feature extraction layers of Gerd-VGGNet.
Diagnostics 12 02827 g007
Figure 8. Comparison of performance of Gerd-VGGNet with proposed GerdNet-RF.
Figure 8. Comparison of performance of Gerd-VGGNet with proposed GerdNet-RF.
Diagnostics 12 02827 g008
Table 1. Performance evaluation of five feature extractors.
Table 1. Performance evaluation of five feature extractors.
ModelFeaturesTrainingValidationTest
AccuracyKappaAccuracyKappaAccuracyKappa
EfficientNetB7-0.348 ± 0.0270.000 ± 0.0000.349 ± 0.0260.000 ± 0.0000.347 ± 0.082−0.016 ± 0.123
ResNet50-0.480 ± 0.0150.208 ± 0.0240.478 ± 0.0700.206 ± 0.1080.369 ± 0.0270.016 ± 0.041
InceptionResNetV2-0.789 ± 0.0160.682 ± 0.0240.687 ± 0.0360.530 ± 0.0540.556 ± 0.0560.329 ± 0.083
VGG16-0.681 ± 0.0130.520 ± 0.0190.653 ± 0.0420.479 ± 0.0640.503 ± 0.0530.254 ± 0.080
Gerd- InceptionResNetV2-0.999 ± 0.0010.999 ± 0.0010.879 ± 0.0220.818 ± 0.0330.684 ± 0.0510.520 ± 0.078
Gerd-VGGNet-0.999 ± 0.0010.999 ± 0.0010.903 ± 0.0220.854 ± 0.0340.737 ± 0.0370.603 ± 0.057
EfficientNetB7-RF25601.000 ± 0.0001.000 ± 0.0000.607 ± 0.0450.407 ± 0.0680.334 ± 0.037−0.026 ± 0.057
ResNet50-RF20481.000 ± 0.0001.000 ± 0.0000.618 ± 0.0760.422 ± 0.1140.503 ± 0.0330.235 ± 0.052
InceptionResNetV2-RF15361.000 ± 0.0001.000 ± 0.0000.602 ± 0.0190.397 ± 0.0280.444 ± 0.0570.164 ± 0.085
VGG16-RF5121.000 ± 0.0001.000 ± 0.0000.720 ± 0.0490.577 ± 0.0740.509 ± 0.0610.255 ± 0.093
GerdNet-RF5121.000 ± 0.0001.000 ± 0.0000.982 ± 0.0170.973 ± 0.0260.900 ± 0.0130.851 ± 0.019
Table 2. Performance evaluation of 11 classifiers based on 512 features of Gerd-VGGNet feature extractor.
Table 2. Performance evaluation of 11 classifiers based on 512 features of Gerd-VGGNet feature extractor.
ClassifierTrainingValidationTest
AccuracyKappaAccuracyKappaAccuracyKappa
Random Forest1.000 ± 0.0001.000 ± 0.0000.982 ± 0.0170.973 ± 0.0260.900 ± 0.0130.851 ± 0.019
HistGB1.000 ± 0.0001.000 ± 0.0000.982 ± 0.0130.973 ± 0.0200.891 ± 0.0210.836 ± 0.031
Naive Bayes0.984 ± 0.0020.976 ± 0.0030.981 ± 0.0130.971 ± 0.0200.872 ± 0.0090.809 ± 0.014
MLP0.982 ± 0.0030.973 ± 0.0040.981 ± 0.0180.971 ± 0.0270.859 ± 0.0470.790 ± 0.070
XGB1.000 ± 0.0001.000 ± 0.0000.979 ± 0.0170.969 ± 0.0250.853 ± 0.0340.781 ± 0.051
KNN0.982 ± 0.0030.973 ± 0.0040.976 ± 0.0150.964 ± 0.0230.844 ± 0.0000.765 ± 0.001
RBF SVM0.982 ± 0.0020.974 ± 0.0030.981 ± 0.0180.971 ± 0.0270.844 ± 0.0140.766 ± 0.021
GB1.000 ± 0.0001.000 ± 0.0000.978 ± 0.0170.966 ± 0.0250.841 ± 0.0330.762 ± 0.048
Linear SVM0.997 ± 0.0010.995 ± 0.0020.982 ± 0.0170.973 ± 0.0260.803 ± 0.0140.705 ± 0.021
Decision Tree1.000 ± 0.0001.000 ± 0.0000.957 ± 0.0280.935 ± 0.0420.800 ± 0.0510.701 ± 0.075
Logistic Regression0.994 ± 0.0010.991 ± 0.0020.981 ± 0.0220.971 ± 0.0330.787 ± 0.0190.682 ± 0.028
Table 3. Performance evaluation of parameter optimization of GerdNet-RF classifier.
Table 3. Performance evaluation of parameter optimization of GerdNet-RF classifier.
Name of ExperimentNumber of
Estimators
Max DepthRandom StateMax FeaturesTrainingValidationTest
AccuracyKappaAccuracyKappaAccuracyKappa
Default100None1sqrt1.000 ± 0.0001.000 ± 0.0000.982 ± 0.0170.973 ± 0.0260.900 ± 0.0130.851 ± 0.019
Tuning 1200None1sqrt1.000 ± 0.0001.000 ± 0.0000.982 ± 0.0170.973 ± 0.0260.909 ± 0.0090.865 ± 0.014
Tuning 2300None1sqrt1.000 ± 0.0001.000 ± 0.0000.982 ± 0.0170.973 ± 0.0260.909 ± 0.0090.865 ± 0.014
Tuning 3400None1sqrt1.000 ± 0.0001.000 ± 0.0000.982 ± 0.0170.973 ± 0.0260.909 ± 0.0090.865 ± 0.014
Tuning 4200161sqrt1.000 ± 0.0001.000 ± 0.0000.982 ± 0.0170.973 ± 0.0260.909 ± 0.0090.865 ± 0.014
Tuning 520081sqrt1.000 ± 0.0001.000 ± 0.0000.982 ± 0.0170.973 ± 0.0260.909 ± 0.0090.865 ± 0.014
Tuning 620041sqrt0.992 ± 0.0020.988 ± 0.0030.982 ± 0.0170.973 ± 0.0260.912 ± 0.0120.869 ± 0.019
Tuning 720041log20.990 ± 0.0020.985 ± 0.0030.984 ± 0.0180.975 ± 0.0270.906 ± 0.0200.860 ± 0.030
Tuning 8200481sqrt0.992 ± 0.0020.988 ± 0.0030.982 ± 0.0170.973 ± 0.0260.912 ± 0.0120.869 ± 0.019
Tuning 9200481log20.991 ± 0.0010.986 ± 0.0020.984 ± 0.0140.975 ± 0.0210.925 ± 0.0210.888 ± 0.031
Table 4. Performance evaluation of five feature extraction layers of Gerd-VGGNet.
Table 4. Performance evaluation of five feature extraction layers of Gerd-VGGNet.
FeaturesTrainingValidationTest
AccuracyKappaAccuracyKappaAccuracyKappa
320.994 ± 0.0020.990 ± 0.0020.978 ± 0.0120.966 ± 0.0180.894 ± 0.0210.841 ± 0.031
640.993 ± 0.0010.989 ± 0.0010.978 ± 0.0120.966 ± 0.0180.909 ± 0.0260.864 ± 0.039
1280.993 ± 0.0010.990 ± 0.0020.981 ± 0.0120.971 ± 0.0180.891 ± 0.0290.835 ± 0.044
2560.993 ± 0.0010.989 ± 0.0020.981 ± 0.0120.971 ± 0.0180.887 ± 0.0320.831 ± 0.048
5120.991 ± 0.0010.986 ± 0.0020.984 ± 0.0140.975 ± 0.0210.925 ± 0.0210.888 ± 0.031
Table 5. Performance comparison of different AI systems for classification of GERD.
Table 5. Performance comparison of different AI systems for classification of GERD.
TaskAlgorithmData UsedEvaluation MethodOverall
Accuracy
Test
Accuracy
Binary
classification
Machine learning (ANN) [18]QUID 1 questionnaire
(577 GERD 2 patients, 94 normal cases)
hold-out99.2%NA
Binary
classification
Machine learning (HHDF-SVM) [19]147 RGB images
(39 GERD patients, 108 normal cases)
10-fold cross-validation93.2%NA
3-class
classification
Deep learning (Gerd-VGGNet) [20]671 NBI 3 images
(GERD A–B: GERD C–D: normal EC-J = 244:229:198)
10-fold cross-validation98.9% ± 0.5%78.8% ± 8.5%
3-class
classification
Deep learning + Machine learning (proposed GerdNet-RF)671 NBI 3 images
(GERD A–B: GERD C–D: normal EC-J = 244:229:198)
10-fold cross-validation99.0% ± 0.1%92.5% ± 2.1%
1 QUestionario Italiano Diagnostico = QUID; 2 Gastroesophageal reflux disease = GERD; 3 Narrow-band imaging = NBI.
Table 6. Three types of examples where both models predict correctly.
Table 6. Three types of examples where both models predict correctly.
ImageGERD Endoscopy GradingGerd-VGGNetGerdNet-RF
Diagnostics 12 02827 i001LA grade A–BLA grade A–BLA grade A–B
Diagnostics 12 02827 i002LA grade C–DLA grade C–DLA grade C–D
Diagnostics 12 02827 i003LA grade normalLA grade normalLA grade normal
Table 7. Three types of examples where two models predict incorrectly.
Table 7. Three types of examples where two models predict incorrectly.
ImageGERD Endoscopy GradingGerd-VGGNetGerdNet-RF
Diagnostics 12 02827 i004LA grade A–BLA grade C–DLA grade C–D
Diagnostics 12 02827 i005LA grade A–BLA grade C–DLA grade A–B
Diagnostics 12 02827 i006LA grade normalLA grade C–DLA grade normal
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yen, H.-H.; Tsai, H.-Y.; Wang, C.-C.; Tsai, M.-C.; Tseng, M.-H. An Improved Endoscopic Automatic Classification Model for Gastroesophageal Reflux Disease Using Deep Learning Integrated Machine Learning. Diagnostics 2022, 12, 2827. https://doi.org/10.3390/diagnostics12112827

AMA Style

Yen H-H, Tsai H-Y, Wang C-C, Tsai M-C, Tseng M-H. An Improved Endoscopic Automatic Classification Model for Gastroesophageal Reflux Disease Using Deep Learning Integrated Machine Learning. Diagnostics. 2022; 12(11):2827. https://doi.org/10.3390/diagnostics12112827

Chicago/Turabian Style

Yen, Hsu-Heng, Hui-Yu Tsai, Chi-Chih Wang, Ming-Chang Tsai, and Ming-Hseng Tseng. 2022. "An Improved Endoscopic Automatic Classification Model for Gastroesophageal Reflux Disease Using Deep Learning Integrated Machine Learning" Diagnostics 12, no. 11: 2827. https://doi.org/10.3390/diagnostics12112827

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop