Next Article in Journal
Detecting Audio Adversarial Examples in Automatic Speech Recognition Systems Using Decision Boundary Patterns
Previous Article in Journal
Image Decomposition Technique Based on Near-Infrared Transmission
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Disease Recognition in X-ray Images with Doctor Consultation-Inspired Model

1
Department of Computer Science, University of Dayton, Dayton, OH 45469, USA
2
Faculty of Software Engineering, University of Information Technology, Linh Trung Ward, Thu Duc District, Ho Chi Minh City 70000, Vietnam
*
Author to whom correspondence should be addressed.
J. Imaging 2022, 8(12), 323; https://doi.org/10.3390/jimaging8120323
Submission received: 25 August 2022 / Revised: 24 November 2022 / Accepted: 30 November 2022 / Published: 5 December 2022

Abstract

:
The application of chest X-ray imaging for early disease screening is attracting interest from the computer vision and deep learning community. To date, various deep learning models have been applied in X-ray image analysis. However, models perform inconsistently depending on the dataset. In this paper, we consider each individual model as a medical doctor. We then propose a doctor consultation-inspired method that fuses multiple models. In particular, we consider both early and late fusion mechanisms for consultation. The early fusion mechanism combines the deep learned features from multiple models, whereas the late fusion method combines the confidence scores of all individual models. Experiments on two X-ray imaging datasets demonstrate the superiority of the proposed method relative to baseline. The experimental results also show that early consultation consistently outperforms the late consultation mechanism in both benchmark datasets. In particular, the early doctor consultation-inspired model outperforms all individual models by a large margin, i.e., 3.03 and 1.86 in terms of accuracy in the UIT COVID-19 and chest X-ray datasets, respectively.

1. Introduction

Coronavirus disease 2019 (COVID-19) [1] is a contagious disease caused by a coronavirus called SARS CoV-2. This disease quickly spread worldwide, causing a global pandemic. Symptoms of COVID-19 appear 2–4 days after exposure. People with COVID-19 may experience fever or chills, cough or shortness of breath, breathing difficulties, headache, fatigue, and loss of smell or taste. According to the World Health Organization, COVID-19 infection can be detected by testing specimens from nose or mouth swabs. Real-time reverse transcription-polymerase chain reaction (RT-PCR) is used to detect nucleic acids in secretory fluids obtained from specimens. Because coinfection with other viruses can impact RT-PCR prediction performance, repetitive testing may be recommended to prevent false negatives. The RT–PCR test has a three-day turnaround time, as RT–PCR test tools have been scarce in recent months. There has been a pressing need for additional procedures to quickly and reliably identify COVID-19 patients. Furthermore, the swabbing operation is highly susceptible to expert errors, and it must be performed repeatedly [2]. Therefore, X-rays or CT scans of the chests are suitable complements to RT-PCR because they can be gathered and processed considerably more quickly [3]. Relative to CT scans, taking chest X-ray images is less expensive, radiation-exposing, and time-consuming. In addition, CT nuclear scanning delivers larger radiation doses than traditional X-rays scanning [4]. An X-ray of the chest produces 0.1 mSv, whereas a CT produces 70 times the amount. X-ray machines are widely available and quickly provide images for diagnosis. Thus, in this work, we focus on recognizing diseases in X-ray images.
Since the reintroduction of convolution neural networks (CNN) [5], deep learning has become dominant in many research fields, such as computer vision, natural language processing, and video/speech recognition. To date, deep learning has been adopted by a wide range of applications, owing to its scalability, speed, and efficiency, even outperforming humans in specific industrial processes [6]. As reviewed in [7], medical science is a relatively new field that is attempting to leverage the success of artificial intelligence and deep learning models. Developments in digital data collection, computer vision, and computation infrastructure have enabled AI applications to move into areas that were previously regarded as entirely human domains [8]. Deep learning in radiology is a game changer in terms of both quality and quantity when it comes to biomedical imaging explanation and data processing. Although machine learning and deep learning algorithms have demonstrated their ability to classify tumors and cancer progression, radiologists are still hesitant to use them [9]. One of the numerous advantages of machine learning in radiology is its capacity to automate or even replace radiologist scanning methods. Deep learning algorithms produce outcomes that are comparable to those of a top radiologist. However, situations in which resources are limited and requirements are particularly demanding, such as the COVID-19 pandemic, exemplify the need for a robust algorithm to assist medical professionals.
Having witnessed the extraordinary performance of deep learning in various tasks [10,11,12,13,14,15,16,17,18,19,20,21,22], we investigated deep learning models in this paper. Inspired by the efforts and experience of healthcare professionals such as doctors and specialists during the pandemic, we propose a doctor consultation-inspired model to fuse various deep learning models to produce accurate outputs.
The novelty of this work is as follows. First, the proposed framework is motivated from the perspective of physicians. The doctor consultation-inspired method is formulated in the form of fusion models. The proposed method considers each individual deep learning model as a medical doctor. Then, a consultation is performed based on inputs from multiple individual models. In this regard, the proposed method leverages the strengths of available methods in order to boost the performance. Second, the proposed method is open in the sense that any future individual methods can be integrated into our method. Third, we evaluate the proposed method on two benchmark datasets with different consultation modes, namely early consultation and late consultation.
The remainder of this paper is organized as follows. In Section 2, we summarize related works. In Section 3, we introduces the proposed doctor consultation-inspired model. The experiments and the experimental results are presented and discussed in Section 4. Finally, Section 5 concludes the paper.

2. Related Works

Many efforts to diagnose COVID-19 and pneumonia from X-ray images have been reported in the literature. In [10], deep neural network techniques were used in conjunction with X-ray imaging to identify COVID-19 infection. The main goal of this effort was to help alleviate doctor shortages in rural areas by providing resources to fill the gap. Shibly et al. [10] used VGG-16 [11] architecture to identify COVID-19 patients from chest X-ray images. The proposed method may aid medical professionals in screening COVID-19 patients. In another work, Sethy et al. [12] sought to detect coronavirus-infected patients using X-ray images. This method involves radiographic analysis using support vector machines (SVMs) with deep features extracted from ResNet50 [13]. The efficacy of a multi-CNN in automatically detecting COVID-19 from X-ray images was examined by Abraham et al. [14], who employed naive Bayes, SVM, AdaBoost, logistic regression, and random forests before settling on the Bayes net. The best-performing method is Xception [15]. Mei et al. [16] proposed a machine learning strategy that uses diagnostic imaging and clinical studies to accurately detect COVID-19-positive patients. The authors created a DCNN to learn the initial imaging characteristics of COVID-19 patients (18-layer residual network: ResNet-18 [13]). In the next stage, random forest, SVM, and MLP classifiers were used to categorize COVID-19 patients. Multilayer perceptron (MLP) performed best on the tuning set, and a neural network model was utilized to evaluate COVID-19 status based on radiographic and clinical data.
Hurt et al. [17] improved their method by just using frontal chest X-ray images. They discovered that the probabilities in their model that are matched to the quality of the imaging data are remarkably general and reliable. According to recent research, machine learning algorithms can distinguish COVID-19 from other pneumonia strains. Tuncer et al. [18] developed a technique for COVID-19 recognition using X-ray scans of the lungs. This technique is broken down into stages, which are detailed as follows. Residual example local binary pattern [19] is the name given to this method (ResExLBP). In the feature selection step, the IRF-based attribute selection method is used. Decision trees, linear classifiers, SVM, k-NN, and SD approaches are utilized in the classification step. Using 10-fold cross validation, the SVM classifier once again achieved the best performance.
Recently, Hemdan et al. [20] developed a deep learning framework to aid radiologists in detecting COVID-19 in X-ray scan. They investigated many deep artificial neural networks to classify the patient’s COVID-19 status as negative or positive. Machine learning classifiers VGG19 [11] and DenseNet201 [21] achieved the best results in predicting COVID-19 using two-dimensional X-ray images. Recently, a rapid COVID-19 diagnosis technique was proposed by Ardakani et al. [22]. The authors used ten well-known pre-trained CNNs for this purpose. They trained and tested the 10 CNNs using the same dataset and compared the results to a radiologist’s classifications. For COVID-19 individuals, ResNet-101 [13] achieved the best performance. Additionally, there are many deep learning models proposed for classification [23,24].
Many optimization and refinement steps have been proposed to improve the performance of classifiers. For example, data augmentation [5] enhances the size and quality of training datasets. Waheed et al. [25] proposed a GAN-based model to synthesize medical images, with the aim of increasing the number of training samples required to train a CNN-based model to detect COVID-19 from medical images. In another study, Oh et al. [26] proposed a patch-based deep neural network architecture that can be trained with a small dataset. Teixeira et al. [27] used a UNet-based lung segmentation model [28] to segment the lung first. Then, they used a CNN-based model to classify X-ray images. Similarly, Tartaglione et al. [29] adopted segmented lung images. Then, they used a feature extractor pretrained on CXR pathology datasets and fine-tuned it on COVID datasets. Balaha et al. [30] introduced a framework with a segmentation phase to segment lung regions. Then, data augmentation such as rotation, skewing, translation, and shifting was applied. Finally, a genetic algorithm was used to learn combinations of hyperparameters. Baghdadi et al. [31] presented an algorithm for COVID-19 classification using a CNN, pre-trained model, and Sparrow search algorithm on CT lung images. Perumal et al. [32] proposed a transfer learning model with Haralick features [33] to speed up the prediction process and assist medical professionals. Transfer learning alleviated the problem of the lack of COVID-19-positive data to some extent. A comparison of related works provided in Table 1. However, a review of all models used for COVID-19 detection is beyond the scope of this paper. Additional research works involving COVID-19, CNNs, and data augmentation were covered in [34,35,36].

3. Proposed Framework

3.1. Individual Doctor Models

In this work, we use the aforementioned deep learning models to simulate medical doctors. Figure 1 shows the architecture of the deep learning models. We adopt the available source code of these models for implementation.
He et al. [13] found that it is very difficult to train deep neural networks, indicating that models struggle with saturation and are very difficult to optimize. Therefore, they proposed a framework to reduce training via residual learning called ResNet. In particular, ResNet considers previous layers as the input layer reformulated as learning residual functions. Many ResNet variants have been developed, with the main difference being the number of layers. for example, ResNet-18, -50, -101, and -152.
Huang et al. [2] proposed a dense convolutional network (DenseNet), which connects each layer to every other layer in a feed-forward fashion. In particular, DenseNet distills the insight of a simple connectivity pattern, which optimizes information flow between layers and then directly connects all layers. DenseNet directly connects any layer to all subsequent layers to further improve the information flow between layers. In particular, the l-th layer has l inputs, consisting of the feature maps of all previous convolutional blocks. Then, the feature maps are passed on to all L-l subsequent layers, introducing L(L+1)/2 connections in an L-layer network. We investigate two DenseNet variants, namely DenseNet-169 and DenseNet-201.
Xie et al. [23] presented a strategy to expose a dimension called “cardinality” (i.e., the size of the set of transformations). They proposed ResNeXt, including a stack of residual blocks. ResNeXt is homogeneous and multibranched, with only a few hyperparameters to set. ResNeXt’s blocks follow two simple rules: the blocks share hyperparameters (width and filter sizes), and the width is multiplied by a factor of two. These rules are constructed while producing spatial maps of the same size and downsampling the spatial map by a factor of two each time. In addition, ResNeXt introduces the revisiting of simple neurons, which is the elementary transformation performed by fully-connected and convolutional layers. In particular, the inner product is a form of aggregating transformation. The analysis of a simple neuron replaces the elementary transformation with a more generic function. In this paper, we consider the widely used ResNeXt-101 variant.
Most state-of-the-art image classification [13,21,23] models follow the same general framework, which first encodes the image to achieve a low-resolution representation and then recovers the high-resolution representation. Wang et al. [24] proposed a framework with a high-resolution network (HRNet). HRNet maintains high-resolution representations throughout the whole process. In particular, it contains parallel multiresolution convolutions and multiresolution fusions. Parallel multiresolution convolutions start from a high-resolution convolution stream and gradually add high-to-low-resolution streams one by one, constructing new stages. Finally, it connects the multiresolution streams in parallel. As a result, the resolutions of a later stage consist of the resolutions from the previous stages. HRNet introduces repeated multiresolution fusions that exchange information across multiresolution representations. In our experiment, we leverage HRNetV2-W48, with a high-resolution width of 48.

3.2. Doctor Consultation-Inspired Model

In this work, multiple deep learning models are used to recognize diseases such as pneumonia and COVID-19 in X-ray images. In reality, a consultation session allows a team of healthcare professionals such as doctors and specialists to limit the damage of the acute respiratory distress syndrome (ARDS), evaluate recovery, manage lingering symptoms, and prevent a recurrence in the future. Follow-up is recommended post COVID syndrome to help the patient get back on track. Inspired by such consultation sessions, in this work, we consider each model as a doctor. Then, we combine the decisions of the various models to output diagnostic and prognostic results. There are two strategies available for doctor consultation models, namely late consultation and early consultation. The details are provided below.
Late Consultation. In the late consultation model, each doctor makes his/her own final decision. The consultation simply combines all of these decisions. In our method, to simulate this strategy, we fuse the prediction scores from individual models to output a final decision. In particular, we first train n   individual models on the training data. Then, we feed the images in the training set to each model ( i ) to output a prediction score ( p i ), where p i is a vector containing m values corresponding to m disease classes. We consider p i as the ith doctor’s final decision. We further train a classification model ( f l a t e ) to output the final decision ( y ^ l a t e ), as shown in Equation (1):
y ^ l a t e = f l a t e ( p 1 p 2 | | p n ] .
In Equation (1), [ . | | . ] denotes the concatenate operation.
Early Consultation. For the early consultation strategy, the decision is made based on the observations and discussions among all health professionals in the consultation. To simulate this strategy, we first train n individual models on the training data. Then, we feed the images in the training set to each model ( i ). Instead of obtaining the prediction scores ( p i s ), we fetch the deep-learned features ( x i ) of the individual model ( i ). The deep-learned features are extracted in the layer prior to the fully connected layer. The deep-learned features have been shown to be effective in classification tasks [5,37,38]. We normalize each individual deep-learned feature with l 2 normalization. We consider x i as the ith doctor’s observation/discussion. We further train a classification model ( f e a r l y ) to output the final decision ( y ^ e a r l y ), as expressed by Equation (2):
y ^ e a r l y = f e a r l y ( [ x 1 | |   x 2   x n ] ) .
For the classification model such, i.e., f l a t e or f e a r l y , we adopt support vector machine (SVM), which is popular for COVID-19 recognition tasks [12,15,16,18]. Specifically, SVM seeks a hyperplane in a high-dimensional space to maximize the margin between classes.
Figure 2 illustrates the two aforementioned consultation strategies. In this work, we investigate both consultation strategies in the evaluation. For reading clarity, the abbreviations and symbols used are listed on Table 2 and Table 3, respectively.

4. Experiments

4.1. Experimental Settings

In this work, we first use the available UIT COVID-19 dataset [39]. This dataset consists of 1317 images annotated in 3 classes, namely COVID-19, pneumonia, and normal. Because COVID-19 illness symptoms are similar to pneumonia symptoms, and pneumonia disease is responsible for many of COVID-19 virus-related deaths, it is logical to combine the two diagnostic techniques. The two subsets, i.e., the training set and testing set, comprise 1053 and 264 images, respectively.
We also conduct experiments on a chest X-ray dataset [40]. This dataset consists of 6432 X-ray images with 3 classes: COVID-19, pneumonia, and normal. The dataset is organized into 2 subsets, namely a training set and a testing set. In particular, there are 5144 images in the training set and 1288 images in the testing set.
Regarding the performance metrics, we report the results in terms of accuracy, precision, recall, and F1 score. In particular, A c c u r a c y describes the number of correct predictions over all predictions.
A c c u r a c y = T r u e   P o s i t i v e + T r u e   N e g a t i v e   T r u e   P o s i t i v e + F a l s e   P o s i t i v e + T r u e   N e g a t i v e + F a l s e   N e g a t i v e .
Here, a true/false positive is an outcome for which the model correctly/incorrectly predicts the positive class. Similarly, a true/false negative is an outcome for which the model correctly/incorrectly predicts the negative class. The second metric, P r e c i s i o n , is a measure of true positives.
P r e c i s i o n = T r u e   P o s i t i v e T r u e   P o s i t i v e + F a l s e   P o s i t i v e .
R e c a l l is a measure of the number of correctly predicted positive cases over all positive cases in the dataset.
R e c a l l = T r u e   P o s i t i v e T r u e   P o s i t i v e + F a l s e   N e g a t i v e .
F1 score is a measure combining both precision and recall. It is generally described as the harmonic mean of the two. The formula for the F1 score is expressed as:
F 1 = 2 × P r e c i s i o n × R e c a l l   P r e c i s i o n + R e c a l l .
We conduct our experiments on a CPU Intel (R) Core(TM) i9-10900X CPU @ 3 with 64 GB of RAM and one GeForce RTX 2080 Ti 12GB. The experimental configurations are primarily leveraged from MMClassification Toolbox and Benchmark version 0.24.0 based on PyTorch V1.8.1 version [41]. In particular, we adopt the configuration achieving the best performance on the Image-Net classification task. The SVM classifier is trained by using the scikit-learn library.

4.2. Experimental Results

We first conduct an ablation study to evaluate the performance of individual doctor models and the proposed doctor consultation-inspired method with both early and late consultation mechanisms.
Table 4 shows the performance of the various methods on the UIT COVID-19 benchmark dataset [39] in terms of accuracy, precision, recall, and F1 score, as described in Equations (3)–(6), respectively. Generally, complicated models achieve superior performance. For example, ResNet-152 performs better than other ResNet variants. Among the individual models, DenseNet-201 and HRNet obtain the top performance, owing to their advanced architectures. The doctor consultation models achieve better performance than the individual models, indicating the effectiveness of the proposed method for the task of anomaly analysis in medical image processing. The early consultation model outperforms the late consultation model, i.e., 94.32, 94.36, 94.32, and 94.31 vs. 92.42, 92.42, 92.42, and 92.42 in terms of accuracy, precision, recall, and F1 score, respectively. These results imply that the concatenation of features from individual doctor models is useful in making a final prediction. The fusional nature of the late consultation model may be biased by the “good” individual models, for example, HRNet or DenseNet.
We then conduct experiments on the chest X-ray dataset [40]. The results are shown in Table 5. DenseNet-201 and HRNet-W48 achieve the top-2 performance among the individual models, i.e., 93.17 and 92.62 in terms of accuracy, respectively. Unlike the UIT COVID-19 dataset, the ResNet variants are outperformed by ResNeXt-101 on this benchmark. The late consultation and early consultation mechanisms obtain the top-2 highest scores across all metrics. The early consultation mode once again surpasses the late consultation model.
We further visualize the classification results of the various methods in the UIT COVID-19 benchmark. Figure 3 shows the visualization of the prediction results from the baselines and our two consultation modes. As shown in the figure, all models perform well in predicting the results in the first two columns. The third column shows failure predictions of individual models. However, both consultation models output correct predictions. The last column demonstrates the advantage of the early consultation over the late consultation strategy. In particular, whereas the late consultation follows all incorrect decisions of individual models, the early consultation model yields the correct prediction, indicating the effectiveness of the proposed model in handling difficult cases.
We observe the consistent performance of the early consultation mode in both UIT COVID-19 and Chest X-ray datasets, outperforming all individual models by a large margin, i.e., 3.03 and 1.86 in terms of accuracy on the UIT COVID-19 and chest X-ray datasets, respectively. However, the individual models are inconsistent. For example, ResNet-50 does not perform well on the UIT COVID-19 dataset but achieves a high performance on the chest X-ray dataset. Furthermore, we evaluate the performance of state-of-the-art baselines on the two benchmark datasets. As shown in Table 6, our proposed method achieves the best performance on both sets. Specifically, the early consultation method outperforms the late consultation method. The baselines are inconsistent between both sets. Here, we would like to highlight the limitations of the proposed work. First, the performance of the late and early consultation models heavily relies on the performance of the individual models. If all of individual models achieve a low performance, this hurts the overall performance of the doctor consultation-inspired model. Second, because SVM is adopted for fusion, the proposed framework lacks explainability.

5. Conclusions

In this paper, we propose a doctor consultation-inspired method for recognizing disease from X-ray images. Inspired by doctor consultation practice, we explore two modes, namely late fusion and early fusion. The proposed method takes advantage of multiple state-of-the-art networks to efficiently recognize disease from an input X-ray image. The early fusion mechanism combines the deep-learned features of various models, whereas the late fusion method combines the confidence scores of all individual models. Experiments show the superiority of the proposed method over individual methods. Both fusion mechanisms outperform baselines by a large margin. In addition, the early fusion model consistently outperforms the late fusion mechanism on the two benchmark datasets. In particular, the early doctor consultation-inspired model outperforms all individual models by a large margin, i.e., 3.03 and 1.86 in terms of accuracy on the UIT COVID-19 and chest X-ray datasets, respectively.
In the future, we intend to extend our model for different diseases. Moreover, we plan to explore different kinds of medical imaging, such as CT scans or MRI. The proposed method also has the potential to integrate additional individual models to better recognize disease from an input X-ray image. The proposed method addresses the classification problem. Therefore, we intend to investigate the effectiveness of the proposed method on various tasks, such as semantic segmentation or instance segmentation in medical images.

Author Contributions

Conceptualization, K.A.P., N.W., and K.N; methodology, K.A.P., N.W., T.T.N., and N.D.V.; software, K.A.P., N.W., T.T.N., and N.D.V.; validation, K.A.P., S.B., T.T.N., and N.D.V.; formal analysis, K.A.P., S.B., T.T.N., and N.D.V.; investigation, K.A.P., S.B., T.T.N., and N.D.V.; resources, K.N.; data curation, T.T.N., N.D.V., and K.N.; writing—original draft preparation, K.A.P., N.W., T.T.N., and N.D.V.; writing—review and editing, K.A.P., N.W., T.T.N., S.B., N.D.V., and K.N.; visualization, K.A.P.; supervision, K.N.; project administration, K.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

COVID-19 dataset is available at https://github.com/nguyenvd-uit/uit-together-dataset/blob/main/COVID-19.md (accessed on 10 November 2022). Chest X-ray dataset is available at: https://www.kaggle.com/datasets/prashant268/chest-xray-covid19-pneumonia (accessed on 10 November 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. COVID-19 Pandemic. Available online: https://www.who.int/emergencies/diseases/novel-coronavirus-2019 (accessed on 20 August 2022).
  2. Ai, T.; Yang, Z.; Hou, H.; Zhan, C.; Chen, C.; Lv, W.; Tao, Q.; Sun, Z.; Xia, L. Correlation of chest CT and RT-PCR testing in coronavirus disease 2019 (COVID-19) in China: A report of 1014 cases. Radiology 2020, 296, E32–E40. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Liu, R.; Han, H.; Liu, F.; Lv, Z.; Wu, K.; Liu, Y.; Feng, Y.; Zhu, C. Positive rate of RT-PCR detection of SARS-CoV-2 infection in 4880 cases from one hospital in Wuhan, China, from Jan to Feb 2020. Clin. Chim. Acta 2020, 505, 172–175. [Google Scholar] [CrossRef] [PubMed]
  4. Radiation Risk from Medical Imaging. Available online: https://www.health.harvard.edu/cancer/radiation-risk-from-medical-imaging (accessed on 20 August 2022).
  5. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  6. Davenport, T.; Kalakota, R. The potential for artificial intelligence in healthcare. Future Healthc. J. 2019, 6, 94. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Phung, K.A.; Kirbas, C.; Dereci, L.; Nguyen, T.V. Pervasive Healthcare Internet of Things: A Survey. J. Inf. 2022, 13, 360. [Google Scholar] [CrossRef]
  8. Yu, K.H.; Beam, A.L.; Kohane, I.S. Artificial intelligence in healthcare. Nat. Biomed. Eng. 2018, 2, 719–731. [Google Scholar] [CrossRef]
  9. Noguerol, T.M.; Paulano-Godino, F.; Martín-Valdivia, M.T.; Menias, C.O.; Luna, A. Strengths, weaknesses, opportunities, and threats analysis of artificial intelligence and machine learning applications in radiology. J. Am. Coll. Radiol. 2019, 16, 1239–1247. [Google Scholar] [CrossRef]
  10. Shibly, K.H.; Dey, S.K.; Islam, M.T.U.; Rahman, M.M. COVID faster R–CNN: A novel framework to Diagnose Novel Coronavirus Disease (COVID-19) in X-Ray images. Inform. Med. Unlocked 2020, 20, 100405. [Google Scholar] [CrossRef] [PubMed]
  11. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  12. Sethy, P.K.; Behera, S.K. Detection of Coronavirus Disease (COVID-19) Based on Deep Features. Preprints 2020, 2020030300. [Google Scholar] [CrossRef] [Green Version]
  13. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  14. Abraham, B.; Nair, M.S. Computer-aided detection of COVID-19 from X-ray images using multi-CNN and Bayesnet classifier. Biocybern. Biomed. Eng. 2020, 40, 1436–1445. [Google Scholar] [CrossRef] [PubMed]
  15. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  16. Mei, X.; Lee, H.C.; Diao, K.Y.; Huang, M.; Lin, B.; Liu, C.; Xie, Z.; Ma, Y.; Robson, P.M.; Chung, M.; et al. Artificial intelligence–enabled rapid diagnosis of patients with COVID-19. Nat. Med. 2020, 26, 1224–1228. [Google Scholar] [CrossRef] [PubMed]
  17. Hurt, B.; Kligerman, S.; Hsiao, A. Deep learning localization of pneumonia: 2019 coronavirus (COVID-19) outbreak. J. Thorac. Imaging 2020, 35, 87–89. [Google Scholar] [CrossRef]
  18. Tuncer, T.; Dogan, S.; Ozyurt, F. An automated Residual Exemplar Local Binary Pattern and iterative ReliefF based COVID-19 detection method using chest X-ray image. Chemom. Intell. Lab. Syst. 2020, 203, 104054. [Google Scholar] [CrossRef] [PubMed]
  19. Zhao, G.; Pietikainen, M. Dynamic texture recognition using local binary patterns with an application to facial expressions. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 915–928. [Google Scholar] [CrossRef] [Green Version]
  20. Hemdan, E.E.D.; Shouman, M.A.; Karar, M.E. Covidx-net: A framework of deep learning classifiers to diagnose COVID-19 in X-ray images. arXiv 2020, arXiv:2003.11055. [Google Scholar]
  21. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  22. Ardakani, A.A.; Kanafi, A.R.; Acharya, U.R.; Khadem, N.; Mohammadi, A. Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: Results of 10 convolutional neural networks. Comput. Biol. Med. 2020, 121, 103795. [Google Scholar] [CrossRef]
  23. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  24. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep high-resolution representation learning for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3349–3364. [Google Scholar] [CrossRef] [Green Version]
  25. Waheed, A.; Goyal, M.; Gupta, D.; Khanna, A.; Al-Turjman, F.; Pinheiro, P.R. Covidgan: Data augmentation using auxiliary classifier gan for improved COVID-19 detection. IEEE Access 2020, 8, 91916–91923. [Google Scholar] [CrossRef]
  26. Oh, Y.; Park, S.; Ye, J.C. Deep learning COVID-19 features on CXR using limited training data sets. IEEE Trans. Med. Imaging 2020, 39, 2688–2700. [Google Scholar] [CrossRef]
  27. Teixeira, L.O.; Pereira, R.M.; Bertolini, D.; Oliveira, L.S.; Nanni, L.; Cavalcanti, G.D.; Costa, Y.M. Impact of lung segmentation on the diagnosis and explanation of COVID-19 in chest X-ray images. Sensors 2021, 21, 7116. [Google Scholar] [CrossRef]
  28. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the 18th International Conference on Medical Image Computing and Computer Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Munich, Germany, 2015; pp. 234–241. [Google Scholar]
  29. Tartaglione, E.; Barbano, C.A.; Berzovini, C.; Calandri, M.; Grangetto, M. Unveiling COVID-19 from chest x-ray with deep learning: A hurdles race with small data. Int. J. Environ. Res. Public Health 2020, 17, 6933. [Google Scholar] [CrossRef]
  30. Balaha, H.M.; Balaha, M.H.; Ali, H.A. Hybrid COVID-19 segmentation and recognition framework (HMB-HCF) using deep learning and genetic algorithms. Artif. Intell. Med. 2021, 119, 102156. [Google Scholar] [CrossRef] [PubMed]
  31. Baghdadi, N.A.; Malki, A.; Abdelaliem, S.F.; Balaha, H.M.; Badawy, M.; Elhosseini, M. An automated diagnosis and classification of COVID-19 from chest CT images using a transfer learning-based convolutional neural network. Comput. Biol. Med. 2022, 144, 105383. [Google Scholar] [CrossRef] [PubMed]
  32. Perumal, V.; Narayanan, V.; Rajasekar, S.J.S. Detection of COVID-19 using CXR and CT images using Transfer Learning and Haralick features. Appl. Intell. 2021, 51, 341–358. [Google Scholar] [CrossRef] [PubMed]
  33. Porebski, A.; Vandenbroucke, N.; Macaire, L. Haralick feature extraction from LBP images for color texture classification. In Proceedings of the 2008 First Workshops on Image Processing Theory, Tools and Applications, Sousse, Tunisia, 23–26 November 2008; pp. 1–8. [Google Scholar]
  34. Yu, W.; Hargreaves, C.A. A review study of the deep learning techniques used for the classification of chest radiological images for COVID-19 diagnosis. Int. J. Inf. Manag. Data Insights 2022, 2, 100100. [Google Scholar]
  35. Clement, J.C.; Ponnusamy, V.; Sriharipriya, K.C.; Nandakumar, R. A survey on mathematical, machine learning and deep learning models for COVID-19 transmission and diagnosis. IEEE Rev. Biomed. Eng. 2021, 15, 325–340. [Google Scholar]
  36. Mohamad, Y.I.; Baraheem, S.S.; Nguyen, T.V. Olympic Games Event Recognition via Transfer Learning with Photobombing Guided Data Augmentation. J. Imaging 2021, 7, 12. [Google Scholar] [CrossRef]
  37. Karpathy, A.; Toderici, G.; Shetty, S.; Leung, T.; Sukthankar, R.; Fei-Fei, L. Large-scale video classification with convolutional neural networks. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1725–1732. [Google Scholar]
  38. Zhou, B.; Lapedriza, A.; Khosla, A.; Oliva, A.; Torralba, A. Places: A 10 million image database for scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 1452–1464. [Google Scholar] [CrossRef]
  39. COVID-19 Dataset. Available online: https://github.com/nguyenvd-uit/uit-together-dataset/blob/main/COVID-19.md (accessed on 20 August 2022).
  40. Chest X-ray Dataset. Available online: https://www.kaggle.com/datasets/prashant268/chest-xray-covid19-pneumonia (accessed on 20 October 2022).
  41. MMClassification Toolbox. Available online: https://github.com/open-mmlab/mmclassification (accessed on 20 October 2022).
Figure 1. Architecture of deep learning models referred to as individual doctors utilized in our framework. From top to bottom: ResNet, DenseNet, ResNeXt, and HRNet.
Figure 1. Architecture of deep learning models referred to as individual doctors utilized in our framework. From top to bottom: ResNet, DenseNet, ResNeXt, and HRNet.
Jimaging 08 00323 g001
Figure 2. The framework of the proposed doctor consultation-inspired model. In particular, there are two fusion mechanisms, namely early fusion and late fusion. The final output of either mechanism is the prediction label.
Figure 2. The framework of the proposed doctor consultation-inspired model. In particular, there are two fusion mechanisms, namely early fusion and late fusion. The final output of either mechanism is the prediction label.
Jimaging 08 00323 g002
Figure 3. Visualization of the prediction results of various models (best viewed online in color with zoom). From top to bottom: ResNeXt101, ResNet152, DenseNet201, HRNet-W48, late consultation, and early consultation. Incorrect predictions are marked with a red rectangle.
Figure 3. Visualization of the prediction results of various models (best viewed online in color with zoom). From top to bottom: ResNeXt101, ResNet152, DenseNet201, HRNet-W48, late consultation, and early consultation. Incorrect predictions are marked with a red rectangle.
Jimaging 08 00323 g003
Table 1. Comparison of related works.
Table 1. Comparison of related works.
MethodYearClassificationLung SegmentationRefinement/Remarks
Shibly et al. [10]2020VGG-16NoNo
Sethy et al. [12]2020ResNet-50, SVMNoNo
Abraham et al. [14]2020Xception, Bayes NetNoNo
Mei et al. [16]2020ResNet-18, MLPNoNo
Tuncer et al. [18]2020Local Binary Pattern, SVMNoIRF-based feature selection
Hemdan et al. [20]2020DenseNet-201NoNo
Ardakani et al. [22]2020ResNet-101NoNo
Waheed et al. [25]2020CNNNoGAN-based data augmentation
Tartaglione et al. [29]2020ResNet-18YesSegmented lung
Perumal et al. [32]2021CNNNoTransfer learning with Haralick features, CT scan
Teixeira et al. [27]2021InceptionV3YesSegmented lung
Balaha et al. [30]2021CNNYesGeometric transformation-based data augmentation, segmented lung, genetic algorithm
Baghdadi et al. [31]2022CNNNoSparrow search algorithm, CT scan
Ours2022CNN, SVMNoDoctor consultation-inspired fusion
Table 2. Table of abbreviations/acronyms used in this paper.
Table 2. Table of abbreviations/acronyms used in this paper.
AbbreviationMeaning
CXRChest X-ray
CT scanComputed tomography scan
MRIMagnetic resonance imaging
RT-PCRReal-time reverse transcription-polymerase chain reaction
ARDSAcute respiratory distress syndrome
CNNConvolutional neural network
ResNetResidual neural network
HRNetHigh-resolution network
DenseNetDense convolutional network
SVMSupport vector machine
Table 3. Table of symbols used in this paper.
Table 3. Table of symbols used in this paper.
SymbolMeaning
n The number of models (doctors)
p i The prediction score of model i
m The number of classes, such as COVID, pneumonia, and normal
[ . | | . ] Concatenation operation
f l a t e The classification function for late fusion
x i The deep-learned features extracted from model i
f e a r l y The classification function for early fusion
l 2 normThe square root of the inner product of a vector with itself
Table 4. Ablation study on the UIT COVID-19 dataset [39]. The performance of individual doctor models and two implementations of doctor consultation-inspired models. The top-two methods are marked in red and blue, respectively.
Table 4. Ablation study on the UIT COVID-19 dataset [39]. The performance of individual doctor models and two implementations of doctor consultation-inspired models. The top-two methods are marked in red and blue, respectively.
ModelAccuracyPrecisionRecallF1 Score
ResNet-1889.3989.7089.3989.34
ResNet-5090.1590.1890.1590.09
ResNet-10190.9190.8790.9190.87
ResNet-15290.5390.6490.5390.46
ResNeXt-10190.5390.6090.5390.46
DenseNet-16990.5392.0592.0592.03
DenseNet-20191.6791.8291.6791.64
HRNet-W4891.2991.2991.2991.26
Late Consultation92.4292.4292.4292.42
Early Consultation94.7094.7094.7094.70
Table 5. Ablation study on the chest X-ray dataset [40]. The performance of individual doctor models and two implementations of doctor consultation-inspired models. The top-two methods are marked in red and blue, respectively.
Table 5. Ablation study on the chest X-ray dataset [40]. The performance of individual doctor models and two implementations of doctor consultation-inspired models. The top-two methods are marked in red and blue, respectively.
ModelAccuracyPrecisionRecallF1 Score
ResNet-1889.7590.4289.7589.93
ResNet-5092.4792.4992.4792.42
ResNet-10190.5390.4990.5390.37
ResNet-15289.5289.4589.5289.37
ResNeXt-10192.5592.5492.5592.49
DenseNet-16992.0091.9592.0091.95
DenseNet-20193.1793.1693.1793.12
HRNet-W4892.6292.6090.6292.56
Late Consultation93.9493.9293.9493.93
Early Consultation95.0395.0395.0395.03
Table 6. Comparison with state-of-the-art baselines. The top-two methods are marked in red and blue, respectively.
Table 6. Comparison with state-of-the-art baselines. The top-two methods are marked in red and blue, respectively.
MethodUIT COVID-19 DatasetChest X-ray Dataset
AccuracyPrecisionRecallAccuracyPrecisionRecall
Shibly et al. [17]90.2490.2490.2490.6890.6090.68
Sethy et al. [18]90.1590.1890.1592.4792.4992.47
Abraham et al. [38]88.2489.2888.2492.0091.9992.00
Mei et al. [39]89.3989.7089.3989.7590.4289.75
Hemdan et al. [32]91.6791.8291.6793.1793.1693.17
Ardakani et al. [22]90.9190.8790.9190.5390.4990.53
Late Consultation92.4292.4292.4293.9493.9293.94
Early Consultation94.7094.7094.7095.0395.0395.03
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Phung, K.A.; Nguyen, T.T.; Wangad, N.; Baraheem, S.; Vo, N.D.; Nguyen, K. Disease Recognition in X-ray Images with Doctor Consultation-Inspired Model. J. Imaging 2022, 8, 323. https://doi.org/10.3390/jimaging8120323

AMA Style

Phung KA, Nguyen TT, Wangad N, Baraheem S, Vo ND, Nguyen K. Disease Recognition in X-ray Images with Doctor Consultation-Inspired Model. Journal of Imaging. 2022; 8(12):323. https://doi.org/10.3390/jimaging8120323

Chicago/Turabian Style

Phung, Kim Anh, Thuan Trong Nguyen, Nileshkumar Wangad, Samah Baraheem, Nguyen D. Vo, and Khang Nguyen. 2022. "Disease Recognition in X-ray Images with Doctor Consultation-Inspired Model" Journal of Imaging 8, no. 12: 323. https://doi.org/10.3390/jimaging8120323

APA Style

Phung, K. A., Nguyen, T. T., Wangad, N., Baraheem, S., Vo, N. D., & Nguyen, K. (2022). Disease Recognition in X-ray Images with Doctor Consultation-Inspired Model. Journal of Imaging, 8(12), 323. https://doi.org/10.3390/jimaging8120323

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop