Next Article in Journal
Non-A Non-B Acute Aortic Dissection: Is There Some Confusion in the Radiologist’s Mind?
Next Article in Special Issue
Residual Lung Abnormalities in Survivors of Severe or Critical COVID-19 at One-Year Follow-Up Computed Tomography: A Narrative Review Comparing the European and East Asian Experiences
Previous Article in Journal
Transnasal Endoscopic Pituitary Surgery—The Role of a CT Scan in Individual Tailoring of Posterior Septum Size Resection
Previous Article in Special Issue
Diagnostic Performance in Differentiating COVID-19 from Other Viral Pneumonias on CT Imaging: Multi-Reader Analysis Compared with an Artificial Intelligence-Based Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MSTAC: A Multi-Stage Automated Classification of COVID-19 Chest X-ray Images Using Stacked CNN Models

by
Thanakorn Phumkuea
1,*,†,
Thakerng Wongsirichot
2,*,†,
Kasikrit Damkliang
2,
Asma Navasakulpong
3 and
Jarutas Andritsch
4
1
College of Digital Science, Prince of Songkla University, Songkhla 90110, Thailand
2
Division of Computational Science, Faculty of Science, Prince of Songkla University, Songkhla 90110, Thailand
3
Division of Respiratory and Respiratory Critical Care Medicine, Prince of Songkla University, Songkhla 90110, Thailand
4
Faculty of Business, Law and Digital Technologies, Solent University, Southampton SO14 0YN, UK
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Tomography 2023, 9(6), 2233-2246; https://doi.org/10.3390/tomography9060173
Submission received: 2 November 2023 / Revised: 8 December 2023 / Accepted: 8 December 2023 / Published: 13 December 2023
(This article belongs to the Special Issue The Challenge of Advanced Medical Imaging Data Analysis in COVID-19)

Abstract

:
This study introduces a Multi-Stage Automated Classification (MSTAC) system for COVID-19 chest X-ray (CXR) images, utilizing stacked Convolutional Neural Network (CNN) models. Suspected COVID-19 patients often undergo CXR imaging, making it valuable for disease classification. The study collected CXR images from public datasets and aimed to differentiate between COVID-19, non-COVID-19, and healthy cases. MSTAC employs two classification stages: the first distinguishes healthy from unhealthy cases, and the second further classifies COVID-19 and non-COVID-19 cases. Compared to a single CNN-Multiclass model, MSTAC demonstrated superior classification performance, achieving 97.30% accuracy and sensitivity. In contrast, the CNN-Multiclass model showed 94.76% accuracy and sensitivity. MSTAC’s effectiveness is highlighted in its promising results over the CNN-Multiclass model, suggesting its potential to assist healthcare professionals in efficiently diagnosing COVID-19 cases. The system outperformed similar techniques, emphasizing its accuracy and efficiency in COVID-19 diagnosis. This research underscores MSTAC as a valuable tool in medical image analysis for enhanced disease classification.

1. Introduction

The Coronavirus Disease 2019 (COVID-19) was formally named Severe Acute Respiratory Syndrome Coronavirus-2 (SARS-CoV-2) by the International Committee on Taxonomy of Viruses [1]. According to the primary studies, SARS-CoV-2 was originally transmitted from bats to humans by unknown intermediate animals [2,3]. Most patients exhibit symptoms such as fever, dry cough, headache, chest pain, and shortness of breath [4,5], even though some patients may experience asymptomatic COVID-19 [6,7]. For diagnosing COVID-19, a throat swab is the gold standard [8,9,10]. Typically, a chest X-ray (CXR) is essential for high-risk patients or Patients Under Inspection (PUI) to detect lung infection. The CXR examination is widely considered the gold standard for diagnosing pneumonia [11]. Specifically, the severity of COVID-19 is exacerbated when patients have risk factors such as older age, high blood pressure, and chronic respiratory disease [12]. The causes of pneumonia include viruses, bacteria, and fungi [13].
Deep learning (DL) serves as an advanced tool for image analysis, object classification, and pattern recognition. DL models accurately identify and classify objects within images, automatically learning complex patterns and representations directly from raw input data. Moreover, they can differentiate patterns and features that are challenging to extract from large datasets. Consequently, these models can be leveraged to develop computer-aided diagnosis (CADx) systems to effectively analyze medical information. With the assistance of artificial intelligence algorithms, CADx systems enable physicians to provide valuable information and support to fellow healthcare professionals. Furthermore, CADx reduces the workload on them by efficiently analyzing a large amount of medical data. However, it is essential to note that CADx systems are designed to assist rather than substitute for physicians. Final diagnostic decisions should always be made by qualified healthcare professionals, drawing on their clinical knowledge, patient history, and other relevant information.
According to previous studies, CADx may help clinicians identify COVID-19 pneumonia using CXR images [14,15]. Currently, CADx of medical images is becoming increasingly viable. State-of-the-art DL approaches have shown promising results in identifying CXR images of COVID-19 infections [16,17,18]. Chowdhury et al. [19] proposed a transfer learning approach with binary classification: normal and COVID-19, and three-class classification: normal, viral pneumonia, and COVID-19. The binary classification achieved 99.7% accuracy and sensitivity, while the three-class classification obtained 97.4% accuracy, sensitivity, and F-measure. Afshar et al. [20] proposed the COVID-CAPS modeling framework based on capsule networks, which achieved accuracy and sensitivity of 95.7% and 90.0%, respectively. Wang et al. [21] introduced COVID-net, a deep convolutional neural networks (CNN) design tailored to classify COVID-19, non-COVID-19, and healthy. COVID-net achieved an accuracy of 93.3% and a sensitivity of 93.3%. Ahmad et al. [22] proposed a hybrid ensemble model for identifying COVID-19 infections in CXR images. MobileNet and InceptionV3 models were fine-tuned for the learned weights. The authors performed four-fold cross-validation. The results showed accuracy, precision, and specificity of 94.2%, 89.9%, and 88.3%, respectively.
Furthermore, Khuzani, Heidari, and Shariati [23] used a dimensionality reduction method to generate a set of optimal features of CXR images to develop a COVID-classifier that distinguishes COVID-19 cases from non-COVID-19 cases. The proposed COVID-classifier achieved an accuracy of 94.05%. Yang et al. [24] used VGG16 to detect multiclass COVID-19 CXR and CT-scan images. The accuracy of the three-class classification of CXR images was 97%. Hong et al. [25] proposed a lightweight CNN, MGMADS-CNN that classifies COVID-19 CXR images with an accuracy of 96.75%. Nishio et al. [26] developed EfficientNet to classify COVID-19 pneumonia, non-COVID-19 pneumonia, and healthy CXR images. Three-class classification accuracy was 86.67%. According to Table 1, DL and pre-trained models for COVID-19 detection on CXR images dominated this research field. However, several studies used small and unbalanced datasets to develop models. This raises potential problems with generalization and bias. Small datasets may restrict the model’s ability to understand the full complexity of cases, and unbalanced datasets can result in biased performance towards the majority class. Even with the prevalence of CXR-pretrained models, researchers should be careful about the size and balance of their datasets to ensure strong and unbiased model performance in detecting COVID-19.
Our retrospective study mainly aimed to classify COVID-19, non-COVID-19, and healthy CXR images using a stacked CNN model. Using stacked CNN models, multi-stage automated classification (MSTAC) of COVID-19 CXR images was realized based on two CNN-based binary classification models. The first model differentiates healthy and unhealthy CXR images. Consequently, all the unhealthy CXR images are passed to the second CNN-based binary classification model. The second model mainly classifies COVID-19 and non-COVID-19 CXR images. We compared MSTAC with the CNN model that classifies COVID-19 pneumonia, non-COVID-19, and healthy. However, only a few studies have used multi-stage automated detection for classifying COVID-19 based on CXR images.
The following are the main contributions of our study that aim to address these issues:
(1)
We propose the Multi-Stage Automated Classification (MSTAC) for COVID-19 detection using two stacked CNN-based binary classification models and a large dataset of CXR images.
(2)
The first model for MSTAC (stage-1) is a DL model that classifies between healthy and unhealthy CXR images.
(3)
The second model for MSTAC (stage-2) is a DL model that classifies between COVID-19 and non-COVID-19 CXR images.
(4)
The proposed MSTAC model stacks stage-1 and stage-2. This stacking allows MSTAC to identify three classes of CXR images: COVID-19, non-COVID-19, and healthy.
(5)
MSTAC achieves high accuracy in identifying COVID-19 pneumonia cases from CXR images. This proposed model enhances the efficiency and accuracy of COVID-19 detection.
(6)
We also compare MSTAC with previous studies that classify COVID-19 pneumonia, non-COVID-19, and healthy cases. This comparison demonstrates the promising results of MSTAC in COVID-19 classification and its effectiveness for medical image analysis and diagnosis.
Overall, the MSTAC for COVID-19 model aims to make a precise contribution to the ongoing struggle with the COVID-19 pandemic.
The study’s structure is organized as follows: The methodology and proposed models are discussed in Section 2. The experimental findings are presented in Section 3. In Section 4, discussions are provided. Finally, Section 5 presents the conclusion.

2. Materials and Methods

2.1. Dataset

Three publicly available data repositories were used: (1) CXR images (pneumonia) dataset available on Mendeley data, which contains CXR images of healthy and non-COVID-19 pneumonia [27]; (2) augmented COVID-19 X-ray images dataset [28]; and (3) COVID-19 image data [29] for CXR images of COVID-19 pneumonia. The representative CXR images of healthy, non-COVID-19 pneumonia, and COVID-19 in the combined dataset are shown in Figure 1.
The dataset used comprised 3150 CXR images. Specifically, 1050 images were randomly selected from dataset (1) for both the healthy and non-COVID-19 pneumonia classes. For COVID-19 pneumonia, 912 CXR images were collected from dataset (2). A total of 138 CXR images were randomly selected from dataset (3) to prevent a class imbalance. Our dataset included 3150 CXR images from all three classes, as shown in Table 2, and was used to develop and validate our proposal.

2.2. DL Architecture

The DL architecture employed in our study utilizes a sequential CNN for the classification of CXR images. As the CXR images were gathered from various public sources, it is necessary to standardize the image size. To ensure uniformity, the CXR images were standardized to dimensions of 224 × 224. Scaling CXR images to 224 × 224, despite their relatively low resolution, this study was conducted under the influence of resource constraints. Despite the reduced resolution, these previous studies have consistently demonstrated the feasibility of developing highly efficient models for tasks such as classification. The initial batch normalization layer played a crucial role in normalizing inputs, significantly reducing the required training epochs for deep networks. The sequential CNN includes three convolutional layers (Conv) of 3 × 3 size. There are different sizes of Conv filters, the first two layers have 64 filters, and the third has 32 filters. Next to each Conv is a 3 × 3 max pooling layer. The third Conv’s pooling output is fed to the first fully connected layer, which includes 128 neurons. A dropout of 0.35 was applied in the second Conv, the third Conv, and the fully connected layers. We used the Adam optimizer to update weights and perform the cross-entropy loss function. The CNN model was trained using 20 epochs with a batch size of 32. The architecture of the proposed CNN model is shown in Figure 2.

2.3. Multi-Stage Automated Classification (MSTAC)

The MSTAC stacked multiple binary classification models to classify COVID-19, non-COVID-19, and healthy CXR images. The model stacking approach was inspired by the stacked machine learning model used to classify sleep stages automatically [30]. In this study, MSTAC was designed to have two stages. The stage-1 model performs the classification of healthy CXR images from unhealthy ones. The stage-2 model classifies between COVID-19 and non-COVID-19 CXR images. Figure 3. illustrates the block diagram of our proposed experiments and the evaluation of MSTAC. From the block diagram, the CXR images were diagnosed as unhealthy by the stage-1 model. After that, unhealthy images were diagnosed by the stage-2 model to determine whether they were COVID-19 or non-COVID-19. The development and test set were used to develop and evaluate classification model performance. We randomly split the CXR image dataset into training and test sets at an 80:20 ratio. The test set had 630 CXR images, comprising COVID-19 pneumonia, non-COVID-19 pneumonia, and normal.
To elaborate, in the initial step, the stage-1 model successfully classified 630 CXR images into categories of healthy and unhealthy. Upon completion of the stage-1 predictions, the algorithm automatically stored the images predicted as belonging to the healthy class. Subsequently, the stage-2 model took charge of classifying the unhealthy class images to discern whether they were indicative of COVID-19 or non-COVID-19 conditions. The predictions generated by the stage-2 model were then saved utilizing the same algorithm. Finally, we employed the recorded predictions, stored by the algorithm, to conduct a comparative analysis with the true labels of the CXR images. This facilitated the creation of a confusion matrix, serving as a robust tool to evaluate the overall performance of MSTAC.
For the stage-1 model, we used the training set by combining the COVID-19 class with the non-COVID-19 class. Considering the design, this process imbalanced the classes and caused bias in the evaluation. We applied the synthetic minority oversampling technique (SMOTE) to solve the imbalanced classes problem [31]. SMOTE synthesized new data from existing data using k-nearest neighbors (kNN) and inserted them into the original dataset [32]. Originally, there were 840 subjects in the healthy class and 1680 in the unhealthy class. The healthy class became a minority in the training set for the stage-1 model. Consequently, we synthesized 840 healthy images using SMOTE to balance both classes. The stage-1 model was finally performed with 1680 CXR healthy and 1680 CXR unhealthy images. To build the stage-2 model, we eliminated the healthy class from the training set. The stage-2 model was trained using 840 COVID-19 and 840 non-COVID-19 CXR images. Table 3. summarizes the CXR images used for developing the CNN model. The use of SMOTE was intended to address the effects of class imbalance. This ensures that the CNN model of the stage-1 could effectively learn and make accurate predictions for both the healthy and unhealthy classes in the CXR images.

2.4. CNN-Multiclass

In our comparison, we also made a separate deep model named CNN-Multiclass to classify CXR images into COVID-19, non-COVID-19, and healthy classes. The architecture of CNN-Multiclass mirrors that of MSTAC’s CNN model, with a singular modification—expanding the input dimension from 2 to 3 to accommodate the multiclass prediction task. This adjustment ensures that CNN-Multiclass aligns with the structure of MSTAC, enabling us to assess and compare the classification performance of each model for specific health conditions in CXR images.
CNN-Multiclass was trained using 840 CXR images for each of the three categories: COVID-19, non-COVID-19, and healthy, as detailed in Table 4. To maintain consistency in the evaluation process, we employed the same test set used for MSTAC to assess CNN-Multiclass. This test set consisted of 630 CXR images, covering cases of COVID-19, non-COVID-19, and healthy conditions. This test set of images served as the standardized benchmark for evaluating the performance of both CNN-Multiclass and MSTAC, as shown in the diagram in Figure 4.
By using the identical test set, potential biases were minimized. This ensures a fair comparison of how well each model can accurately classify and distinguish among COVID-19, non-COVID-19 pneumonia, and normal CXR images. This approach enhances the reliability of the evaluation results. It provides a robust basis for comparing the effectiveness of the two models in the context of their diagnostic performance on standardized data.

3. Results

In this section, we present the experimental results to demonstrate the ability of the MSTAC model to identify COVID-19 instances from CXR images.

3.1. Performance Evaluation

The evaluation included several performance measures: accuracy (ACC), sensitivity (TPR), precision (PPV), F1-score (F1), and the area under the curve (AUC) [33]. ACC gauges the overall correctness of classifications, indicating the proportion of accurately classified cases among all assessed cases. TPR quantifies the proportion of correctly identified positive cases among all actual positive cases, signaling low false negatives when high. PPV measures the proportion of true positive cases among all cases classified as positive by the system, highlighting low false positives when high. F1 strikes a balance between sensitivity and precision, offering a single metric to assess both true positives and false positives. AUC reflects the system’s overall discriminative power in distinguishing between different classes, particularly crucial for imbalanced datasets.
The performance measure formulae are presented in Equations (1)–(5), involving true positive (TP), true negative (TN), false positive (FP), and false negative (FN). Additionally, we employed the area under the receiver operating characteristic curve (AUC of ROC) to assess performance. The DL algorithms were implemented using Python 3.8.13 with TensorFlow 2.4, CUDA 11.0, cuDNN v8.0.5 on an NVIDIA GeForce-1050 GPU (Nvidia, Santa Clara, CA, USA) with 4 GB GDDR5 onboard memory. The system ran on an Intel Core i3-8100 3.6 GHz processor (Intel, Santa Clara, CA, USA), paired with 8 GB DDR4 RAM.
A C C = T P + T N T P + F P + T N + F N
  T P R = T P T P + F N  
P P V = T P T P + F P  
F 1 = 2 T P 2 T P + F P + F N
  A U C = 1 2 T P T P + F N + T N T N + F P

3.2. Performance Results

Figure 5 shows the training and validation performance of the stage-1 model. The stage-1 model was trained with 20 epochs to classify healthy and unhealthy CXR images. The training accuracy was stable at 99.2% after epoch 15. The training and validation losses were similar, representing a good fit for the stage-1 model, and were not overfitted. In Figure 6, the accuracy plot of the stage-2 model shows a good training and validation trend. Despite a slight difference in the gap between the training and validation curves in both accuracy and loss plots, this indicates that the stage-2 model is effective. It is effective in distinguishing between COVID-19 and non-COVID-19 CXR images. The results of the combined stage-1 and stage-2 models are showcased in MSTAC, as depicted in Figure 3. To assess the performance of MSTAC and CNN-Multiclass, a set of 630 CXR images was utilized. The training and validation results, displayed in the figures, suggest the efficacy of the proposed multi-stage approach in classifying different health conditions in CXR images.
Table 5 compares the outcomes of the three-class classification for CNN-Multiclass and MSTAC. The test set comprised 210 CXR images for each of the three classes that had not been part of the training or testing sets. Initially, MSTAC attained accuracy, precision, recall, f1-score, and AUC values of 97.30%, 97.34%, 97.30%, 97.30%, and 97.98%, respectively. MSTAC showed a three-class classification result that was better than CNN-Multiclass in terms of effectiveness. Figure 7 displays the confusion matrix, showcasing robust sensitivity for COVID-19 and non-COVID-19 pneumonia. In confusion matrix (a), CNN-Multiclass accurately classified 207 out of 210 COVID-19 cases and 208 out of 210 non-COVID-19 cases, achieving sensitivities of 98.57% and 99.05%, respectively. In confusion matrix (b), MSTAC correctly classified 208 out of 210 COVID-19 pneumonia cases and 206 out of 210 non-COVID-19 pneumonia cases, with sensitivities of 99.05% and 98.10%, respectively. However, this suggests that the accuracy in classifying COVID-19 and non-COVID-19 was higher than that for the healthy class. Nevertheless, the high sensitivity of COVID-19 detection can reduce the burden on the healthcare system.
The ROC curves displayed in Figure 8a,b offer a comprehensive overview of the classification performance of our proposed MSTAC and the CNN-Multiclass model. For CNN-Multiclass, the Area Under the Curve (AUC) values for healthy, non-COVID-19, and COVID-19 were 0.930, 0.961, and 0.992, respectively. These values collectively resulted in an overall AUC of 0.961 for the three-class classification. The AUC of MSTAC for healthy, non-COVID-19, and COVID-19 was 0.969, 0.9774, and 0.993, respectively. This represents 0.980 of the classification. Consequently, the findings suggest that our proposed MSTAC outperforms CNN-Multiclass in terms of overall performance, specifically in COVID-19 detection and AUC-ROC.

4. Discussion

This retrospective study represents a significant advance in the development of MSTAC, a model designed for the challenging task of three-class classification—distinguishing between COVID-19 pneumonia, non-COVID-19 pneumonia, and healthy cases through the analysis of CXR images. The utilization of stacked multiple binary models was a detailed process, demanding considerable effort in the reconstruction of the development set and the subsequent training of each model. The proposed CNN exhibited efficiency, converging in 20 epochs, each lasting approximately 18 s. The paramount outcome of our investigation is the noteworthy diagnostic accuracy achieved by MSTAC, surpassing the 97% threshold across the three distinctive classes. In the field of deep learning, numerous studies have investigated a range of preprocessing techniques, feature extraction methodologies, and classification approaches. Notably, customized CNNs and ensemble learning have emerged as pivotal strategies. Given the gravity of potential risks associated with misdiagnoses, our primary objective was to craft a model that excels in accuracy. Table 6 compares the performances achieved in compared studies. With an accuracy exceeding 95%, our model establishes itself as a statistical benchmark. We observed that our proposed approach of stacking multiple binary models achieved better performance than that achieved in studies that used a similar number of CXR image datasets. Comparatively, the performance of MSTAC places it among the top-performing models in the field. It outperforms the accuracy achieved in several notable studies including Wang et al. [21], Khuzani et al. [23], Yang et al. [24], and Nishio et al. [26] in terms of three-class classification. Another strength of MSTAC is its competitive sensitivity of 97.30%, indicating its effectiveness in correctly identifying true positive cases across all three classes. This sensitivity outperforms several existing models, highlighting MSTAC’s robustness in detecting COVID-19, non-COVID-19 pneumonia, and healthy CXR images. Compared to the various transfer learning techniques utilized by Yang et al. [24] and Nishio et al. [26], MSTAC demonstrates notably higher performance. This underscores the effectiveness of our approach in achieving superior results. Additionally, the robustness of the MSTAC model is significantly influenced by the dataset size. The use of a well-balanced dataset, comprising 1050 images for each class and totaling 3150 CXR images, contributes to a more comprehensive and diverse set for both training and evaluation. This stands in clear contrast to studies employing smaller datasets, as shown by Khuzani et al. [23]. It emphasizes MSTAC’s advantage in efficiently addressing diverse features and enhancing its overall effectiveness in detecting COVID-19.
The MSTAC architecture is intentionally designed for optimal performance in detecting COVID-19. The two-stage classification process initially categorizes cases into healthy and unhealthy, followed by a precise classification into COVID-19 and non-COVID-19 categories. The stacking approach integrates the strengths of individual models. This allows for a comprehensive analysis of the three distinct classes within CXR images.
The clinical importance of MSTAC extends beyond its technical strengths. The model’s potential for high accuracy and efficiency promises to significantly reduce misdiagnoses, thereby minimizing risks for patients. The efficient diagnostic workflow streamlined by MSTAC shows potential for integration into current healthcare systems. This is especially important in environments where rapid and accurate COVID-19 detection is crucial.
While MSTAC has demonstrated considerable success, it is crucial to acknowledge a limitation in its accuracy for healthy cases when compared to its robust performance in identifying COVID-19 and non-COVID-19 cases. In the future, we could improve the model by looking at more detailed features, using advanced training methods, and maybe including more data sources to make it better at identifying healthy cases. Our approach utilizes existing resources without the need for extensive processing, providing a straightforward and efficient means of detecting COVID-19 CXR images.
Our distinctive stacking model technique, which has played a crucial role in the success of MSTAC, explores a less-traveled path in the field of CXR image analysis for detecting COVID-19. The simplicity and efficiency of our approach allow it to utilize existing resources without the need for extensive processing. These attributes position MSTAC as a promising system in the realm of diagnostic tools for respiratory conditions.

5. Conclusions

The MSTAC model offers a promising method for accurately and efficiently detecting COVID-19 from CXR images. By using stacked CNN models and a two-stage classification strategy, MSTAC achieves high diagnostic performance. The system showcases impressive accuracy, sensitivity, precision, F1-score, and AUC. This suggests its potential as an effective tool for COVID-19 diagnosis. Our study makes significant contributions, such as developing a three-class classification system that can distinguish COVID-19, non-COVID-19, and healthy cases. Additionally, we validated MSTAC on publicly available datasets to ensure transparency and reproducibility. Comparisons with traditional CNN models underscore MSTAC’s superiority in COVID-19 detection.
In the future, we will implement image data augmentation techniques on CXR images to enhance classification accuracy. Additionally, our current study focused on classifying three categories. However, our future plans involve expanding the dataset and creating a model capable of classifying four classes. This will involve incorporating a new binary classification model for distinguishing between viral and bacterial pneumonia.
Furthermore, our future objectives include developing a classification model that utilizes a pretrained model to identify abnormal regions in CXR images. Lastly, we aim to create a prototype of a web-based CADx system. This system could serve as an alternative, bringing the technique into clinical settings as a valuable tool to support medical staff in improving COVID-19 detection accuracy. At the core of this system is the diagnosis of CXR images uploaded by users, which are classified by MSTAC into healthy, pneumonia, or COVID-19 cases.

Author Contributions

Conceptualization, T.P. and T.W.; methodology, T.P. and T.W.; software, T.P.; validation, T.P. and T.W.; formal analysis, T.P. and T.W.; writing—original draft preparation, T.P. and T.W.; writing—review and editing, T.P., T.W., K.D., A.N., and J.A.; supervision, T.W.; funding acquisition, T.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset utilized in this study is openly accessible to the public. Additionally, all the code developed to support the findings presented in this publication has been made publicly available. The code repository can be accessed at the following URL: https://github.com/tanthanakorn2541/MSTAC_Deep_Learning_COVID19 (accessed on 21 October 2022)”.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

MSTACMulti-Stage Automated Classification.
CXRChest X-Ray.
CNNConvolutional Neural Networks.
DLDeep Learning.
CADxComputer-Aided Diagnosis.

References

  1. Gorbalenya, A.E.; Baker, S.C.; Baric, R.S.; de Groot, R.J.; Drosten, C.; Gulyaeva, A.A.; Haagmans, B.L.; Lauber, C.; Leontovich, A.M.; Neuman, B.W.; et al. Severe Acute Respiratory Syndrome-Related Coronavirus: The Species and Its Viruses—A Statement of the Coronavirus Study Group. bioRxiv 2020. [Google Scholar] [CrossRef]
  2. Singhal, T. A Review of Coronavirus Disease-2019 (COVID-19). Indian J. Pediatr. 2020, 87, 281–286. [Google Scholar] [CrossRef] [PubMed]
  3. Li, Q.; Guan, X.; Wu, P.; Wang, X.; Zhou, L.; Tong, Y.; Ren, R.; Leung, K.S.M.; Lau, E.H.Y.; Wong, J.Y.; et al. Early Transmission Dynamics in Wuhan, China, of Novel Coronavirus–Infected Pneumonia. N. Engl. J. Med. 2020, 382, 1199–1207. [Google Scholar] [CrossRef] [PubMed]
  4. Chen, N.; Zhou, M.; Dong, X.; Qu, J.; Gong, F.; Han, Y.; Qiu, Y.; Wang, J.; Liu, Y.; Wei, Y.; et al. Epidemiological and Clinical Characteristics of 99 Cases of 2019 Novel Coronavirus Pneumonia in Wuhan, China: A Descriptive Study. Lancet 2020, 395, 507–513. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, W.; Xu, Y.; Gao, R.; Lu, R.; Han, K.; Wu, G.; Tan, W. Detection of SARS-CoV-2 in Different Types of Clinical Specimens. JAMA 2020, 323, 1843–1844. [Google Scholar] [CrossRef] [PubMed]
  6. Trouillet-Assant, S.; Viel, S.; Gaymard, A.; Pons, S.; Richard, J.C.; Perret, M.; Villard, M.; Brengel-Pesce, K.; Lina, B.; Mezidi, M.; et al. Type I IFN Immunoprofiling in COVID-19 Patients. J. Allergy Clin. Immunol. 2020, 146, 206–208.e2. [Google Scholar] [CrossRef] [PubMed]
  7. Castellanos-Bermejo, J.E.; Cervantes-Guevara, G.; Cervantes-Pérez, E.; Cervantes-Cardona, G.A.; Ramírez-Ochoa, S.; Fuentes-Orozco, C.; Delgado-Hernández, G.; Tavares-Ortega, J.A.; Gómez-Mejía, E.; Chejfec-Ciociano, J.M.; et al. Diagnostic Efficacy of Chest Computed Tomography with a Dual-Reviewer Approach in Patients Diagnosed with Pneumonia Secondary to Severe Acute Respiratory Syndrome Coronavirus 2. Tomography 2023, 9, 1617–1628. [Google Scholar] [CrossRef]
  8. Lan, L.; Xu, D.; Ye, G.; Xia, C.; Wang, S.; Li, Y.; Xu, H. Positive RT-PCR Test Results in Patients Recovered from COVID-19. JAMA 2020, 323, 1502–1503. [Google Scholar] [CrossRef]
  9. Baratella, E.; Ruaro, B.; Marrocchio, C.; Poillucci, G.; Pigato, C.; Bozzato, A.M.; Salton, F.; Confalonieri, P.; Crimi, F.; Wade, B.; et al. Diagnostic Accuracy of Chest Digital Tomosynthesis in Patients Recovering after COVID-19 Pneumonia. Tomography 2022, 8, 1221–1227. [Google Scholar] [CrossRef]
  10. Radpour, A.; Bahrami-Motlagh, H.; Taaghi, M.T.; Sedaghat, A.; Karimi, M.A.; Hekmatnia, A.; Haghighatkhah, H.R.; Sanei-Taheri, M.; Arab-Ahmadi, M.; Azhideh, A. COVID-19 Evaluation by Low-Dose High Resolution CT Scans Protocol. Acad. Radiol. 2020, 27, 901. [Google Scholar] [CrossRef]
  11. Spogis, J.; Fusco, S.; Hagen, F.; Kaufmann, S.; Malek, N.; Hoffmann, T. Repeated Lung Ultrasound versus Chest X-Ray—Which One Predicts Better Clinical Outcome in COVID-19? Tomography 2023, 9, 706–716. [Google Scholar] [CrossRef]
  12. Yang, J.; Zheng, Y.; Gou, X.; Pu, K.; Chen, Z.; Guo, Q.; Ji, R.; Wang, H.; Wang, Y.; Zhou, Y. Prevalence of Comorbidities in the Novel Wuhan Coronavirus (COVID-19) Infection: A Systematic Review and Meta-Analysis. Int. J. Infect. Dis. 2020, 94, 1016. [Google Scholar] [CrossRef]
  13. Salama, C.; Han, J.; Yau, L.; Reiss, W.G.; Kramer, B.; Neidhart, J.D.; Criner, G.J.; Kaplan-Lewis, E.; Baden, R.; Pandit, L.; et al. Tocilizumab in Patients Hospitalized with COVID-19 Pneumonia. N. Engl. J. Med. 2021, 384, 20–30. [Google Scholar] [CrossRef]
  14. Hwang, E.J.; Kim, K.B.; Kim, J.Y.; Lim, J.K.; Nam, J.G.; Choi, H.; Kim, H.; Yoon, S.H.; Goo, J.M.; Park, C.M. COVID-19 Pneumonia on Chest X-Rays: Performance of a Deep Learning-Based Computer-Aided Detection System. PLoS ONE 2021, 16, e0252440. [Google Scholar] [CrossRef]
  15. Pereira, R.M.; Bertolini, D.; Teixeira, L.O.; Silla, C.N.; Costa, Y.M.G. COVID-19 Identification in Chest X-ray Images on Flat and Hierarchical Classification Scenarios. Comput. Methods Programs Biomed. 2020, 194, 105532. [Google Scholar] [CrossRef]
  16. Kassania, S.H.; Kassanib, P.H.; Wesolowskic, M.J.; Schneidera, K.A.; Detersa, R. Automatic Detection of Coronavirus Disease (COVID-19) in X-Ray and CT Images: A Machine Learning Based Approach. Biocybern. Biomed. Eng. 2021, 41, 867–879. [Google Scholar] [CrossRef]
  17. Shalev-Shwartz, S.; Ben-David, S. Understanding Machine Learning: From Theory to Algorithms; Cambridge University Press: Cambridge, UK, 2013; Volume 9781107057135. [Google Scholar]
  18. Yousri, D.; Abd Elaziz, M.; Abualigah, L.; Oliva, D.; Al-qaness, M.A.A.; Ewees, A.A. COVID-19 X-Ray Images Classification Based on Enhanced Fractional-Order Cuckoo Search Optimizer Using Heavy-Tailed Distributions. Appl. Soft Comput. 2021, 101, 107052. [Google Scholar] [CrossRef] [PubMed]
  19. Chowdhury, M.E.H.; Rahman, T.; Khandakar, A.; Mazhar, R.; Kadir, M.A.; Bin Mahbub, Z.; Islam, K.R.; Khan, M.S.; Iqbal, A.; Emadi, N.A.; et al. Can AI Help in Screening Viral and COVID-19 Pneumonia? IEEE Access 2020, 8, 132665–132676. [Google Scholar] [CrossRef]
  20. Afshar, P.; Heidarian, S.; Naderkhani, F.; Oikonomou, A.; Plataniotis, K.N.; Mohammadi, A. COVID-CAPS: A Capsule Network-Based Framework for Identification of COVID-19 Cases from X-Ray Images. Pattern Recognit. Lett. 2020, 138, 638–643. [Google Scholar] [CrossRef] [PubMed]
  21. Wang, L.; Lin, Z.Q.; Wong, A. COVID-Net: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases from Chest x-Ray Images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef]
  22. Ahmad, F.; Farooq, A.; Ghani, M.U. Deep Ensemble Model for Classification of Novel Coronavirus in Chest X-Ray Images. Comput. Intell. Neurosci. 2021, 2021, 1–17. [Google Scholar] [CrossRef] [PubMed]
  23. Khuzani, A.Z.; Heidari, M. COVID-Classifier: An Automated Machine Learning Model to Assist in the Diagnosis of COVID-19 Infection in Chest X-ray Images. Sci. Rep. 2021, 11, 9887. [Google Scholar] [CrossRef] [PubMed]
  24. Yang, D.; Martinez, C.; Visuña, L.; Khandhar, H. Detection and Analysis of COVID-19 in Medical Images Using Deep Learning Techniques. Sci. Rep. 2021, 11, 19638. [Google Scholar] [CrossRef] [PubMed]
  25. Hong, G.; Chen, X.; Chen, J.; Zhang, M.; Ren, Y.; Zhang, X. A Multi-Scale Gated Multi-Head Attention Depthwise Separable CNN Model for Recognizing COVID-19. Sci. Rep. 2021, 11, 18048. [Google Scholar] [CrossRef] [PubMed]
  26. Nishio, M.; Kobayashi, D.; Nishioka, E.; Matsuo, H.; Urase, Y.; Onoue, K.; Ishikura, R.; Kitamura, Y.; Sakai, E.; Tomita, M.; et al. Deep Learning Model for the Automatic Classification of COVID-19 Pneumonia, Non-COVID-19 Pneumonia, and the Healthy: A Multi-Center Retrospective Study. Sci. Rep. 2022, 12, 8214. [Google Scholar] [CrossRef]
  27. Kermany, D.S.; Zhang, K.; Goldbaum, M.H. Labeled Optical Coherence Tomography (OCT) and Chest X-ray Images for Classification. Available online: https://data.mendeley.com/datasets/rscbjbr9sj/2 (accessed on 12 November 2021).
  28. Alqudah, A.M.; Qazan, S. Augmented COVID-19 X-ray Images Dataset—Mendeley Data. Available online: https://data.mendeley.com/datasets/2fxz4px6d8/4 (accessed on 12 November 2021).
  29. Cohen, J.P.; Morrison, P.; Dao, L.; Roth, K.; Duong, T.Q.; Ghassemi, M. COVID-19 Image Data Collection: Prospective Predictions Are the Future. J. Mach. Learn. Biomed. Imaging 2020, 2020, 2–3. [Google Scholar] [CrossRef]
  30. Wongsirichot, T.; Hanskunatai, A. A Multi-Layer Hybrid Machine Learning Model for Automatic Sleep Stage Classification. Biomed. Eng. Appl. Basis Commun. 2018, 30, 1850041. [Google Scholar] [CrossRef]
  31. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority over-Sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  32. Pan, T.; Zhao, J.; Wu, W.; Yang, J. Learning Imbalanced Datasets Based on SMOTE and Gaussian Distribution. Inf. Sci. 2020, 512, 1214–1233. [Google Scholar] [CrossRef]
  33. Hossin, M.; Sulaiman, M.N. A Review on Evaluation Metrics for Data Classification Evaluations. Int. J. Data Min. Knowl. Manag. Process 2015, 5, 1. [Google Scholar] [CrossRef]
Figure 1. Sample of CXR images: (A) healthy, (B) non-COVID-19, and (C) COVID-19.
Figure 1. Sample of CXR images: (A) healthy, (B) non-COVID-19, and (C) COVID-19.
Tomography 09 00173 g001
Figure 2. Architecture of the CNN model.
Figure 2. Architecture of the CNN model.
Tomography 09 00173 g002
Figure 3. Illustration of MSTAC.
Figure 3. Illustration of MSTAC.
Tomography 09 00173 g003
Figure 4. Illustration of Performance Evaluation: CNN-Multi Class vs. MSTSC.
Figure 4. Illustration of Performance Evaluation: CNN-Multi Class vs. MSTSC.
Tomography 09 00173 g004
Figure 5. Stage-1 model performance for the healthy and unhealthy CXR images.
Figure 5. Stage-1 model performance for the healthy and unhealthy CXR images.
Tomography 09 00173 g005
Figure 6. Stage-2 model performance for the COVID-19 and non-COVID-19 CXR images.
Figure 6. Stage-2 model performance for the COVID-19 and non-COVID-19 CXR images.
Tomography 09 00173 g006
Figure 7. Confusion matrix for the three-class classification (a) Confusion matrix of the MSTAC (b) Confusion matrix of the CNN-Multiclass.
Figure 7. Confusion matrix for the three-class classification (a) Confusion matrix of the MSTAC (b) Confusion matrix of the CNN-Multiclass.
Tomography 09 00173 g007
Figure 8. ROC curves of different classifications: (a) MSTAC; (b) CNN-Multiclass.
Figure 8. ROC curves of different classifications: (a) MSTAC; (b) CNN-Multiclass.
Tomography 09 00173 g008
Table 1. Summary of previous studies on COVID-19 CXR detection.
Table 1. Summary of previous studies on COVID-19 CXR detection.
AuthorMethodsClassesDataset
Chowdhury et al. [19]Different pre-trained CNN model3COVID-19 = 423, Viral pneumonia = 1485, and Normal = 1579
Afshar et al. [20]COVID-CAPS framework consists of 4 convolutional layers and 3 capsule layers4Not specified
Wang et al. [21]COVID-Net network architecture313,975 CXR images
Ahmad et al. [22]Hybrid ensemble model4COVID-19 = 1050, Viral pneumonia = 1050, Bacterial pneumonia = 1050, and Normal = 1050
Khuzani et al. [23]Dimensionality reduction method of CXR images to build classifier3COVID-19 = 140, non-COVID-19 = 140, and Normal = 140
Yang et al. [24]VGG163COVID-19 = 3616, Pneumonia = 1345, and Normal = 3500
Hong et al. [25]MGMADS-CNN model4COVID-19 = 1341, Viral pneumonia = 1733, Bacterial pneumonia = 2705, and Normal = 2916
Nishio et al. [26]Several transfer learning techniques3Three datasets
Our MSTACStacked CNN-based3COVID-19 = 1050, non-COVID-19 = 1050, and Normal = 1050
Table 2. Dataset distribution of CXR images.
Table 2. Dataset distribution of CXR images.
DatasetClassesNumber of Images
CXR Images (Pneumonia) dataset [27]Healthy1050
CXR Images (Pneumonia) dataset [27]Non-COVID-191050
Augmented COVID-19 X-ray Images Dataset [28]COVID-19912
COVID-19 Image Data Collection [29]COVID-19138
Table 3. CXR image distribution used to develop each CNN model.
Table 3. CXR image distribution used to develop each CNN model.
CNN-BasedClassesNo. of Images in Training Set (80%)
OriginalSMOTE
stage-1Healthy8401680
Unhealthy16801680
stage-2COVID-19840-
Non-COVID-19840-
Table 4. CXR image distribution used to develop CNN-Multiclass model.
Table 4. CXR image distribution used to develop CNN-Multiclass model.
ModelClassesNo. of Images in Training Set (80%)
CNN-basedHealthy840
Non-COVID-19840
COVID-19840
Table 5. Classification performance comparison in classifying the three-class CXR images. The best results are shown in bold.
Table 5. Classification performance comparison in classifying the three-class CXR images. The best results are shown in bold.
Three-Classes ClassificationPerformance Metric (%)
AccuracyPrecisionSensitivityF1-Score
CNN-Multiclass94.7695.2294.7694.75
MSTAC97.3097.3497.3097.30
Bold characters indicate higher values.
Table 6. Comparison of classification results of our MSTAC with other COVID-19 DL studies on CXR images.
Table 6. Comparison of classification results of our MSTAC with other COVID-19 DL studies on CXR images.
AuthorMethodsClassesDatasetAccuracy (%)Sensitivity (%)
Chowdhury et al. [19]Different pre-trained CNN model3COVID-19 = 423, Viral pneumonia = 1485, and Normal = 157997.997.9
Afshar et al. [20]COVID-CAPS framework consists of 4 convolutional layers and 3 capsule layers4Not specified95.790
Wang et al. [21]COVID-Net network architecture313,975 CXR images93.393.3
Ahmad et al. [22]Hybrid ensemble model4COVID-19 = 1050, Viral pneumonia = 1050, Bacterial pneumonia = 1050, and Normal = 105094.288.4
Khuzani et al. [23]Dimensionality reduction method of CXR images to build classifier3COVID-19 = 140, non-COVID-19 = 140, and Normal = 14094.05-
Yang et al. [24]VGG163COVID-19 = 3616, Pneumonia = 1345, and Normal = 35009797
Hong et al. [25]MGMADS-CNN model.4COVID-19 = 1341, Viral pneumonia = 1733, Bacterial pneumonia = 2705, and Normal = 291696.7596.60
Nishio et al. [26]Several transfer learning techniques3Three datasets86.76-
MSTAC (our proposed model)Stacked CNN-based3COVID-19 = 1050, non-COVID-19 = 1050, and Normal = 105097.3097.30
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Phumkuea, T.; Wongsirichot, T.; Damkliang, K.; Navasakulpong, A.; Andritsch, J. MSTAC: A Multi-Stage Automated Classification of COVID-19 Chest X-ray Images Using Stacked CNN Models. Tomography 2023, 9, 2233-2246. https://doi.org/10.3390/tomography9060173

AMA Style

Phumkuea T, Wongsirichot T, Damkliang K, Navasakulpong A, Andritsch J. MSTAC: A Multi-Stage Automated Classification of COVID-19 Chest X-ray Images Using Stacked CNN Models. Tomography. 2023; 9(6):2233-2246. https://doi.org/10.3390/tomography9060173

Chicago/Turabian Style

Phumkuea, Thanakorn, Thakerng Wongsirichot, Kasikrit Damkliang, Asma Navasakulpong, and Jarutas Andritsch. 2023. "MSTAC: A Multi-Stage Automated Classification of COVID-19 Chest X-ray Images Using Stacked CNN Models" Tomography 9, no. 6: 2233-2246. https://doi.org/10.3390/tomography9060173

APA Style

Phumkuea, T., Wongsirichot, T., Damkliang, K., Navasakulpong, A., & Andritsch, J. (2023). MSTAC: A Multi-Stage Automated Classification of COVID-19 Chest X-ray Images Using Stacked CNN Models. Tomography, 9(6), 2233-2246. https://doi.org/10.3390/tomography9060173

Article Metrics

Back to TopTop