Next Article in Journal
Clinical Risk Factors for Dysphagia and Esophageal Dysmotility in Systemic Sclerosis
Previous Article in Journal
Burden and Risk Factors of Postpartum Depression in Southwest Saudi Arabia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Deep Learning Methods for Interpretation of Pulmonary CT and X-ray Images in Patients with COVID-19-Related Lung Involvement: A Systematic Review

1
School of Engineering and Digital Sciences, Nazarbayev University, Kabanbay Batyr Ave. 53, Astana 010000, Kazakhstan
2
School of Medicine, Nazarbayev University, 5/1 Kerey and Zhanibek Khandar Str., Astana 010000, Kazakhstan
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2023, 12(10), 3446; https://doi.org/10.3390/jcm12103446
Submission received: 24 February 2023 / Revised: 25 April 2023 / Accepted: 6 May 2023 / Published: 13 May 2023
(This article belongs to the Section Nuclear Medicine & Radiology)

Abstract

:
SARS-CoV-2 is a novel virus that has been affecting the global population by spreading rapidly and causing severe complications, which require prompt and elaborate emergency treatment. Automatic tools to diagnose COVID-19 could potentially be an important and useful aid. Radiologists and clinicians could potentially rely on interpretable AI technologies to address the diagnosis and monitoring of COVID-19 patients. This paper aims to provide a comprehensive analysis of the state-of-the-art deep learning techniques for COVID-19 classification. The previous studies are methodically evaluated, and a summary of the proposed convolutional neural network (CNN)-based classification approaches is presented. The reviewed papers have presented a variety of CNN models and architectures that were developed to provide an accurate and quick automatic tool to diagnose the COVID-19 virus based on presented CT scan or X-ray images. In this systematic review, we focused on the critical components of the deep learning approach, such as network architecture, model complexity, parameter optimization, explainability, and dataset/code availability. The literature search yielded a large number of studies over the past period of the virus spread, and we summarized their past efforts. State-of-the-art CNN architectures, with their strengths and weaknesses, are discussed with respect to diverse technical and clinical evaluation metrics to safely implement current AI studies in medical practice.

1. Introduction

The novel coronavirus (SARS-CoV-2) emerged in early December 2019 and has since become a global pandemic threatening humanity [1]. COVID-19, the disease caused by this virus, presents with non-specific symptoms such as cough, fever, myalgia, headache, gastrointestinal dysfunction, and other flu-like symptoms, thereby making it challenging to differentiate from other viral upper respiratory diseases, especially in the early stages [2]. However, the timely diagnosis and management of affected patients could reduce mortality and the spread of COVID-19 [1,2].
The standard diagnostic tests used to confirm COVID-19 diagnosis are real-time RT-PCR and the reverse transcription polymerase chain reaction (RT-PCR) [3,4]. Chest-computed tomography (CCT) is a radiological diagnostic tool used to detect the pulmonary manifestations and complications of COVID-19, such as pneumonia and acute respiratory distress syndrome. The main radiological features of COVID-19 are asymmetric peripheral ground-glass opacities (GGOs) in the absence of pleural effusions [5]. However, given the high volume of cases that radiologists need to analyze, the manual interpretation of chest CT scans can be monotonous, tedious, and prone to diagnostic errors. This is where automated artificial intelligence tools can come in handy by facilitating and double-checking physicians’ tasks, as well as reducing the likelihood of diagnostic errors.
Recent advances in artificial intelligence, especially deep learning (DL) algorithms, have shown great potential in accurately interpreting medical images, including chest CT scans [6]. CNN, one of the DL algorithms, can learn the most important representations through layer-by-layer feature analysis [7] and has been successfully adapted to analyze chest CT images for COVID-19 detection [8,9].
Although the characteristics of the current CNN models have been widely tested and described in the literature, there is no systematic review and meta-analysis focusing on the application of CNN in the detection of pulmonary involvement in COVID-19 patients. Therefore, the aim of this systematic review is to study the performance, benefits, and risks of CNN for the detection of pulmonary manifestations in COVID-19 patients.
One of the major challenges in developing CNN models for COVID-19 detection is the limited size of medical datasets, especially in COVID-19 cases. To address this issue, advanced augmentation approaches based on conventional generative adversarial networks (GANs) have been successfully applied to acquire a sufficient number of clinical data samples. Recently, publicly available datasets have been created by many research foundations, and CNN architecture with data-invariant performance has become one of the most challenging issues in clinical CNN study.
Another challenge is constructing a parameter-invariant CNN framework that assures robust performance on the variations of medical imaging data. Techniques such as semi-automated image processing software or CNN-based segmentation have been developed to overcome this issue by extracting the region of interest. More recent end-to-end CNN architectures have shown comparable or even superior performance to conventional pipeline systems.
For the performance evaluation, we utilized popular CNN architectures designed primarily for the computer vision field as our baseline models. These models included VGG, Inception, DenseNet, GoogLeNet, ResNet, and more extensive versions of the individual networks. Diverse state-of-the-art CNN models have been proposed, which vary in terms of model structures such as activation functions, filter/kernel size, model parameters, and depth of the network. These models aim to improve the representational ability of feature maps to overcome known problems such as overfitting, vanishing gradient, computational cost, etc. Transfer learning, ensemble metal-classifiers, and fine-tuning based on pre-trained CNN models have been successfully implemented in clinical imaging data. However, the lack of a common framework, as well as a benchmark dataset, make it challenging to provide a reliable comparison of the state-of-art methodologies.
Meanwhile, ensuring the reliability of the network’s decision making through quantitative evaluation is critical in clinical studies. Therefore, interpretable AI, which can visually represent how and why a CNN makes certain decisions at a human-understandable level, is an essential part of clinical applications. Most recent studies have implemented a class activation map (CAM) to provide visual explanations of CNN output by representing a heatmap that can highlight the decision-relevant regions on the input image.
Our study aims to systematically review published studies that applied deep learning approaches for the diagnosis and prognosis of COVID-19 based on CT and X-ray images. The main contributions of our systematic review are as follows:
We analyzed publicly available CT and X-ray imaging datasets for COVID-19 cases and evaluated their overall performance with conventional and state-of-the-art CNN approaches. We provided a detailed description of CNN architecture in terms of the structural network design and its pros and cons. We examined visualization models that can quantitatively evaluate CNN model decisions. We identified further challenging issues based on our findings of the review to be importantly addressed in the medical AI community to advance the field.
This review is structured as follows. In Section 2, we present the methodology used to carry out the review protocol, search strategy, and data extraction. In Section 3, we provide a summarization of the results in terms of the research perspectives identified in the previous section. Section 4 discusses the limitations of this literature review and suggests future research directions. We believe that this review will aid clinical practice by informing future research and development about improved diagnostic and treatment techniques for patients with COVID-19.

2. Methodology

2.1. Protocol and Literature Search

The scope of this review included articles reporting the use of convolutional neural network (CNN) methods for chest-computed tomography and X-ray imaging interpretation in COVID-19 patients. We conducted this scoping review in accordance with the PRISMA guidelines [10]: Patient population—critical/intensive care patients; Intervention/diagnostic tools—deep learning models; Comparison—non-deep learning prediction models; Outcomes—prediction of mortality (in ICU); Secondary—the value of deep learning methods in decision making, characteristics of deep learning models, and resources, including data and code availability, AI mode/approach, DL structure, transfer learning, end-to-end learning, explainability, pulmonary involvement (injury) measurement, pulmonary involvement classification, deep learning architectures used in biomedical image analysis, image processing techniques, and the opinions of certified radiologists.
The main goals of this systematic review are: (1) to analyze the original studies and publicly available datasets reporting the application of deep learning in the analysis of COVID-19-related lung involvement, (2) to provide a comprehensive analysis of detection of COVID-19-related pneumonia and segmentation methods based on deep learning, (3) to discuss advantages and disadvantages of the methods and DL methods used (4) to discuss recommendations and future challenges based on the results of the systematic review, and (5) to summarize the limitations of the study and the significance of the application of the deep learning.
Inclusion and exclusion criteria were as follows:
  • The included studies must have assessed the diagnostic or prognostic potential using deep learning algorithms in COVID-19 patients with pulmonary manifestations.
  • Only original studies were included in the systematic review.
  • Abstracts, case reports, case series, invited reviews, narrative and systematic reviews, meta-analyses, animal studies, editorials, letters to the editors, conference papers, commentaries, comparative studies, and expert views were excluded.
  • Articles in non-English language were excluded.
  • Studies that examined non-deep learning applications for diagnosis of pulmonary manifestations in COVID-19 patients were also excluded;
  • The review study excluded papers that did not present sufficient information on related classification performance metrics.
  • Finally, we also excluded studies focusing on non-radiological methods of diagnosis of pulmonary manifestations in COVID-19 patients, even if deep learning methods were used.
The literature search resulted in 1116 citations. Following the removal of duplicates and screening of titles with abstracts, as well as reading the full articles, we analyzed 74 articles in the final version of our systematic review (as depicted in Figure 1 and Figure 2).
The screening of the potential studies was conducted by three reviewers (DV, AS, and MHL). After removal of duplicated titles, we screened the titles and abstracts. The data extraction process was conducted by two reviewers (AS and ML). The discrepancies were resolved by consensus.
We performed a literature search in the following databases: PubMed, EMBASE, and Google Scholar. We used the following free text and medical subject headings (MeSH) terms: ‘lung’, ‘respiratory system’, ‘pneumonia’, ‘respiratory complications’, ‘classification’, ‘artificial intelligence’, and ‘tomography’. The terms were combined with the following words: ‘convolutional neural network’, ‘CNN’, ‘deep learning’, ‘COVID-19’, ‘pneumonia’, ‘detection’, and ‘diagnosis’. We considered all original studies that have been published from inception to May 2022. Articles that did not meet the review criteria were excluded. We examined and cross-referenced bibliographies to identify additional papers of relevance.
In our systematic review, we covered a wide range of attributes related to CT/CXR datasets and DNN approaches. The DNN approaches can be categorized into two directions: image segmentation and classification. Segmentation of chest CXR/CT images involves the automatic localization of boundaries that limit the area of pulmonary involvement (pathological part of the lung), which is then separated from the intact part of the image for further analysis. Image classification entails the ability to capture pixel-level features that may not be easily detected by human eyes. This includes feature representation, pixel-level heatmap visualization, and validation of decoding accuracy.

2.2. Specification of Public Dataset

The growth of the deep learning approach has been further facilitated by the creation of large, annotated open datasets. In our review, we focused on deep learning classification models for COVID-19 based on available CT and X-ray scan datasets. We conducted a literature review of papers that used both publicly available open datasets and private datasets. From dataset descriptions, we extracted information on their origin, data types, sample size, image resolutions, populations, available links, and the highest performance achieved in the current research with their CNN approach. Patient scans and other materials included in datasets were typically collected from various sources, such as public domains, hospitals, and physicians, with the collection and distribution of patient information regulated by the ethics review board in each of the reviewed studies. Authors in the reviewed studies reported the use of annotated datasets in which each slice of the CT or X-ray scan was labeled with various classes, including binary classification of COVID-19 pneumonia/normal, COVID-19 pneumonia/bacterial pneumonia, as well as studies reporting multi-class classifications, including three cases of COVID-19, normal, and bacterial pneumonia. For studies reporting the use of public datasets for COVID-19 classification, information on RT-CPR confirmed COVID-19, or radiologist-verified inclusion criteria were checked.
The sample size of datasets varied significantly across different studies. Generally, models fit on smaller datasets showed poor generalization compared to large datasets. Nevertheless, some models were able to achieve high accuracy even on small/moderate-sized datasets.

2.3. Performance Evaluation and Baseline Models

Performance evaluation metrics, such as accuracy, sensitivity, specificity, positive predicted value (PPV), negative predicted value (NPV), the area under the curve (AUC), and F-score were extracted. Individual papers presented several baseline models for comparing the performance with the actual proposed model. For review purposes, these evaluation results were recorded for the best-performed baseline models. In some studies, authors only reported a subset of the presented metrics. For some of the studies that reported the performance metrics, we were able to recover the missing value of the metric with the aid of the reported confusion matrix. In the reviewed papers, we additionally gathered information on the evaluation strategy. The evaluation strategy included the subdivision of the dataset into training, validation, and testing subsets. We looked at the percentage proportion of the dataset splits. Another approach that we focused on during the review was selecting information about the cross-validation approach to evaluate the performance. These approaches most commonly included 5-fold and 10-fold cross-validation procedures.
The actual review study excluded the papers in which an insufficient amount of information on related classification performance metrics was presented. Among other factors, we have not included, in the review, studies that only presented segmentation results in their studies.

2.4. CNN Architecture

Our systematic review covers various aspects related to neural network structures, including the CNN model, end-to-end learning, explainability, and transfer learning. We examined the features of different CNN architectures, such as layerwise components, parameter size, preprocessing algorithms, optimization techniques, and so on. Feature representation methods helped us to gain insight into the manifold space by transforming the non-linear and high-dimensional feature space into a subspace dimension where we could visually observe the feature distribution for each class. In this context, explainable models were essential to consider, as they allowed us to intuitively understand the CNN decision by highlighting the contribution score on the input medical image at the pixel level.
Other crucial aspects of our survey were transfer learning and fine-tuning approaches. Medical image data have high resolutions, but small sample sizes compared to the computer vision domain, which could lead to overfitting or generalization problems. Transfer learning has been reported to have a significant impact on knowledge transfer across domains, as it allows for the sharing of equivalent convolution filters and the optimal setting of initial parameters in the network, thus resulting in superior performance.
Therefore, our study includes relevant information regarding model explainability and knowledge transfer perspective as well. The results of the included studies are summarized in Table 1 (see Figure 1). Based on the reviewed papers, we compiled a comprehensive table that summarizes the key details of the most notable approaches published before 2022. The table includes the following information, which we believe represents the main contributions of the reviewed studies:
  • CNN architecture, i.e., the main underlying baseline model.
  • Imaging modality, i.e., CT scan or X-ray.
  • The prediction classes, i.e., COVID-19, normal, and pneumonia.
  • Pre-processing steps, i.e., data augmentation, image processing methods, feature extraction, and image segmentation.
  • Explainability, i.e., the activation heatmaps and grad-CAM visualization.
  • Availability of code.
  • Performance evaluation metrics, including accuracy, sensitivity, specificity, F-score, AUC, and others.
  • Use of transfer learning techniques.
  • Specific details of the datasets used.
The reviewed articles included a wide range of network architectures, with specific descriptions of models and their fine-tuning parameters. The most widely used models were ResNet-50, DenseNet-161, VGG-16, ConvLSTM, 3D ResNet-50, InceptionV3, and DenseNet-201. These models were presented as backbone models or as modified architectures with transformations to the model architecture or to the high-level optimization hyperparameters.
Deep learning is often considered a black box, because it is difficult to understand how a neural network arrives at a certain classification choice. To gain insight into the underlying features that contribute to the decision-making of neural networks, the authors of the reviewed papers analyzed the output of the neural networks using visualization methods that mapped regions of the input images. These methods allowed the prediction score of each pixel in the image to be depicted, thus showing how it contributed to the classification decision. The authors relied on techniques such as grad-CAM, AM, RISE, and OS.

3. Results

3.1. Definition of the Classification Target Classes

The presented models exhibited variability in target class definitions. The majority of publications used two or three target classes, with COVID-19 (C) and non-COVID-19 (N) classes being the most commonly used. Some studies opted for four classes, thereby aiming to distinguish between smaller subcategories of diseases. For instance, in their work [38], the authors used a transfer learning approach based on the ResNet architecture to classify chest X-ray images into four categories, namely, COVID-19 (C), non-COVID-19 (N), bacterial pneumonia (BP), and viral pneumonia (VP). Some studies focused on pneumonia cases and classified scan images into COVID-19 pneumonia (CP) versus non-COVID-19 pneumonia (nCP).

3.2. Overall Performance

The main tasks used to train CNN models included detecting, classifying, and segmenting chest CT and chest X-ray images of COVID-19 patients. Most studies defined the classification problem as binary decoding, i.e., COVID-19 (C) vs. non-COVID-19 (N), and also used more specific disease definitions such as bacterial pneumonia (BP), viral pneumonia (VP), COVID-19 pneumonia (CP), non-COVID-19 pneumonia (nCP), COVID-19 stage (ML), medium COVID-19 stage (MD), severe stage of COVID-19 (SC), and non-COVID-19 stage (N).
In our review, the evaluation factor performances ranged from 0.57 to 1.0 for sensitivity, 0.58 to 0.99 for specificity, 0.55 to 0.99 for positive predictive value (PPV), 0.67 to 1.0 for area under the curve (AUC), and 0.7 to 1.0 for the F-score. There was inconsistency in using the assessment phases, with most authors reporting using the training phase followed by validation and testing. Others reported training and validation or training and testing only. The proportion and volume of data used for each of these phases also varied widely.
Several authors reported high performance. For instance, Serte et al. [59] developed a ResNet-50 deep learning model to detect COVID-19 in 3D chest CT scan images and compared it with other deep learning models. The proposed ResNet-50 outperformed all other models with a sensitivity of 100%, specificity of 0.98% (0.95), and AUC value of 96% for detecting COVID-19. Sharifrazi et al. [60] proposed a COVID-19 detection model using X-ray images that were fed to a CNN deep learning model followed by ten-fold cross-validation by an SVM classifier. The proposed CNN-SVM model with a Sobel filter (CNN-SVM + Sobel) achieved the highest classification sensitivity of 100%, specificity of 95.23%, and accuracy of 99.02%.

3.3. Evaluation of DNN Architectures

Interestingly, only 20% of CNN studies implemented an end-to-end learning framework, with many studies instead suggesting the use of diverse machine learning techniques on raw imaging data, such as data augmentation (DA), feature extraction (FE), segmentation (SG), or image processing (IP). Table 1 displays reported CNN architectures that have demonstrated superior performance on certain datasets, including CNN and its various modifications (Inception V3, Xception, ResNeXt, CFRCF, 2D-CNN, 3D-CNN, ResNet-50, VGG-16, FBSED, EfficientNetB0-3, Inception_resnet_v2, Xception convolutional auto-encoder neural network, DenseCapsNet (DenseNet + CapsNet), MWSR, CTnet-10, AlexNet, CTnet-10, and CovidAid_V2).
In general, the authors reported improved classification accuracy by combining models with advanced pre-processing methods, using an ensemble of classifiers, or employing advanced fine-tuning to the model architecture. We can categorize the contributions that led to superior performance in the reviewed studies into the following types: (1) Preprocessing and data augmentation, (2) Transfer learning and fine-tuning, (3) Ensemble learning, etc.
Several studies combined two or more approaches in their works. For instance, in [61], the authors used an ensemble of 15 pre-trained models (EfficientNets(B0-B5), NasNetLarge, NasNetMobile, InceptionV3, ResNet-50, SeResnet 50, Xception, DenseNet121, ResNext50, and Inception_resnet_v2) in conjunction with data augmentation (DA) for the binary classification task of COVID-19 on a relatively small dataset.

3.4. Ensemble Learning

The purpose of an ensemble system is to correct errors made by other classifiers within the system, and, thus, the diversity of the classifiers is crucial [62]. If all classifiers give the same output, the ensemble system would provide no additional benefit. Therefore, it is essential that the individual classifiers within an ensemble system make different errors in different instances to reduce the total error by accounting for the different errors of each classifier in the ensemble.
Several studies suggested and applied the ensemble learning approach to improve classification accuracy [61,63]. This approach involves combining multiple distinct CNN models to provide more reliable results. The results reported by these studies indicate that the ensemble approach is highly successful in solving the COVID-19 classification problem. One example of the success of the ensemble approach is that it can learn complex non-linear boundaries by combining the classification results of several linear models. For instance, one study used a stacked ensemble of CNN classifiers consisting of ResNet50, ResNet101, VGG16, VGG19, Xception, MobileNetV1, MobileNetV2, DenseNet121, and DenseNet169 models and achieved 99.31% accuracy. In another study, the authors utilized a different ensemble approach that calculated the fuzzy membership values of InceptionV3, DenseNet121, and VGG19 classifiers.

3.5. Transfer Learning and Data Augmentation

There have been numerous DNN models used for COVID-19 detection, including pre-trained and novel architectures. The former and the latter are the two most common approaches for addressing the issue. When pre-trained features need to be preserved, transfer learning is used by freezing the convolution part (in the case of a CNN) and unfreezing the dense or classification part of the architecture. Several well-known CNN architectures, such as ResNet, Inception, VGG, AlexNet, DenseNet, ConvNet, DarkNet, 3D-CNN, and LSTM, have been adapted using transfer learning for COVID-19 detection (see Table 1) [64]. On the other hand, fine-tuning is used by freezing the classification part of the architecture and unfreezing the convolution part of the model. Models such as AlexNet, GoogleNet, and SqueezeNet were fine-tuned for COVID-19 detection [65].
The second most common approach is creating a new model specifically for COVID-19 recognition. Models such as CSEn, PF-BAT FKNN, NASNet, DeepSense, DeCoVNet, and ConvLSTM were proposed for the problem. Additionally, popular architectures such as Xception with auto-encoder, DenseCapsNet (DenseNet + CapsNet), DWS-CNN+Deep SVM, VGG-16+nCOV-NET, DenseNet-161+nCOV, ResNet-50+nCOV-NET, UNet+SCOAT, 27-layer 3D-CNN, MobileNetV2+FKNN, and MobileNetV2+FKNN were modified for COVID-19 detection (see Table 1). However, according to Pham [65], the fine-tuned models, such as AlexNet, GoogleNet, and SqueezeNet, outperformed new models proposed for COVID-19 detection, such as CoroNet, CovidGAN, and DarkCovidNet.
Many studies employed additional pre-processing steps before feeding data into deep CNN models, which typically provided better classification results. Data augmentation (DA) methods were used in nearly half of the surveyed studies, with most reporting high accuracy scores after their application. For example, authors [14] reported a 99% accuracy for the ternary classification task (non-COVID-19, COVID-19, pneumonia) using ResNet-V2 architecture by applying a set of augmentations based on the random segmentation of original scans to 480 × 480 pixels, followed by random horizontal flips and normalization. Additionally, many studies employed different approaches for lung segmentation, such as the “snake” technique used by Saad et al. [66], which involves segmenting the lung section to detach the pixels of interest from noise and other unnecessary areas.

3.6. Explainability of CNN Model

To add transparency to the proposed models, the majority of reviewed papers implemented visualization techniques. The gradient-weighted class activation mapping (grad-CAM) [67] approach has been the most commonly used method for model explainability. This approach utilizes the gradient information received by the last convolutional layer in a CNN to comprehend the importance of each neuron during the decision-making process [67].
In [12], the authors utilized the occlusion sensitivity (OS) method for visualization purposes. According to the authors, this method provides a clear illustration of the lung regions that are the most crucial for COVID-19 prediction. They concluded that the occlusion sensitivity visualization method they used provided evidence of significant lung regions clearly within the lungs, thereby avoiding sensitivity on corners or edges.
In [13], studies showed the effectiveness of RISE visualization when applied to X-ray images. The results indicated that the RISE technique successfully emphasizes the changes in lung X-ray images caused by COVID-19 infection.
Multiple studies in this survey paper used the activation map visualization technique. In [15], the authors used a spatial activation map which, as they suggested, successfully identified key areas relevant to infected regions in their proposed DA-CMIL model.
In [23], activation maps were also used to illustrate the infected lung regions. To demonstrate the effectiveness of the method, they compared the images obtained using an activation map with the RALE score of those X-ray images. The RALE score represents the severity of the disease. The authors concluded that the visual representation of infected regions obtained by the activation map matched perfectly with the RALE score that had been calculated.
Several papers reviewed did not provide a detailed description of the specific explainable model, and, thus, their approaches were marked as a heat map (HM). In 46% of the papers, the visualization results were provided by highlighting the pixel-level confidence score on the 2D-CXR or 3D-CT image. The heat maps in these studies showed convincing results, as the intra-zone and middle-zone of the pulmonary region have the greatest influence on the decision process.

3.7. Performance Enhancement: Novel CNN Strategy

The heterogeneity of the dataset plays a crucial role in the classification performance of state-of-the-art DNN models, as evidenced by the CC-CCII dataset [68]. The highest accuracy achieved on this dataset was 93% by a certain CNN model [41], which is good, but not the best result, for the same task compared to DNN models on other large datasets. Our survey suggests that the differences in classifier performances are related to model structure and resource availability.
To overcome the limitations of unimodal CNN architecture and achieve general performance on diverse datasets, ensemble learning has been proposed. Ensemble learning combines multiple distinct CNN models to represent complex non-linear features and robust decision boundaries. The goal is to reduce total error by strategically combining the model outputs, which is similar to the low pass filtering of noise. For example, one study proposed a stacked ensemble of CNN classifiers consisting of ResNet50, VGG16, Xception, MobileNetV1, and DenseNet169, while other studies used a different ensemble approach by calculating the fuzzy membership values from InceptionV3, DenseNet121, and VGG19 models. Ensemble learning approaches have achieved superior performance compared to a sole CNN model.
Recent studies have proposed various CNN approaches to boost performance by combining several classification strategies and machine learning techniques. For instance, [69] suggested transfer learning and parameter optimization to simultaneously classify X-ray and CT images in a hierarchical architecture, and [61] proposed transferring an ensemble of 15 pre-trained models with data augmentation techniques.
Gupta et al. [70] claimed the effectiveness of CNN models and proposed COVID-WideNet, which is a capsule network with 20 times fewer trainable parameters that are computationally less expensive while preserving model performance. Feature optimization, channel boosting, and recurrent units such as RNN or LSTM have also reported enhanced performance in studies by [71] (Advanced Squirrel Search Optimization Algorithm), [72] (CB-STM-RENet), and [73] (Gated Recurrent Unit).

3.8. Open COVID19 Dataset

Radiological images such as chest X-ray (CXR) and chest computed tomography (CT) are often captured at different resolutions, with some of the most common being 512 × 512, 580 × 335, 333 × 308, 299 × 299, 1024 × 1024, 333 × 308, 1766 × 1349, 224 × 224, and 1857 × 1317. These images were obtained from various sources, including intrinsic hospital data sets such as the Wuhan Pulmonary Hospital (China) [37], Honghu and Nanchang hospitals (China) [48], National China Hospitals [68], municipal hospitals in Moscow (Russia) [74], Sao Paulo hospitals (Brazil) [75], the Metropolitan Hospital of Lapa (Brazil) [76], the Indiana University Hospital (USA) [77], and radiology hospitals (Italy) [78].
Additionally, the study utilized open datasets such as the “China Consortium of Chest CT Image Investigation (CC-CCII) Dataset” [68], MosMedData [74], COVID-CT [79], SARS-COV-2 CT Scan [75], Harvard Dataverse [76], LIDC-IDRI [80], iCTCF [81], three versions of COVIDx-CT [47,82,83], COVID-19-CXR [84], COVID-19 RD [85], ChestX-ray8 [86], CXR (COVID-19 and pneumonia) [84,87,88], CT [89], RSNA PDC [86], COVID-19 IDC [88], CDGC [90], and SIRM [78]. Population information, gender ratio, sample size, classes, resolution, and links to public datasets were extracted from the descriptions of these datasets, as are summarized in Table 2.
The majority of datasets and codes are readily accessible and available (see Table 1 and Table 2). Among the surveyed datasets of CT and X-ray scans, the largest in terms of sample size and patient populations were CC-CCII (with 411529 CT slices), CT (with 120968 chest X-ray images), iCTCF (with 256356 CT slices), LIDC-IDRI (with 244527 CT slices), COVIDNet-CT (with 194922 CT slices), COVIDx-CT (with 201103 CT slices), and ChestX-ray8 (with 112120 chest X-ray images).
A certain proportion of these datasets included a combined collection of images from several open-source COVID-19-related or other pulmonary conditions datasets. For example, the CT dataset consisted of multi-source chest X-ray data from the BIMCV, COVID-19 RD, and RSNA PDC datasets. Other large-scale dataset collections were mostly a product of multi-hospital efforts to provide an appropriate and urgent solution to analyze the rapidly growing scan collections within a certain country or region. For instance, the CC-CCII (China Consortium of Chest CT Image Investigation) dataset included patient CT scans from CC-Sun Yat-sen Memorial Hospital, the Third Affiliated Hospital of Sun Yat-sen University, the First Affiliated Hospital of Anhui Medical University, the West China Hospital, the Nanjing Renmin Hospital, the Yichang Central People’s Hospital, and the Renmin Hospital of Wuhan University in China.
New variants of the SARS-CoV-2 virus have emerged in addition to the original strain causing COVID-19. The most prevalent ones are the delta variant, which emerged in late 2020, and the omicron variant, which emerged in late 2022. While all currently available datasets provide CT and CXR images of typical COVID-19 disease, there is a dearth of datasets featuring its variants. In one study [97], an open-access dataset featuring two classes of data—CT scans [98] and X-rays [99]—of the delta and omicron variants was used. For the initial phase of their framework, the researchers used an X-ray database obtained from Kaggle [100,101], along with a limited local database to test it. They also collected a comprehensive database of CT scan images from the radiology centers of Tehran University Hospitals to train and test the model’s second phase, thereby making it entirely native. The definitive status of the cases in this dataset was determined after conducting PCR tests. Another dataset featuring CT scan images of children with delta variant cases is available upon contacting the corresponding author of the paper [102], but it is not open-access. In their study, the authors analyzed delta variant cases in children without using CNN classification.

3.9. Review of the Validity and Applicability of the DL Models

The majority of studies across all specialties were deemed high risk according to the ’Quality Assessment of Diagnostic Accuracies Studies 2’ (QUADAS-2) due to significant deficiencies in patient selection, flow, and timing. The PROBAST assessment tool [103] also revealed that all included studies had a high risk of bias, thereby indicating that the model’s performance in practice may be lower than reported. Most studies did not provide specific details about patients and interventions, thus resulting in a high-risk rating for the participant domain.

4. Discussion and Conclusions

From the original articles analyzed in this systematic review, automated AI-aided radiological image interpretation has become increasingly useful in highly active and low-resource clinical settings. The reviewed models demonstrated high performance and reliability in detecting pulmonary manifestations of COVID-19 and their differentiations from non-COVID-19 pneumonia.
This study aimed to (1) estimate the diagnostic or/and predictive performance of DL algorithms to identify distinct radiological features of the pulmonary manifestation of COVID-19 using chest CT and chest X-ray images; and to (2) review the variation in the study reporting DL in radiological diagnosis in the published studies. We found that DL algorithms demonstrate a high diagnostic and predictive performance and are acceptable to be used in clinical settings. The high diagnostic and predictive accuracy of DL approaches was identified in all articles, thus suggesting that DL algorithms can be deployed and used for assistance in overloaded clinical settings.
DL models can provide valuable support to doctors in making an accurate diagnosis and facilitate heavy workloads, especially when the healthcare system is overloaded or in resource-constrained regions with a shortage of radiologists [104]. However, the diagnostic accuracy of DL is not significantly higher compared with experienced radiologists. The time spent per CT image interpretation and description was 10 s per scan, while experienced radiologists spent about 10 min for the same task [104].
One study showed that it was feasible to rapidly develop reliable lung segmentation for COVID-19 using deep CNN. Satisfactory results were achieved with less than 50 cases used for training. Deep CNN was successful in developing the fully automated quantification of pulmonary manifestations of COVID-19. The authors trained the deep CNN-based segmentation algorithm and implemented a threshold-based quantification assessing lung opacity load into the clinical practice [105].
One research group tested the performance of 17 neural networks including AlexNet, ResNet-50, DarkNet-53, DarkNet-19, SqueezNet, GoogleNet, Place365-GoogLeNet, MobileNet-v2, ShuffleNet, NasNet-Mobile, Xception, Inception-ResNet-v28, Inception-v3, DenseNet-201, ResNet-18, VGG-19, and ResNet-101. The authors reported that DarkNet-19 outperformed other neural networks in the interpretation of chest X-ray imaging of COVID-19 patients. DarkNet-19 achieved an accuracy value of 94.28% on 5854 X-ray images [106].
The next research group developed a multiple-instance learning method (based on a deep learning method) to accurately predict the disease severity of COVID-19 using quantitative CT data. This model was efficient in identifying patients at high risk for progression in the early phase of COVID-19, which was useful in preventing disease progression and decreasing mortality. The authors recommended that COVID-19 patients undergo CT screening after admittance to the hospital to identify patients at high risk before disease progression [48].
COVIDNet-CT is a deep CNN architecture customized for the detection of pulmonary injury in COVID-19 patients using chest CT images through machine-driven design. The authors created COVIDx-CT, a CT image dataset containing 104,009 images from 1489 patients [47]. Nevertheless, because of high heterogeneity across studies, there was substantial uncertainty of the diagnostic accuracy estimates.
Although CT scanning has shown high performance in the diagnosis of COVID-19, chest X-ray also has many benefits. Thus, it is an easily available approach that can assist radiologists in emergency clinical settings and those widely used worldwide throughout the pandemic. The model has been reported to diagnose COVID-19 for several seconds. Despite the high performance of CT, it is expensive and not widely accessible in low-resource settings. Moreover, the dose of radiation is higher in CT.
Chest X-rays can enable the rapid triaging of patients with pulmonary manifestation and can be performed simultaneously with other laboratory tests to assist medical staff in identifying COVID-19 patients among the large numbers of patients [107]. Existing evidence showed that the chest X-ray plays a fundamental role in the diagnosis of COVID-19. Although the radiological features of pulmonary manifestations in COVID-19 patients and non-COVID-19 pneumonia may have some similarities, key radiological patterns allow clinicians to differentiate between these two diseases [107].
Chest X-rays patterns in COVID-19 patients include (1) ground-glass opacities (usually bilateral, multifocal, peripheral subpleural, medial, basal, and posterior location), (2) a crazy paving appearance (ground-glass opacities with inter- or/and intra-lobular septal thickening), (3) traction bronchiectasis, (4) air space consolidation, and (5) bronchovascular thickening (in the lesion). Chest X-ray features of pneumonia include (1) ground-glass opacities (however, they are opposite to COVID-19 central distribution and unilateral), (2) distribution more along the bronchovascular bundle, (3) vascular thickening, (4) reticular opacity, and (5) bronchial wall thickening. A wide variety of models and solutions have been proposed. One of the proposed models was completely automated and did not require manual feature extraction. The reported accuracy was 98.08% for binary and 87.02% for multi-class tasks [107]. The best and the most commonly used framework for automated classification and detection of the pulmonary manifestations of COVID-19 using chest CT and chest X-ray is the detection and classification CNN.
Chest X-rays are mostly used for detection, while CT scans are mostly used for classification tasks. The segmentation of radiological patterns is primarily executed using CT scans [108]. Transfer learning has been used to accelerate model learning and diminish the requirement for large training data sets by using pre-trained CNN models [108]. Sekeroglu et al. demonstrated that a CNN with minimized fully connected convolutional layers was capable of identifying COVID-19 within two classes with mean ROC AUC scores of higher than 96.0% [56].
Dansana et al. also achieved high performance of the models by using the least computationally intensive deep learning architecture models to detect COVID-19 on chest X-ray images. The VGG-16 model achieved the highest precision of 100% [57]. The next group developed and assessed an AI model using a large dataset including more than ten thousand CT volumes from patients with different pulmonary pathologies (COVID-19, non-COVID-19 viral pneumonia, such as influenza-A/B, non-viral community-acquired pneumonia, and non-pneumonia). This deep CNN model achieved an area under the curve on two publicly available datasets of 92.99% on CC-CCII and 93.25% on MosMedData. The AI model outperformed radiologists [109].
Heidari et al. developed a transfer deep-learning-based CNN model to classify pulmonary involvement in COVID-19 patients using chest X-ray images. The image preprocessing generated better input image data for developing deep learning models. The authors achieved high classification performance, which can be further optimized to detect COVID-19 cases and validated using large and diverse image datasets [110].
The next research group used a decision fusion approach combining the predictions of each of the individual deep CNN models to improve the predictive performance. Such an approach achieved an F1-Score of 0.853 and ROC AUC of 0.824 and reduced false positives. The performance of this model could be further boosted by applying image augmentation transfer learning and feature level fusion [111].
Afshar et al. developed and evaluated a capsule network framework for the diagnosis of COVID-19 using X-ray images. The framework consisted of several capsules and convolutional layers. The model achieved excellent performance with a low number of used parameters. The pre-training process improved the accuracy, specificity, and AUC. The model achieved a sensitivity of 90%, a specificity of 95.8%, an accuracy of 95.7%, and an area under the curve of 0.97. The model is publically available [112].
Jin et al. developed an AI system consisting of five main networks: (1) pulmonary segmentation, (2) slice diagnosis, (3) COVID-infectious slice location, (4) visualization, and (5) image phenotype analysis [109]. The system showed excellent performance on the test sets, with an AUC of 0.9745 for COVID-19, 0.9804 for community-acquired pneumonia, 0.9885 for influenza, and 0.9752 for non-pneumonia. Moreover, for diagnosing COVID-19, the sensitivity was 0.8703, the specificity was 0.9660, and the multi-way AUC was 0.9781. The authors also performed radiomics 36 feature extraction from the attentional regions and identified 665-dimensional imaging features to find the twelve most discriminative features to differentiate COVID-19 from other types of pneumonia. However, there was no significant difference in radiomics features in differentiating influenza from COVID-19 [109].
Although hypoxemic COVID-19 patients share the same etiologic factor (SARS-CoV-2), for severely hypoxemic patients, despite sharing a single etiology (SARS-CoV-2), their pulmonary manifestations can be quite different from one another. The spectrum of respiratory failure can vary from normal breathing (“silent” hypoxemia) to markedly dyspneic, as well as from responsive to certain treatments, such as nitric oxide or prone positioning, to not. Therefore, even the same disease can be characterized by heterogeneous clinical manifestation. The same disease actually presents itself with impressive difference. Two phenotypes (L and H) were proposed, which are best differentiated by CT. Therefore, artificial intelligence can potentially be useful in phenotype identification and can assist in understanding pathophysiology, which is crucial to establish appropriate treatment [113]. Even though PCR is being used globally for COVID-19 diagnostics and tracking, CCT processed by CNN showed much better diagnostic accuracy. Thus, Lacerda et al. developed a CNN model that achieved a sensitivity of 97%, precision of 82%, and an accuracy of 88%, thereby outperforming the diagnostic accuracy of human experts (72%) based on CCT interpretation and the sensitivity of the PCR tests (53–88%) [28].

4.1. Limitations of CNN in COVID-19 Detection

A majority of studies did not perform an external validation of the algorithm on a separate test dataset and used results from the internal validation data. This might have led to an overestimation of the diagnostic performance. The overfitting was well described across many studies [111].
Although some studies reported high model performance values, they had been validated (tested) using the intrinsic database, whereas the true performance value of the algorithms can be ultimately concluded after the use of external validation (test) on separate external test datasets with previously unused data from representatives of the target population. Another limitation was an imbalance between the study groups, as well as a small sample size of the severe patients included in the study [114].
CNN models require large datasets that have all possible variants of data for achieving the highest accuracy. The shortage of available datasets is a barrier to their training. A majority of models have been trained and validated on small datasets [108]. To counteract the inefficiency of the training datasets and increase their size, several authors reported using GAN models for data augmentation. Such data augmentation made the model more robust to overfitting [108]. Another limitation includes the so-called “black box”, which is the basic feature of deep neural networks. Even though attention maps assist in interpretation by highlighting the dominant areas, they are not fully sufficient to visualize the unique features used by CNNs to differentiate between COVID-19 and non-COVID-19 pneumonia. Since the majority of studies primarily focused on radiological diagnostics (chest CT and X-ray), clinical information was not included [108].
There was also an extensive variation in the terminology, study methodology, data interpretability, and outcome. It was very difficult to formally assess the performance of algorithms due to the variation in reporting. The variability in study methodology and reporting can be partially explained by the fact that these studies were conducted by researchers from different specialties, including medical specialists (intensivists, emergency physicians, pulmonologists, radiologists), information technology specialists, and engineers. Therefore, the study goals and the experience of researchers also varied from study to study. Even though DL algorithms showed high diagnostic accuracy in medical imaging, it is, at present, difficult to determine if they are fully applicable in real clinical settings.

4.2. Potential for Future Research

The presence of large image datasets is necessary to improve generalizability and limit overfitting in training CNNs. Although techniques have improved learning on small datasets, large medical datasets are essential. The majority of achievements of CNNs are mainly based on a large amount of data.
Therefore, creating large datasets of radiological images is one of the main challenges that must be addressed by the research community of clinicians and computer scientists, as well as hospital administrators and other staff involved in AI research. Although building large clinical datasets is expensive and requires enormous work by specialists, an international collaborative effort of the medical network is necessary to advance this field. Such collaborative work should follow common guidelines for the systematic acquirement and annotation of radiological images without interfering with the clinical routine [115].
Medical datasets should ideally be annotated by certified radiologists, and this process is very time-consuming. Applications such as GTCreator [116] can contribute to ground truth creation and facilitate image annotations, sharing, and revising conclusions among different radiologists [115]. It is generally recommended to establish a bootstrapping methodology for testing sets, indicating the number of images, which images should be considered in each iteration, and the sample size and repetitions [115]. Further improvement of the diagnostic performance of the DL models could be achieved of including clinical and laboratory data.
Building large clinical datasets is expensive and requires enormous work by specialists. It is an essential procedure in the DNN approach for model reliability. The international collaborative effort of the medical network is necessary to advance this field. A more effective method is to use a generative adversarial network for generating new CT/CXR images that can contribute to the continuation of the network learning.
While numerous models have been developed, the inter-exchange of data between the sources would result in achieving better performance. Moreover, the customization of these models according to the local patient population using regionally collected data would also improve the models’ performance. Finally, the most important task for the future would be shifting from research-based model development and testing to actual implementation and continuation of testing such models in the real-world environment.
Although deep learning is currently one of the fastest developing fields in medicine with the potential to improve the diagnosis of pulmonary manifestations in COVID-19 patients, some conclusions about its superiority over clinicians seem to be over-promising. Moreover, since the beginning of the pandemic, hundreds of studies have been conducted; however, the evidence on the implementation of deep learning models into real-world clinical settings and the assessment if their value is still missing. Almost all studies reported excellent performance, and some of them even outperformed certified radiologists in making image-based diagnoses. However, many studies either did not describe exactly how they assessed DL models against doctors, or the process of assessment was not rigorous enough to support the conclusions. One study reported the opinions of radiologists about the DL models, but these opinions were rather subjective and not standardized. Despite there being a massive explosion of research focusing on DL in the diagnosis of COVID-19, there is not enough evidence of the actual deployment of these models in clinical practice. A few questions should be addressed before these models get actively used in clinical practice:
  • Can these models actually take over radiologist, or should their usage be strictly limited to a doctor’s assistantship, second opinions, or double-checking?
  • How deeply can healthcare workers rely on these models depending on the level of hospital and doctor availability? Can these models be used without the radiologist’s opinion or only in the case that the radiologist is absolutely unavailable? Or, in this case, should a radiologist definitely double-check the model-made diagnosis?
  • Can these models be used in the region where a particular model has not been validated or accustomed?
Therefore, the conduction of high-quality studies with transparent reporting of the results is necessary to avoid hype, assist healthcare workers, and reduce or avoid harm to patients. Future research should be focused on post-deployment, region-specific validation, and safety in clinical settings. The additional potential pathway that should be considered for achieving advancements in the field is solving patients’ privacy and ethical requirements.

Author Contributions

M.-H.L. and D.V. contributed to the design and implementation of the research. A.S. and M.K. contributed to survey all articles and analyzed their methodologies. M.-H.L. and D.V. took the lead in writing the manuscript. All authors provided critical feedback and helped shape the research, analysis, and manuscript. D.V. supervised the project. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to thank Nazarbayev University for funding this article under the Social Policy Grant (201705 SPG).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was supported by the Faculty Development Competitive Research Grant Program (No. 080420FD1909) at Nazarbayev University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, N.; Zhang, D.; Wang, W.; Li, X.; Yang, B.; Song, J.; Zhao, X.; Huang, B.; Shi, W.; Lu, R.; et al. A novel coronavirus from patients with pneumonia in China, 2019. N. Engl. J. Med. 2020, 382, 727–733. [Google Scholar] [CrossRef] [PubMed]
  2. Peñarrubia, L.; Ruiz, M.; Porco, R.; Rao, S.N.; Juanola-Falgarona, M.; Manissero, D.; López-Fontanals, M.; Pareja, J. Multiple assays in a real-time RT-PCR SARS-CoV-2 panel can mitigate the risk of loss of sensitivity by new genomic variants during the COVID-19 outbreak. Int. J. Infect. Dis. 2020, 97, 225–229. [Google Scholar] [CrossRef]
  3. Guan, W.J.; Ni, Z.Y.; Hu, Y.; Liang, W.H.; Ou, C.Q.; He, J.X.; Liu, L.; Shan, H.; Lei, C.L.; Hui, D.S.; et al. Clinical characteristics of coronavirus disease 2019 in China. N. Engl. J. Med. 2020, 382, 1708–1720. [Google Scholar] [CrossRef] [PubMed]
  4. Tsuchida, T.; Fujitani, S.; Yamasaki, Y.; Kunishima, H.; Matsuda, T. Development of a protective device for RT-PCR testing SARS-CoV-2 in COVID-19 patients. Infect. Control Hosp. Epidemiol. 2020, 41, 975–976. [Google Scholar] [CrossRef] [PubMed]
  5. Li, Y.; Xia, L. Coronavirus disease 2019 (COVID-19): Role of chest CT in diagnosis and management. Am. J. Roentgenol. 2020, 214, 1280–1286. [Google Scholar] [CrossRef]
  6. Sinonquel, P.; Eelbode, T.; Bossuyt, P.; Maes, F.; Bisschops, R. Artificial intelligence and its impact on quality improvement in upper and lower gastrointestinal endoscopy. Dig. Endosc. 2021, 33, 242–253. [Google Scholar] [CrossRef]
  7. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  8. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  9. Savaş, S.; Topaloğlu, N.; Kazcı, Ö.; Koşar, P.N. Classification of carotid artery intima media thickness ultrasound images with deep learning. J. Med. Syst. 2019, 43, 1–12. [Google Scholar] [CrossRef]
  10. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. Int. J. Surg. 2021, 88, 105906. [Google Scholar] [CrossRef]
  11. Yamac, M.; Ahishali, M.; Degerli, A.; Kiranyaz, S.; Chowdhury, M.E.H.; Gabbouj, M. Convolutional Sparse Support Estimator-Based COVID-19 Recognition From X-Ray Images. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 1810–1820. [Google Scholar] [CrossRef] [PubMed]
  12. Kaur, T.; Gandhi, T.K.; Panigrahi, B.K. Automated Diagnosis of COVID-19 Using Deep Features and Parameter Free BAT Optimization. IEEE J. Transl. Eng. Health Med. 2021, 9, 1–9. [Google Scholar] [CrossRef] [PubMed]
  13. Rangarajan, K.; Muku, S.; Garg, A.K.; Gabra, P.; Shankar, S.H.; Nischal, N.; Soni, K.D.; Bhalla, A.S.; Mohan, A.; Tiwari, P.; et al. Artificial Intelligence–assisted chest X-ray assessment scheme for COVID-19. Eur. Radiol. 2021, 31, 6039–6048. [Google Scholar] [CrossRef] [PubMed]
  14. Zhao, W.; Jiang, W.; Qiu, X. Deep learning for COVID-19 detection based on CT images. Sci. Rep. 2021, 11, 14353. [Google Scholar] [CrossRef] [PubMed]
  15. Chikontwe, P.; Luna, M.; Kang, M.; Hong, K.S.; Ahn, J.H.; Park, S.H. Dual Attention Multiple Instance Learning with Unsupervised Complementary Loss for COVID-19 Screening. Med. Image Anal. 2021, 72, 102105. [Google Scholar] [CrossRef] [PubMed]
  16. Dong, S.; Yang, Q.; Fu, Y.; Tian, M.; Zhuo, C. RCoNet: Deformable Mutual Information Maximization and High-order Uncertainty-aware Learning for Robust COVID-19 Detection. IEEE Trans. Neural Netw. Learn. Syst. 2021, 2, 3401–3411. [Google Scholar] [CrossRef] [PubMed]
  17. Fridadar, M.; Amer, R.; Gozes, O.; Nassar, J.; Greenspan, H. COVID-19 in CXR: From detection and severity scoring to patient disease monitoring. IEEE J. Biomed. Health Inform. 2021, 25, 1892–1903. [Google Scholar] [CrossRef]
  18. Yang, Y.; Lure, F.Y.; Miao, H.; Zhang, Z.; Jaeger, S.; Liu, J.; Guo, L. Using artificial intelligence to assist radiologists in distinguishing COVID-19 from other pulmonary infections. J. X-ray Sci. Technol. 2021, 29, 1–17. [Google Scholar] [CrossRef]
  19. Zhu, Z.; Xingming, Z.; Tao, G.; Dan, T.; Li, J.; Chen, X.; Li, Y.; Zhou, Z.; Zhang, X.; Zhou, J.; et al. Classification of COVID-19 by compressed chest CT image through deep learning on a large patients cohort. Interdiscip. Sci. Comput. Life Sci. 2021, 13, 73–82. [Google Scholar] [CrossRef]
  20. Zhang, Q.; Chen, Z.; Liu, G.; Zhang, W.; Du, Q.; Tan, J.; Gao, Q. Artificial Intelligence Clinicians Can Use Chest Computed Tomography Technology to Automatically Diagnose Coronavirus Disease 2019 (COVID-19) Pneumonia and Enhance Low-Quality Images. Infect. Drug Resist. 2021, 14, 671. [Google Scholar] [CrossRef]
  21. COVID-19 Patients Lungs X Ray Images 10000. Available online: https://www.kaggle.com/nabeelsajid917/COVID-19-x-ray-10000-images (accessed on 5 May 2023).
  22. Owais, M.; Lee, Y.W.; Mahmood, T.; Haider, A.; Sultan, H.; Park, K.R. Multilevel Deep-Aggregated Boosted Network to Recognize COVID-19 Infection from Large-Scale Heterogeneous Radiographic Data. IEEE J. Biomed. Health Inform. 2021, 25, 1881–1891. [Google Scholar] [CrossRef] [PubMed]
  23. Zhong, A.; Li, X.; Wu, D.; Ren, H.; Kim, K.; Kim, Y.; Buch, V.; Neumark, N.; Bizzo, B.; Tak, W.Y.; et al. Deep metric learning-based image retrieval system for chest radiograph and its clinical applications in COVID-19. Med. Image Anal. 2021, 70, 101993. [Google Scholar] [CrossRef] [PubMed]
  24. Liu, B.; Liu, P.; Dai, L.; Yang, Y.; Xie, P.; Tan, Y.; Du, J.; Shan, W.; Zhao, C.; Zhong, Q.; et al. Assisting scalable diagnosis automatically via CT images in the combat against COVID-19. Sci. Rep. 2021, 11, 4145. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, X.; Wang, D.; Shao, J.; Tian, S.; Tan, W.; Ma, Y.; Xu, Q.; Ma, X.; Li, D.; Chai, J.; et al. A deep learning integrated radiomics model for identification of coronavirus disease 2019 using computed tomography. Sci. Rep. 2021, 11, 3938. [Google Scholar] [CrossRef]
  26. Li, Z.; Zhao, W.; Shi, F.; Qi, L.; Xie, X.; Wei, Y.; Ding, Z.; Gao, Y.; Wu, S.; Liu, J.; et al. A novel multiple instance learning framework for COVID-19 severity assessment via data augmentation and self-supervised learning. Med. Image Anal. 2021, 69, 101978. [Google Scholar] [CrossRef]
  27. Polat, Ç.; Karaman, O.; Karaman, C.; Korkmaz, G.; Balcí, M.C.; Kelek, S.E. COVID-19 diagnosis from chest X-ray images using transfer learning: Enhanced performance by debiasing dataloader. J. X-ray Sci. Technol. 2021, 29, 19–36. [Google Scholar] [CrossRef]
  28. Lacerda, P.; Barros, B.; Albuquerque, C.; Conci, A. Hyperparameter Optimization for COVID-19 Pneumonia Diagnosis Based on Chest CT. Sensors 2021, 21, 2174. [Google Scholar] [CrossRef]
  29. Quan, H.; Xu, X.; Zheng, T.; Li, Z.; Zhao, M.; Cui, X. DenseCapsNet: Detection of COVID-19 from X-ray images using a capsule neural network. Comput. Biol. Med. 2021, 133, 104399. [Google Scholar] [CrossRef]
  30. Sedik, A.; Iliyasu, A.M.; El-Rahiem, A.; Abdel Samea, M.E.; Abdel-Raheem, A.; Hammad, M.; Peng, J.; El-Samie, A.; Fathi, E.; El-Latif, A.; et al. Deploying machine and deep learning models for efficient data-augmented detection of COVID-19 infections. Viruses 2020, 12, 769. [Google Scholar] [CrossRef]
  31. Ozyurt, F.; Tuncer, T.; Subasi, A. An automated COVID-19 detection based on fused dynamic exemplar pyramid feature extraction and hybrid feature selection using deep learning. Comput. Biol. Med. 2021, 132, 104356. [Google Scholar] [CrossRef]
  32. Habib, N.; Rahman, M.M. Diagnosis of corona diseases from associated genes and X-ray images using machine learning algorithms and deep CNN. Inform. Med. Unlocked 2021, 24, 100621. [Google Scholar] [CrossRef] [PubMed]
  33. Shi, W.; Tong, L.; Zhu, Y.; Wang, M.D. COVID-19 Automatic Diagnosis with Radiographic Imaging: Explainable AttentionTransfer Deep Neural Networks. IEEE J. Biomed. Health Inform. 2021, 25, 2376–2387. [Google Scholar] [CrossRef] [PubMed]
  34. Le, D.N.; Parvathy, V.S.; Gupta, D.; Khanna, A.; Rodrigues, J.J.; Shankar, K. IoT enabled depthwise separable convolution neural network with deep support vector machine for COVID-19 diagnosis and classification. Int. J. Mach. Learn. Cybern. 2021, 12, 3235–3248. [Google Scholar] [CrossRef]
  35. Lee, E.H.; Zheng, J.; Colak, E.; Mohammadzadeh, M.; Houshmand, G.; Bevins, N.; Kitamura, F.; Altinmakas, E.; Reis, E.P.; Kim, J.K.; et al. Deep COVID DeteCT: An international experience on COVID-19 lung detection and prognosis using chest CT. NPJ Digit. Med. 2021, 4, 11. [Google Scholar] [CrossRef] [PubMed]
  36. Khuzani, A.Z.; Heidari, M.; Shariati, S.A. COVID-Classifier: An automated machine learning model to assist in the diagnosis of COVID-19 infection in chest x-ray images. Sci. Rep. 2021, 11, 9887. [Google Scholar] [CrossRef] [PubMed]
  37. Fang, C.; Bai, S.; Chen, Q.; Zhou, Y.; Xia, L.; Qin, L.; Gong, S.; Xie, X.; Zhou, C.; Tu, D.; et al. Deep learning for predicting COVID-19 malignant progression. Med Image Anal. 2021, 72, 102096. [Google Scholar] [CrossRef]
  38. Madhavan, M.V.; Khamparia, A.; Gupta, D.; Pande, S.; Tiwari, P.; Hossain, M.S. Res-CovNet: An internet of medical health things driven COVID-19 framework using transfer learning. Neural Comput. Appl. 2021, 1–14. [Google Scholar] [CrossRef]
  39. Kundu, R.; Basak, H.; Singh, P.K.; Ahmadian, A.; Ferrara, M.; Sarkar, R. Fuzzy rank-based fusion of CNN models using Gompertz function for screening COVID-19 CT-scans. Sci. Rep. 2021, 11, 14133. [Google Scholar] [CrossRef]
  40. Rahman, M.M.; Nooruddin, S.; Hasan, K.A.; Dey, N.K. HOG+ CNN Net: Diagnosing COVID-19 and pneumonia by deep neural network from chest X-Ray images. Sn Comput. Sci. 2021, 2, 371. [Google Scholar] [CrossRef]
  41. Li, J.; Zhao, G.; Tao, Y.; Zhai, P.; Chen, H.; He, H.; Cai, T. Multi-task contrastive learning for automatic CT and X-ray diagnosis of COVID-19. Pattern Recognit. 2021, 114, 107848. [Google Scholar] [CrossRef]
  42. Panthakkan, A.; Anzar, S.; Al Mansoori, S.; Al Ahmad, H. A novel DeepNet model for the efficient detection of COVID-19 for symptomatic patients. Biomed. Signal Process. Control 2021, 68, 102812. [Google Scholar] [CrossRef] [PubMed]
  43. Keles, A.; Keles, M.B.; Keles, A. COV19-CNNet and COV19-ResNet: Diagnostic inference Engines for early detection of COVID-19. Cogn. Comput. 2021, 1–11. [Google Scholar] [CrossRef] [PubMed]
  44. Qi, X.; Brown, L.G.; Foran, D.J.; Nosher, J.; Hacihaliloglu, I. Chest X-ray image phase features for improved diagnosis of COVID-19 using convolutional neural network. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 197–206. [Google Scholar] [CrossRef] [PubMed]
  45. Khadidos, A.; Khadidos, A.O.; Kannan, S.; Natarajan, Y.; Mohanty, S.N.; Tsaramirsis, G. Analysis of COVID-19 Infections on a CT Image Using DeepSense Model. Front. Public Health 2020, 8, 599550. [Google Scholar] [CrossRef]
  46. Elsharkawy, M.; Sharafeldeen, A.; Taher, F.; Shalaby, A.; Soliman, A.; Mahmoud, A.; Ghazal, M.; Khalil, A.; Alghamdi, N.S.; Razek, A.A.K.A.; et al. Early assessment of lung function in coronavirus patients using invariant markers from chest X-rays images. Sci. Rep. 2021, 11, 12095. [Google Scholar] [CrossRef]
  47. Gunraj, H.; Wang, L.; Wong, A. COVIDNet-CT: A Tailored Deep Convolutional Neural Network Design for Detection of COVID-19 Cases From Chest CT Images. Front. Med. 2020, 7, 1025. [Google Scholar] [CrossRef]
  48. Xiao, L.S.; Li, P.; Sun, F.; Zhang, Y.; Xu, C.; Zhu, H.; Cai, F.Q.; He, Y.L.; Zhang, W.F.; Ma, S.C.; et al. Development and Validation of a Deep Learning-Based Model Using Computed Tomography Imaging for Predicting Disease Severity of Coronavirus Disease 2019. Front. Bioeng. Biotechnol. 2020, 8, 898. [Google Scholar] [CrossRef]
  49. Harmon, S.A.; Sanford, T.H.; Xu, S.; Turkbey, E.B.; Roth, H.; Xu, Z.; Yang, D.; Myronenko, A.; Anderson, V.; Amalou, A.; et al. Artificial intelligence for the detection of COVID-19 pneumonia on chest CT using multinational datasets. Nat. Commun. 2020, 11, 4080. [Google Scholar] [CrossRef]
  50. Kikkisetti, S.; Zhu, J.; Shen, B.; Li, H.; Duong, T.Q. Deep-learning convolutional neural networks with transfer learning accurately classify COVID-19 lung infection on portable chest radiographs. PeerJ 2020, 8, e10309. [Google Scholar] [CrossRef]
  51. El-Bana, S.; Al-Kabbany, A.; Sharkas, M. A multi-task pipeline with specialized streams for classification and segmentation of infection manifestations in COVID-19 scans. PeerJ Comput. Sci. 2020, 6, e303. [Google Scholar] [CrossRef]
  52. Wang, D.; Mo, J.; Zhou, G.; Xu, L.; Liu, Y. An efficient mixture of deep and machine learning models for COVID-19 diagnosis in chest X-ray images. PLoS ONE 2020, 15, e0242535. [Google Scholar] [CrossRef] [PubMed]
  53. Albadr, M.A.A.; Tiun, S.; Ayob, M.; Al-Dhief, F.T.; Omar, K.; Hamzah, F.A. Optimised genetic algorithm-extreme learning machine approach for automatic COVID-19 detection. PLoS ONE 2020, 15, e0242899. [Google Scholar] [CrossRef] [PubMed]
  54. Li, L.; Qin, L.; Xu, Z.; Yin, Y.; Wang, X.; Kong, B.; Bai, J.; Lu, Y.; Fang, Z.; Song, Q.; et al. Using artificial intelligence to detect COVID-19 and community-acquired pneumonia based on pulmonary CT: Evaluation of the diagnostic accuracy. Radiology 2020, 296, E65–E71. [Google Scholar] [CrossRef] [PubMed]
  55. Pham, T.D. A comprehensive study on classification of COVID-19 on computed tomography with pretrained convolutional neural networks. Sci. Rep. 2020, 10, 16942. [Google Scholar] [CrossRef]
  56. Sekeroglu, B.; Ozsahin, I. Detection of COVID-19 from Chest X-Ray Images Using Convolutional Neural Networks. Slas Technol. Transl. Life Sci. Innov. 2020, 25, 553–565. [Google Scholar] [CrossRef]
  57. Dansana, D.; Kumar, R.; Bhattacharjee, A.; Hemanth, D.J.; Gupta, D.; Khanna, A.; Castillo, O. Early diagnosis of COVID-19-affected patients based on X-ray and computed tomography images using deep learning algorithm. Soft Comput. 2020, 27, 2635–2643. [Google Scholar] [CrossRef]
  58. Wang, X.; Deng, X.; Fu, Q.; Zhou, Q.; Feng, J.; Ma, H.; Liu, W.; Zheng, C. A weakly-supervised framework for COVID-19 classification and lesion localization from chest CT. IEEE Trans. Med. Imaging 2020, 39, 2615–2625. [Google Scholar] [CrossRef]
  59. Serte, S.; Demirel, H. Deep learning for diagnosis of COVID-19 using 3D CT scans. Comput. Biol. Med. 2021, 132, 104306. [Google Scholar] [CrossRef]
  60. Sharifrazi, D.; Alizadehsani, R.; Roshanzamir, M.; Joloudari, J.H.; Shoeibi, A.; Jafari, M.; Hussain, S.; Sani, Z.A.; Hasanzadeh, F.; Khozeimeh, F.; et al. Fusion of convolution neural network, support vector machine and Sobel filter for accurate detection of COVID-19 patients using X-ray images. Biomed. Signal Process. Control 2021, 68, 102622. [Google Scholar] [CrossRef]
  61. Gifani, P.; Shalbaf, A.; Vafaeezadeh, M. Automated detection of COVID-19 using ensemble of transfer learning with deep convolutional neural network based on CT scans. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 115–123. [Google Scholar] [CrossRef]
  62. Sagi, O.; Rokach, L. Ensemble learning: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2018, 8, e1249. [Google Scholar] [CrossRef]
  63. Singh, M.; Bansal, S.; Ahuja, S.; Dubey, R.K.; Panigrahi, B.K.; Dey, N. Transfer learning-based ensemble support vector machine model for automated COVID-19 detection using lung computerized tomography scan data. Med. Biol. Eng. Comput. 2021, 59, 825–839. [Google Scholar] [CrossRef] [PubMed]
  64. Pathak, Y.; Shukla, P.K.; Tiwari, A.; Stalin, S.; Singh, S. Deep transfer learning based classification model for COVID-19 disease. Irbm 2020, 43, 87–92. [Google Scholar] [CrossRef] [PubMed]
  65. Pham, T.D. Classification of COVID-19 chest X-rays with deep learning: New models or fine tuning? Health Inf. Sci. Syst. 2021, 9, 2. [Google Scholar] [CrossRef]
  66. Saad, W.; Shalaby, W.A.; Shokair, M.; El-Samie, F.A.; Dessouky, M.; Abdellatef, E. COVID-19 classification using deep feature concatenation technique. J. Ambient Intell. Humaniz. Comput. 2022, 13, 2025–2043. [Google Scholar] [CrossRef]
  67. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  68. Zhang, K.; Liu, X.; Shen, J.; Li, Z.; Sang, Y.; Wu, X.; Zha, Y.; Liang, W.; Wang, C.; Wang, K.; et al. Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell 2020, 181, 1423–1433. [Google Scholar] [CrossRef]
  69. Ghaderzadeh, M.; Asadi, F.; Jafari, R.; Bashash, D.; Abolghasemi, H.; Aria, M. Deep Convolutional Neural Network-Based Computer-Aided Detection System for COVID-19 Using Multiple Lung Scans: Design and Implementation Study. J. Med. Internet Res. 2021, 23, e27468. [Google Scholar] [CrossRef]
  70. Gupta, P.; Siddiqui, M.K.; Huang, X.; Morales-Menendez, R.; Pawar, H.; Terashima-Marin, H.; Wajid, M.S. COVID-WideNet—A capsule network for COVID-19 detection. Appl. Soft Comput. 2022, 122, 108780. [Google Scholar] [CrossRef]
  71. El-Kenawy, E.S.M.; Mirjalili, S.; Ibrahim, A.; Alrahmawy, M.; El-Said, M.; Zaki, R.M.; Eid, M.M. Advanced meta-heuristics, convolutional neural networks, and feature selectors for efficient COVID-19 x-ray chest image classification. IEEE Access 2021, 9, 36019–36037. [Google Scholar] [CrossRef]
  72. Khan, S.H.; Sohail, A.; Khan, A.; Lee, Y.S. COVID-19 Detection in Chest X-ray Images Using a New Channel Boosted CNN. Diagnostics 2022, 12, 267. [Google Scholar] [CrossRef]
  73. Shah, P.M.; Ullah, F.; Shah, D.; Gani, A.; Maple, C.; Wang, Y.; Shahid; Abrar, M.; Islam, S.U. Deep GRU-CNN model for COVID-19 detection from chest X-rays data. IEEE Access 2021, 10, 35094–35105. [Google Scholar] [CrossRef] [PubMed]
  74. Morozov, S.P.; Andreychenko, A.E.; Blokhin, I.A.; Gelezhe, P.B.; Gonchar, A.P.; Nikolaev, A.E.; Pavlov, N.A.; Chernina, V.Y.; Gombolevskiy, V.A. Mosmeddata: Data set of 1110 chest ct scans performed during the COVID-19 epidemic. Digit. Diagn. 2020, 1, 49–59. [Google Scholar] [CrossRef]
  75. Soares, E.; Angelov, P.; Biaso, S.; Froes, M.H.; Abe, D.K. SARS-CoV-2 CT-scan dataset: A large dataset of real patients CT scans for SARS-CoV-2 identification. medRxiv 2020. [Google Scholar] [CrossRef]
  76. Soares, E. & Angelov, P. A large dataset of real patients CT scans for COVID-19 identification. Harv. Dataverse 2020, 1, 1–8. [Google Scholar] [CrossRef]
  77. Open Access Biomedical Image Search Engine. Available online: https://openi.nlm.nih.gov/ (accessed on 20 September 2021).
  78. Italia Radiology Hospitals Dataset. 2020. Available online: https://sirm.org/category/senza-categoria/COVID-19/ (accessed on 5 May 2023).
  79. He, X.; Yang, X.; Zhang, S.; Zhao, J.; Zhang, Y.; Xing, E.; Xie, P. Sample-Efficient Deep Learning for COVID-19 Diagnosis Based on CT Scans. medRxiv 2020. [Google Scholar] [CrossRef]
  80. Armato III, S.G.; McLennan, G.; Bidaut, L.; McNitt-Gray, M.F.; Meyer, C.R.; Reeves, A.P.; Zhao, B.; Aberle, D.R.; Henschke, C.I.; Hoffman, E.A.; et al. The lung image database consortium (LIDC) and image database resource initiative (IDRI): A completed reference database of lung nodules on CT scans. Med. Phys. 2011, 38, 915–931. [Google Scholar] [CrossRef]
  81. Ning, W.; Lei, S.; Yang, J.; Cao, Y.; Jiang, P.; Yang, Q.; Zhang, J.; Wang, X.; Chen, F.; Geng, Z.; et al. iCTCF: An integrative resource of chest computed tomography images and clinical features of patients with COVID-19 pneumonia. Res. Sq. 2020. [Google Scholar] [CrossRef]
  82. Gunraj, H.; Sabri, A.; Koff, D.; Wong, A. COVID-Net CT-2: Enhanced Deep Neural Networks for Detection of COVID-19 from Chest CT Images Through Bigger, More Diverse Learning. arXiv 2021, arXiv:2101.07433. [Google Scholar] [CrossRef]
  83. Haghanifar, A.; Majdabadi, M.M.; Choi, Y.; Deivalakshmi, S.; Ko, S. COVID-CXNet: Detecting COVID-19 in Frontal Chest X-ray Images using Deep Learning. Multimed. Tools Appl. 2020, 81, 30615–30645. [Google Scholar] [CrossRef]
  84. Linda, W.; Alexander, W.; Zhong, Q.L.; Paul, M.; Audrey, C.; Hayden, G.; James, L.; Ashkan, E.; Kim-Ann, G.; Abdul, A.H. Figure 1 COVID-19 Chest X-ray Dataset Initiative. 2020. Available online: https://github.com/agchung/Figure1-COVID-chestxray-dataset (accessed on 5 May 2023).
  85. Chowdhury, M.E.; Rahman, T.; Khandakar, A.; Mazhar, R.; Kadir, M.A.; Mahbub, Z.B.; Islam, K.R.; Khan, M.S.; Iqbal, A.; Al Emadi, N.; et al. Can AI help in screening viral and COVID-19 pneumonia? IEEE Access 2020, 8, 132665–132676. [Google Scholar] [CrossRef]
  86. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2097–2106. [Google Scholar]
  87. Kermany, D.; Zhang, K.; Goldbaum, M. Labeled Optical Coherence Tomography (OCT) and Chest X-ray Images for Classification, Mendeley Data; Mendeley: London, UK, 2018. [Google Scholar] [CrossRef]
  88. Cohen, J.P.; Morrison, P.; Dao, L.; Roth, K.; Duong, T.Q.; Ghassemi, M. COVID-19 Image Data Collection: Prospective Predictions Are the Future. arXiv 2020, arXiv:2006.11988. [Google Scholar]
  89. Degerli, A.; Ahishali, M.; Yamac, M.; Kiranyaz, S.; Chowdhury, M.E.; Hameed, K.; Hamid, T.; Mazhar, R.; Gabbouj, M. COVID-19 infection map generation and detection from chest X-ray images. Health Inf. Sci. Syst. 2021, 9, 15. [Google Scholar] [CrossRef] [PubMed]
  90. CDGC Dataset. 2020. Available online: https://drive.google.com/uc?id=1coM7x3378f-Ou2l6Pg2wldaOI7Dntu1a (accessed on 5 May 2023).
  91. Vayá, M.d.l.I.; Saborit, J.M.; Montell, J.A.; Pertusa, A.; Bustos, A.; Cazorla, M.; Galant, J.; Barber, X.; Orozco-Beltrán, D.; García-García, F.; et al. Bimcv COVID-19+: A large annotated dataset of rx and ct images from COVID-19 patients. arXiv 2020, arXiv:2006.01174. [Google Scholar]
  92. Sait, U.; Lal, K.G.; Prajapati, S.; Bhaumik, R.; Kumar, T.; Sanjana, S.; Bhalla, K. Curated Dataset for COVID-19 Posterior-Anterior Chest Radiography Images (X-Rays); Mendeley: London, UK, 2020; Volume 1. [Google Scholar] [CrossRef]
  93. Montalbo, F.J.P. Diagnosing COVID-19 chest x-rays with a lightweight truncated DenseNet with partial layer freezing and feature fusion. Biomed. Signal Process. Control 2021, 68, 102583. [Google Scholar] [CrossRef] [PubMed]
  94. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 2018, 172, 1122–1131. [Google Scholar] [CrossRef] [PubMed]
  95. Tsai, E.B.; Simpson, S.; Lungren, M.P.; Hershman, M.; Roshkovan, L.; Colak, E.; Erickson, B.J.; Shih, G.; Stein, A.; Kalpathy-Cramer, J.; et al. The RSNA International COVID-19 Open Radiology Database (RICORD). Radiology 2021, 299, E204–E213. [Google Scholar] [CrossRef]
  96. Umer, M.; Ashraf, I.; Ullah, S.; Mehmood, A.; Choi, G.S. COVINet: A convolutional neural network approach for predicting COVID-19 from chest X-ray images. J. Ambient. Intell. Humaniz. Comput. 2021, 13, 535–547. [Google Scholar] [CrossRef]
  97. Ghadezadeh, M.; Eshraghi, M.A.; Asadi, F.; Hosseini, A.; Jafari, R.; Bashash, D.; Abolghasemi, H. Efficient Framework for Detection of COVID-19 Omicron and Delta Variants Based on Two Intelligent Phases of CNN Models. Comput. Math. Methods Med. 2022, 2022, 4838009. [Google Scholar] [CrossRef]
  98. COVID-19 Omicron and Delta Variant Lung CT Scans. 2022. Available online: https://www.kaggle.com/datasets/mohammadamireshraghi/covid19-omicron-and-delta-variant-ct-scan-dataset (accessed on 5 May 2023).
  99. COVID-19 Chest XRay Images. 2021. Available online: https://www.kaggle.com/datasets/mohammadamireshraghi/covid19-xray-local-data (accessed on 5 May 2023).
  100. Ghaderzadeh, M.; Aria, M.; Hosseini, A.; Asadi, F.; Bashash, D.; Abolghasemi, H. A fast and efficient CNN model for B-ALL diagnosis and its subtypes classification using peripheral blood smear images. Int. J. Intell. Syst. 2022, 37, 5113–5133. [Google Scholar] [CrossRef]
  101. Hussain, Z.; Gimenez, F.; Yi, D.; Rubin, D. Differential Data Augmentation Techniques for Medical Imaging Classification Tasks. AMIA Annu. Symp. Proc. AMIA Symp. 2018, 2017, 979–984. [Google Scholar]
  102. Cheng, Q.; Fan, M.; Hao, J.; Hu, X.; Ge, X.; Hu, Z.; Li, Z. Chest CT features of children infected by B.1.617.2 (Delta) variant of COVID-19. World J. Pediatr. 2022, 18, 37–42. [Google Scholar] [CrossRef] [PubMed]
  103. Wolff, R.F.; Moons, K.G.; Riley, R.D.; Whiting, P.F.; Westwood, M.; Collins, G.S.; Reitsma, J.B.; Kleijnen, J.; Mallett, S.; Group†, P. PROBAST: A tool to assess the risk of bias and applicability of prediction model studies. Ann. Intern. Med. 2019, 170, 51–58. [Google Scholar] [CrossRef] [PubMed]
  104. Yan, T.; Wong, P.K.; Ren, H.; Wang, H.; Wang, J.; Li, Y. Automatic distinction between COVID-19 and common pneumonia using multi-scale convolutional neural network on chest ct scans. Chaos Solitons Fractals 2020, 140, 110153. [Google Scholar] [CrossRef] [PubMed]
  105. Anastasopoulos, C.; Weikert, T.; Yang, S.; Abdulkadir, A.; Schmülling, L.; Bühler, C.; Paciolla, F.; Sexauer, R.; Cyriac, J.; Nesic, I.; et al. Development and clinical implementation of tailored image analysis tools for COVID-19 in the midst of the pandemic: The synergetic effect of an open, clinically embedded software development platform and machine learning. Eur. J. Radiol. 2020, 131, 109233. [Google Scholar] [CrossRef]
  106. Elgendi, M.; Nasir, M.U.; Tang, Q.; Fletcher, R.R.; Howard, N.; Menon, C.; Ward, R.; Parker, W.; Nicolaou, S. The performance of deep neural networks in differentiating chest X-rays of COVID-19 patients from other bacterial and viral pneumonias. Front. Med. 2020, 7, 550. [Google Scholar] [CrossRef] [PubMed]
  107. Ozturk, T.; Talo, M.; Yildirim, E.A.; Baloglu, U.B.; Yildirim, O.; Acharya, U.R. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput. Biol. Med. 2020, 121, 103792. [Google Scholar] [CrossRef]
  108. Kugunavar, S.; Prabhakar, C. Convolutional neural networks for the diagnosis and prognosis of the coronavirus disease pandemic. Vis. Comput. Ind. Biomed. Art 2021, 4, 12. [Google Scholar] [CrossRef]
  109. Jin, C.; Chen, W.; Cao, Y.; Xu, Z.; Tan, Z.; Zhang, X.; Deng, L.; Zheng, C.; Zhou, J.; Shi, H.; et al. Development and evaluation of an artificial intelligence system for COVID-19 diagnosis. Nat. Commun. 2020, 11, 5088. [Google Scholar] [CrossRef]
  110. Heidari, M.; Mirniaharikandehei, S.; Khuzani, A.Z.; Danala, G.; Qiu, Y.; Zheng, B. Improving the performance of CNN to predict the likelihood of COVID-19 using chest X-ray images with preprocessing algorithms. Int. J. Med. Inform. 2020, 144, 104284. [Google Scholar] [CrossRef]
  111. Mishra, A.K.; Das, S.K.; Roy, P.; Bandyopadhyay, S. Identifying COVID19 from chest CT images: A deep convolutional neural networks based approach. J. Healthc. Eng. 2020, 2020, 8843664. [Google Scholar] [CrossRef]
  112. Afshar, P.; Heidarian, S.; Naderkhani, F.; Oikonomou, A.; Plataniotis, K.N.; Mohammadi, A. Covid-caps: A capsule network-based framework for identification of COVID-19 cases from X-ray images. Pattern Recognit. Lett. 2020, 138, 638–643. [Google Scholar] [CrossRef] [PubMed]
  113. Gattinoni, L.; Chiumello, D.; Caironi, P.; Busana, M.; Romitti, F.; Brazzi, L.; Camporota, L. COVID-19 pneumonia: Different respiratory treatments for different phenotypes? Intensive Care Med. 2020, 46, 1099–1102. [Google Scholar] [CrossRef] [PubMed]
  114. Montazeri, M.; ZahediNasab, R.; Farahani, A.; Mohseni, H.; Ghasemian, F. Machine Learning Models for Image-Based Diagnosis and Prognosis of COVID-19: Systematic Review. JMIR Med. Informat. 2021, 9, e25181. [Google Scholar] [CrossRef] [PubMed]
  115. Sánchez-Peralta, L.; Calderón, Á.; Cabezon, V.; Ortega-Morán, J.; Sanchez-Margallo, F.; Polo, F.; Saratxaga, C.; Picon, A. Systematic acquisition and annotation of clinical cases for the generation of a medical image database. Proc. Br. J. Surg. 2019, 106, 16. [Google Scholar]
  116. Bernal, J.; Histace, A.; Masana, M.; Angermann, Q.; Sánchez-Montes, C.; de Miguel, C.R.; Hammami, M.; García-Rodríguez, A.; Córdova, H.; Romain, O.; et al. GTCreator: A flexible annotation tool for image-based datasets. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 191–201. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Illustration of meta-analysis based on published papers from January 2020 to May 2022. The systematic review included COVID-19 classification studies based on DNN approach for the CT and chest X-ray imaging dataset.
Figure 1. Illustration of meta-analysis based on published papers from January 2020 to May 2022. The systematic review included COVID-19 classification studies based on DNN approach for the CT and chest X-ray imaging dataset.
Jcm 12 03446 g001
Figure 2. PRISMA flowchart for systematic review and meta-Analysis with the keywords used in the literature review.
Figure 2. PRISMA flowchart for systematic review and meta-Analysis with the keywords used in the literature review.
Jcm 12 03446 g002
Table 1. List of selected publications for CNN classification of CT/X-ray scans.
Table 1. List of selected publications for CNN classification of CT/X-ray scans.
PublicationDatasetCNN ArchitecturePerformance Evaluation
YearStudySourceImagingClassesEnd-to-EndExplStructureTLCodeSens/Spec/PPV/NPVAUCF1Acc.
2021[11]QaTa-Cov19CXR4, BP, VP, N, Cno/DAnoCSENyesno0.98/0.95/0.64/0.99-0.770.95
2021[12]SARS CoV-2 CTCT2, C, Nno/DAOSPF-BAT FKNNyesno0.99/0.99/0.99/0.990.990.990.99
2021[13]AIIMSCXR2, C, NyesRISECovidAid_V2noyes0.88/0.94/0.88/0.94-0.880.92
2021[14]COVIDx CT-2ACT3, C, P, Nno/DACAMResNet-v2yesno0.98/0.99/0.98/0.99--0.99
2021[13]Shahid Beheshti University of MSCT2, C, Nno/FEnoNASNetnono0.99/0.98/0.99/0.99-0.990.99
2021[15]Yeungnam UniversityCT2, P, Cno/DAAMDA-CMILnono1/0.97/0.96/10.980.980.98
2021[16]COVIDxCXR3, C, P, Nno/DAnoRCoNetnono0.97/0.98/0.97/--0.970.97
2021[17]COVID-19 IDL, RSNA, CXR DICXR3, C, P, Nno/DACAMResNet50yesno0.92/0.97/0.98/-0.98-0.94
2021[18]Guangzhou, HebeiCT2, C, Nno/DAnoResUNetnono0.91/0.90/0.89/0.920.90-0.91
2021[19]Covid-ct-dataset, Guangxi Univ.CT2, C, Nno/IPHMResNet50yesno0.93/0.92/-/-0.930.920.93
2021[20]NLMMCCT2, C, Nno/DAnoGARCDyesno0.967/0.912/-/-0.98--
2021[21]Kaggle Chest X-rayCXR2,C, Nno/DAnoCOVINetnono0.98/0.96/0.98/-0.980.980.97
2021[22]6 Public datasetsCT, CXR2, C, NyesCAMMDA-BNnono0.98/0.92/0.93/0.980.980.950.95
2021[23]COVIDx, 5 US and 4 SK hospitalsCXR3, C, P, Nno/SGAMDL CBIRnono0.85/-/0.95/-0.83-0.83
2021[24]4 CH, 2 GH, ChinaCT2, C, Nno/IPnoCOVIDNetnoyes0.93/0.95/0.93/0.940.980.930.94
2021[25]Jihan Infectious Disease HospitalCT2, C, Pno/FEnoDL-MLPnono0.87/0.90/0.80/-0.920.840.89
2021[26]the Second Xiangya HospitalCT2, SC, nSCno/CRAMMILnono0.93/0.96/-/-0.980.890.95
2021[27]COVID chest X-ray, Chest X-ray14CXR2, C, CPno/IP & DACAMDenseNet-161yesno0.80/1/0.8/0.96-0.980.97
2021[28]MosMed, LUNA16CT2, C, Nno/IPCAMVGG16yesno0.97/0.82/0.79/0.97-0.890.88
2021[29]COVID-chest X-ray, CoronaHack
COVID-19 radiography
CXR3, N, P, Cno/IPCAMDenseCapsNetnono1.00/0.95/0.91/1.00-0.910.91
COVID-CTCT 0.94/0.99/0.99/0.960.960.980.98
2021[30]Covid Chest X-rayCXR2, C, Nno/DAnoConvLSTMnono0.97/0.98/0.98/0.980.810.980.95
2021[31]COVID-CTCT2, C, Nno/IPnoDNNnono0.94/0.96/0.96/0.95-0.950.95
2021[32]COVID-19 RDCXR2, C, Nno/DAnoCheXNetyesno0.99/1.00/1.00/0.99-0.990.99
SIRM, COVID-19 X-rayCXR 0.92/0.97/0.89/0.96-0.900.93
2021[33]COVID-CT, RadiopediaCT3, N, C, Pno/DAHMEATCnono0.86/0.97/0.90/0.89-0.880.87
2021[34]Covid Chest XrayCXR2, C, Nno/FEnoDWS-CNNnono0.98/0.98/-/--0.980.98
2021[35]Multi-center datasetCT2, C, NyesCAM3D-CNNyesyes0.90/0.98/0.96/0.940.880.930.88
2021[36]UCSD (California)CXR3, N, C, Pno/IPnoANNnono1.00/0.98/0.96/1.000.770.980.94
2021[37]Wuhan Pulmonary HospitalCT2, SC, nSCyesAM3D ResNetyesyes0.86/0.88/-/-0.92-0.88
2021[38]Guangzhou W&C MCCXR4, N, C, BP, VPnonoRes-CovNetyesno0.97/-/0.97/--0.980.98
SARS-COV-2CT 0.98/0.98/0.98/0.990.980.980.98
2021[39]Harvard DataverseCT2, C, NyesnoVGG-11+Inceptionv3+
WideResnet-50
yesyes0.99/0.99/.0.98/0.980.980.980.98
2021[40]Kaggle Chest X-rayCXR3, N, P, CnonoHOG+CNNnono0.96/0.99/0.94/0.990.990.990.96
CC-CCII DatasetCT 0.90/0.90/-/-0.89-0.93
2021[41]COVID-19 IDL, HwaMei HospitalCXR3, N, P, Cno/IPAMCMT-CNNnoyes0.92/0.91/-/-0.92-0.97
COVID-19 IDL, Chest X-ray-NIHCCCXR 0.99/0.99/0.99/0.99-0.990.97
2021[42]Curated X-ray DatasetCXR3, N, C, PyesnoCOVID-DeepNetnono0.97/0.98/0.97/0.980.990.990.98
2021[43]COVID-19 RD, MendeleyCXR3, C, VP, NnonoCov19-CNnetnono0.94/0.96/0.98/1.00-0.940.98
2021[44]BIMCV, COVIDx, COVID-CXNetCXR3, N, C, Pno/IPCAMFus-ResNet50nono0.95/0.99/0.94/0.99-0.950.95
2020[45]IEEE8023, COVID-CT, CORD-19CT2, C, Nno/FEnoDeepSensenono0.97/0.97/-/--0.920.98
2020[46]COVID-chestxray, CORD-19CXR2, C, Pno/SG & IPAMDarkNet-19noyes1.00/0.97/-/---0.98
2020[47]COVIDx-CTCT3, NC, NCP, CPno/CRnoCOVIDNet-CTnono0.97/0.99/0.99/0.99-0.980.99
2020[48]Honghu and Nanchang hospitalsCT3, NS, S, DPno/CRnoResNet34yesyes0.98/0.83/0.85/0.970.950.910.90
2020[49]LIDCCT2, N, Cno/CRCAMDenseNet-121noyes0.91/0.93/0.85/0.950.940.880.90
2020[50]COVID-19 CXRCT3, N, C, Pno/SG & CRHMVGG16yesyes0.91/0.93/-/-0.890.910.88
2020[51]COVID-chest X-rayCXR3, C, P, Nno/IPnoInception-V3yesyes1.00/0.99/0.99/0.99-0.990.99
COVID-chest X-rayCT 0.81/0.99/0.97/-0.950.880.96
2020[52]Daegu, South KoreaCT2, N, Cno/DA & IPnoXceptionyesno0.95/0.88/0.94/0.910.920.940.95
2020[53]COVID-19 X-rayCXR2, N, Cno/IPnoOGA-ELMnoyes0.97/0.95/0.95/0.970.980.960.96
Wuhan People’s HospitalCT 0.94/0.96/-/-0.98--
2020[54]COVIDxCXR3, C, P, NonPno/IPCAMCOVNetyesyes0.95/0.95/0.90/0.97-0.920.93
2020[55]COVID-19 CTCT2, C, Nno/UNet SGnoDensNet-201yesyes0.87/0.95/-/-0.970.920.92
2020[56]COVID chest X-rayCXR3, C, N, Pno/IPnoConvNetyesno0.94/0.99/0.93/0.990.960.940.98
2020[57]COVID chest X-rayCXR3, N, P, Cno/DAnoVGG-16nono0.94/0.94/1.00/0.84-0.970.91
2020[58]PACS Union HospitalCT2, C, NyesCAMDeCoVNetnoyes0.90/0.91/0.84/0.980.950.870.90
Table 2. CT/X-ray public datasets and state-of-the-art performance.
Table 2. CT/X-ray public datasets and state-of-the-art performance.
LinkNameOriginTypeResolutionN. PatientsClassesSample SizeH.Per.CNN Model
[68]CC-CCIINational hospitals (China)CT512 × 51227423; C, P, N411,5290.93 CMT-CNN [41]
[74]MosMedDataMunicipal hospitals in Moscow (Russia)CT512 × 51211104; ML, MD, SC, CC, N-0.88ED-VGG16 [28]
[79]COVID-CTmedRxiv and bioRxiv (USA)CT1853 × 14852162; C, N8120.98ConvLSTM [30]
[75]SARS-COV-2 CT ScanSao Paulo hospitals (Brazil)CT327  ×  30712522; C, N24810.99PF-BAT FKNN [12]
[76]Harvard DataverseHSPM (Brazil),
Metropolitan Hospital of Lapa (Brazil)
CT-2103; C, N, OLI41730.98VGG-11+Inceptionv3+
WideResnet-50 [39]
[80]LIDC-IDRICombined: NCI (Malaysia),
FNIH and FDA (USA)
CT512 × 51210104; nonP, CAP, Infl, C244,5270.90DenseNet-121 [49]
[47]COVIDx CT 1CNCB (China),
ITAC (Canada),
LIDC-IDRI,
Radiopaedia (Australia)
CT512 × 51237453; N, CP, nonCP194,9220.99ResNet-v2 [14]
[81]iCTCFHUST-UH/HUST-LH (China)CT512 × 51211703; SC, nSC, N256,3560.98GARCD [20]
[82]COVIDx-CTCNCB (China),
ITAC (Canada),
LIDC-IDRI,
Radiopaedia (Australia)
CT512 × 51245013; N, CP, nonCP201,1030.99ResNet-v2 [14]
[91]BIMCV-COVID19Medical Imaging Databank
in Valencian Region MIB (Spain)
CT
CXR
-13543; UD, N, P23,5270.95MDA-BN [22]
[84]COVID-19 CXRBMJ, Doctors Without Borders,
Mount Sinai Health System (Canada)
CXR1594 × 1600483; C, P, N550.96Xception [52]
[85]COVID-19 RDSIRM (Italy), COVID-19 IDC (Canada),
PadChest (Spain), RSNA (USA),
COVID-CXNet
CXR299 × 299-3; C, N, VP36150.99CheXNet [32]
[86]ChestX-ray8National Institute of
Health Clinical Center (USA)
CXR1024 × 102430,8053; C, P, N112,1200.99COVID-DeepNet [42]
[87]OCTaCXRIGuangzhou Women and Children’s
Medical Center (China)
CXRdifferent-2; P, N58560.98Res-CovNet [38]
[92]Curated X-Ray
Dataset
Indian Institute of Science,
PES Uni., Ramaiah IT (India),
Concordia Uni. (Canada)
CXR479 × 479-4; C, N, VP, BP92080.98DenseNet [93]
[94]
[84,88]
CXR (COVID-19
& Pneumonia)
COVID-19 IDC (Canada),
Mendeley (UK),
COVID-19 CXR (Canada)
CXR386 × 386-3; C, P, N64320.96HOG+CNN [40]
[77]Open-i CXRIndiana University Hospital (USA)CXR--2; CovidP, OP74700.97DenseNet-161 [27]
[89]CTBIMCV (Spain),
RSNA PDC (USA),
COVID-19 RD (Bangladesh)
CXR224 × 224-4; C, N, VP, BP120,9680.95CSEN [11]
[86]RSNA PDCRadiological Society of
North America (USA)
CXR-11,2543; N, NLO, LO266840.94ResNet50 [17]
[83]COVIDx CT 2SIRM (Italy), COVID-19 IDC (Canada),
Radiopaedia (Australia),
COVID-19 CXR (Canada),
Hannover Uni. (Germany)
CXR859 × 73045013; CP, CAP, N201,1030.95Fus-ResNet50 [44]
[84,88]
[84,95]
[86]
COVIDxCOVID-19 IDC (Canada), MIDRC (USA),
Actualmed (Spain), COVID-19 RD,
RSNA PDC (USA), COVID-19 CXR (Canada)
CXR-151002; C, NC16,0000.97RCoNet [16]
[88]COVID-19 IDCRadiopaedia (Australia), SIRM (Italy),
RSNA PDC (USA), Eurorad (Germany),
Coronacases (China), COVID-19 CXR (Canada)
CXR604 × 4994122; C, P6790.98DWS-CNN [34]
[90]COVID-19 PL CXR IunreportedCXR--2; C, N980.97COVINet [96]
[90]CDGCunreportedCXR--2; VP, BP790.97COVINet [96]
[78]SIRMRadiology Hospitals (Italy)CXR356 × 3381151; C4500.93DAM [33]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, M.-H.; Shomanov, A.; Kudaibergenova, M.; Viderman, D. Deep Learning Methods for Interpretation of Pulmonary CT and X-ray Images in Patients with COVID-19-Related Lung Involvement: A Systematic Review. J. Clin. Med. 2023, 12, 3446. https://doi.org/10.3390/jcm12103446

AMA Style

Lee M-H, Shomanov A, Kudaibergenova M, Viderman D. Deep Learning Methods for Interpretation of Pulmonary CT and X-ray Images in Patients with COVID-19-Related Lung Involvement: A Systematic Review. Journal of Clinical Medicine. 2023; 12(10):3446. https://doi.org/10.3390/jcm12103446

Chicago/Turabian Style

Lee, Min-Ho, Adai Shomanov, Madina Kudaibergenova, and Dmitriy Viderman. 2023. "Deep Learning Methods for Interpretation of Pulmonary CT and X-ray Images in Patients with COVID-19-Related Lung Involvement: A Systematic Review" Journal of Clinical Medicine 12, no. 10: 3446. https://doi.org/10.3390/jcm12103446

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop