Next Article in Journal
Drug-Induced Acute Pancreatitis in Hospitalized COVID-19 Patients
Next Article in Special Issue
Diagnostic Accuracy of Machine Learning AI Architectures in Detection and Classification of Lung Cancer: A Systematic Review
Previous Article in Journal
Salivary ATR-FTIR Spectroscopy Coupled with Support Vector Machine Classification for Screening of Type 2 Diabetes Mellitus
Previous Article in Special Issue
Developing a Deep-Learning-Based Coronary Artery Disease Detection Technique Using Computer Tomography Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Neural Network Augments Performance of Junior Residents in Diagnosing COVID-19 Pneumonia on Chest Radiographs

1
Institute of High Performance Computing (IHPC), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #16-16 Connexis, Singapore 138632, Singapore
2
Department of Diagnostic Radiology, Tan Tock Seng Hospital, 11, Jalan Tan Tock Seng, Singapore 308433, Singapore
3
Lee Kong Chian School of Medicine, 11, Mandalay Road, Singapore 308232, Singapore
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(8), 1397; https://doi.org/10.3390/diagnostics13081397
Submission received: 2 March 2023 / Revised: 5 April 2023 / Accepted: 7 April 2023 / Published: 12 April 2023

Abstract

:
Chest X-rays (CXRs) are essential in the preliminary radiographic assessment of patients affected by COVID-19. Junior residents, as the first point-of-contact in the diagnostic process, are expected to interpret these CXRs accurately. We aimed to assess the effectiveness of a deep neural network in distinguishing COVID-19 from other types of pneumonia, and to determine its potential contribution to improving the diagnostic precision of less experienced residents. A total of 5051 CXRs were utilized to develop and assess an artificial intelligence (AI) model capable of performing three-class classification, namely non-pneumonia, non-COVID-19 pneumonia, and COVID-19 pneumonia. Additionally, an external dataset comprising 500 distinct CXRs was examined by three junior residents with differing levels of training. The CXRs were evaluated both with and without AI assistance. The AI model demonstrated impressive performance, with an Area under the ROC Curve (AUC) of 0.9518 on the internal test set and 0.8594 on the external test set, which improves the AUC score of the current state-of-the-art algorithms by 1.25 % and 4.26 % , respectively. When assisted by the AI model, the performance of the junior residents improved in a manner that was inversely proportional to their level of training. Among the three junior residents, two showed significant improvement with the assistance of AI. This research highlights the novel development of an AI model for three-class CXR classification and its potential to augment junior residents’ diagnostic accuracy, with validation on external data to demonstrate real-world applicability. In practical use, the AI model effectively supported junior residents in interpreting CXRs, boosting their confidence in diagnosis. While the AI model improved junior residents’ performance, a decline in performance was observed on the external test compared to the internal test set. This suggests a domain shift between the patient dataset and the external dataset, highlighting the need for future research on test-time training domain adaptation to address this issue.

1. Introduction

The Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) outbreak [1], initially detected in Wuhan, Hubei, China, has rapidly escalated into a worldwide pandemic [2]. The National Centre for Infectious Diseases (NCID) [3] has been at the forefront of Singapore’s COVID-19 response. As of the time of writing, the Ministry of Health (MOH) has recorded over 2.2 million confirmed cases in Singapore [4].
Patients with COVID-19 pneumonia often exhibit similar symptoms to other viral diseases, such as Middle East Respiratory Syndrome [5], and their imaging findings are often non-specific, presenting a diagnostic challenge [6,7,8]. Currently, the definitive method for diagnosing COVID-19 infection is the reverse transcriptase polymerase chain reaction (RT-PCR) [9]; however, limitations in diagnostic testing resources can hinder its accuracy [10,11,12]. In this context, chest imaging techniques [13], including computed tomography (CT) and chest radiography (CXR), are essential in the context of patient triaging and making treatment decisions [12,14]. Despite being less sensitive than CT, CXR is more widely adopted as it is faster, exposes patients to lower levels of radiation, and is potentially more cost-effective [7,15,16]. Therefore, junior residents, who are often the first point of contact, are expected to interpret CXRs of COVID-19 patients in many institutions.
In recent years, the application of deep learning models to clinical problems has shown significant potential in facilitating auto-diagnosis of diseases and providing real-time procedural support, particularly in the field of healthcare [17,18,19]. Several studies have explored the use of deep learning models for diagnosing COVID-19 through analysis of chest X-rays [20,21]. For instance, Jordan et al. collected a CXR dataset from NCID to train an AI model with the DenseNet as backbone to detect COVID-19 pneumonia. In response to the rapidly evolving global pandemic during COVID-19, they quickly deployed the trained model to NCID [22]. Abul et al. adopted a convolutional neural network (CNN) to extract feature representations from CXRs, and connected it with various classifiers, such as support vector machine (SVM), pattern recognition network (PRN), decision tree (DT), random forest (RF), and k-nearest neighbours (KNN), to perform COVID-19 detection [23]. Linda et al. introduced a novel COVID-Net for diagnosing COVID-19 from CXRs in publicly available datasets [24]. This approach incorporates an explainability method to not only provide clinicians with a deeper understanding of the critical factors associated with COVID-19 cases and improved screening, but also to enhance the transparency and accountability of COVID-Net by ensuring that its decisions are based on relevant information extracted from CXR images. However, the diagnostic performance of these models has yet to be validated in a clinical setting due to the limited size of CXR datasets used for training [24,25,26]. Moreover, it has been reported that lab-trained models may experience a significant decline in performance when deployed in clinical practice [27].
The motivation of this research is to harness AI’s potential to improve the diagnostic process, optimize healthcare resources, and ultimately enhance patient care. Specifically, we develop a deep neural network to differentiate COVID-19 from other forms of pneumonia and explore the potential of AI techniques to enhance the diagnostic accuracy of junior residents, who often serve as the primary point-of-contact in the diagnostic process. In order to achieve the objectives of this study, an interdisciplinary approach has been adopted that combines clinical research, image diagnostics, and AI models. The primary aim of the study is to collect structured data, including CXRs and RT-PCR results, and use these data to develop a three-class classification AI model that can accurately distinguish between non-pneumonia, non-COVID-19 pneumonia, and COVID-19 pneumonia cases. Subsequently, the study seeks to assess the effectiveness of the AI model in enhancing the diagnostic precision of novice residents by implementing and evaluating its performance. Specifically, we collected a patient dataset and used it to train and validate the AI model. Then, the trained AI model has been deployed in the Tan Tock Seng Hospital (TTSH), where its effectiveness in improving junior residents’ diagnostic accuracy has been evaluated. In addition, the study has investigated the performance of junior residents with different levels of training, both with and without the assistance of the deployed AI model.

2. Materials and Methods

2.1. Dataset

The dataset utilized in this work comprised a total of 5051 CXRs obtained from the NCID Screening Centre and Tan Tock Seng Hospital (TTSH) in Singapore. CXRs that were conducted between February 2020 and early April 2020 were included in the dataset. Two senior radiologists, each with over 15 years of experience, annotated the class label of the CXRs as pneumonia or non-pneumonia and used them as the reference standard. All CXRs were reviewed independently by senior radiologists who were unaware of any clinical details.
Patients were classified into three groups based on their clinical presentation and diagnostic results. Patients were classified as positive for COVID-19 pneumonia if they tested positive on the PCR test and had a CXR positive for pneumonia. Patients were classified as having non-COVID-19 pneumonia if their CXR was positive for pneumonia, but they tested negative on the PCR test. The remaining patients were grouped to the non-pneumonia class.
The dataset contained a total of 607 COVID-19 pneumonia cases, 570 cases of non-COVID-19 pneumonia (including viral, bacterial, and fungal pneumonia), and 3874 non-pneumonia cases. To form the training, validation, and test sets, CXRs were chosen randomly from each category within the dataset. These three sets comprised 70%, 10%, and 20% of the total data, respectively. Table 1 provides a summary of the dataset split statistics.

2.2. External Test Set

In order to assess the potential of AI assistance to improve the diagnostic accuracy of junior residents, a separate dataset consisting of 500 CXRs were used. This external test set was reviewed by three junior residents with varying levels of training, both with and without the assistance of the developed AI model.
The external dataset was also obtained from the same institution but was collected during a different time period. It included 72 cases of COVID-19 pneumonia, 49 cases of non-COVID-19 pneumonia, and 379 cases of non-pneumonia, as detailed in Table 1.

2.3. Neural Network Architecture and Training Strategy

The workflow of the AI model-aided diagnosis is presented in Figure 1. Initially, we developed an end-to-end CNN-based framework [28] to classify CXRs as non-pneumonia, non-COVID-19 pneumonia, or COVID-19 pneumonia. In the second stage, three junior residents, each with varying levels of training, were enlisted to review the external test set both with and without AI assistance. During the AI-assisted review, the junior residents were provided with the probability output and relevant heatmap for each case, generated by the trained model.
To extract the features of the input CXRs, we employed an EffcientNet-b7 [29] model as the backbone, replacing the output layer with one that includes three neurons with the softmax activation function to output the final predictions. The architecture of the EfficientNet-b7, comprising Block1 to Block7, Stem, and Final Layers, is illustrated in Figure 2. Given the imbalanced patient dataset, we minimized the weighted cross-entropy loss (WCEL) to optimize the AI model as follows:
L = i = 1 N j = 1 C w j y i j log ( p i j ) ,
where N is the number of samples in the dataset, C is the number of classes, w j is the weight assigned to class j, y i j is the true label for class j of sample i, and p i j is the predicted probability of sample i belonging to class j obtained from the output of the model. In this work, the weight for each class is w j = N N j with N j as the number of samples in class j.
We trained, validated, and tested our AI model on the internal dataset, choosing the model that displayed the top AUC score during validation to generate probability scores and heatmaps for use in our experiment.

2.4. AI Model Deployment & Diagnosis

In this study, the trained model was executed on a virtual machine running Ubuntu 18.04, which was hosted on a Windows 10 workstation equipped with two Nvidia 2080Ti GPUs, located within the Department of Diagnostic Radiology. In adherence to the hospital’s security guidelines, formal authorization was secured from the Integrated Health Information Systems (IHiS) committee in Singapore, which is in charge of managing the IT risk and security for Tan Tock Seng Hospital.
To evaluate the impact of our AI model in enhancing the performance of junior residents, we recruited three junior residents with varying levels of training for the study. The residents independently reviewed the external dataset of 500 CXRs, initially without the aid of the AI model. After a deliberate 3-month hiatus, the same dataset of 500 CXRs was reviewed by all three junior residents, this time with the aid of AI in the form of probability outputs and relevant heatmaps.

3. Results

3.1. Verify on AI Model

The effectiveness of the developed AI model was verified by comparing it with five deep learning methods [6,30,31,32,33] designed for COVID-19 diagnosis using CT scans and CXRs. The performance of our method and the five peer methods for the test set and the external test set in terms of AUC scores (with the macro-averaging strategy) and 95% confidence interval (95% CI) are reported in Table 2, which show that our model outperforms the other five methods significantly, achieving an AUC score of 0.9520 (95% CI: 0.9479–0.9585) for the internal test set and 0.8588 (95% CI: 0.8570–0.8623) for the external test set. Specifically, it improves the AUC score of the peer methods by 1.25 % and 4.26 % for the test set and the external test set, respectively. This evidence suggests that the proposed AI model demonstrates efficacy in diagnosing pneumonia. All the tested methods demonstrated significant performance degradation on the external test, potentially due to a data distribution mismatch between the TTSH dataset (training, validation, and test set) and the external test set.
The AUC score of our AI model for each of the three classes (non-pneumonia, non-COVID-19 pneumonia, and COVID-19 pneumonia) is reported in Table 3. Our AI model achieved higher AUC scores, sensitivity, and specificity for each class on both the internal and external test sets, outperforming the second-best method (CV19-Net) by a large margin. Specifically, our AI model improves the AUC score of the second-best performing algorithms by 1.63 % and 3.40 % for the test set and the external test set on COVID-19 pneumonia, respectively. Moreover, our AI model improves the sensitivity and specificity of the peer methods by 1.39 % and 7.24 % for the external test set on COVID-19 pneumonia, respectively. The comparison of ROC curves is presented in Figure 3. Additionally, examples of heatmaps from Grad-CAM [34] and predictions generated by our model are shown in Figure 4. Our proposed model is capable of generating a set of three images for reference, pertaining to a given case of interest, such as JRs. The first image in the sequence corresponds to the original CXR, which is augmented with blue circles indicating the suspicious area showing signs of possible COVID-19 infection, computed by the AI model. The second image displays the heatmaps obtained from the model’s predictions, which are overlaid onto the original CXRs. The regions highlighted by the heatmaps correspond to the anatomical areas that exert the greatest impact on the final model predictions. The final image in the set presents the predicted probabilities for each class, thereby providing an integrated summary of the model’s diagnostic accuracy.

3.2. AI-Aided Diagnosis

In order to investigate the extent to which the proposed AI model improves the performance of junior residents (JRs), we conducted a comparison of the performances of the JRs with and without AI assistance, as presented in Table 4. The JRs who participated in this study possessed varying levels of expertise. Specifically, JR1, JR2, and JR3 had approximately 6 months, 1 year, and more than 2 years of experience in interpreting CXRs, respectively. The results indicate that the performance of the junior residents improved in proportion to their level of training. Even without AI assistance, the JR with the most experience (JR3) achieved an AUC score of 0.8657 (95% CI: 0.8633–0.8676), while the JR with the least experience (JR1) attained an AUC score of 0.7813 (95% CI: 0.7785–0.7827). Following AI augmentation, we observed improvements for both JR1 and JR2, achieving AUC scores of 0.8482 (95% CI: 0.8452–0.8511) and 0.8511 (95% CI: 0.8493–0.8526), respectively. Additionally, Table 4 presents Cohen’s kappa score for each JR, with the scores of JR1 (0.5574) and JR2 (0.4651) being smaller than that of JR3 (0.7400), indicating that the AI model had a greater impact on JR1 and JR2 compared to JR3. The detailed performance of all JRs before and after AI assistance is presented in Table 5. From Table 5 we can find that: with the AI model’s assistance, the JR1’s sensitivity has been improved from 0.3889 to 0.6250 on COVID-19 pneumonia diagnosis, the specificity has been improved from 0.7317 to 0.9002 on non-COVID-19 pneumonia, the AUC score has been improved from 0.8121 to 0.8417 and the specificity has been improved from 0.9091 to 0.9339 on non-pneumonia. JR2’s sensitivity has been improved from 0.5000 to 0.5833 on COVID-19 pneumonia diagnosis, the specificity has been improved from 0.8226 to 0.8514 on non-COVID-19 pneumonia, and the specificity has been improved from 0.9008 to 0.9835 on non-pneumonia. Notably, even JR3 with higher training level, the AI model was found to improve sensitivity, from 0.8681 to 0.8902, demonstrating the broad applicability and effectiveness of the model.

4. Discussion

Several recent studies have highlighted the potential of chest CT in diagnosing COVID-19 pneumonia, particularly in its early stages [35,36]. However, in healthcare systems with limited resources, CT may not always be available, and CXR is the more commonly used imaging method in clinical practice. Additionally, concerns have been raised by international workgroups regarding the untested specificity of CT in cases where the pre-test probability of COVID-19 infection is low, and CXR is preferred to reduce the risk of nosocomial transmission [37,38,39]. Recent reports have demonstrated that CXR findings correlate well with clinical severity and can be used to predict severe pneumonia [24]. Therefore, the development of AI models that aid in diagnosing COVID-19 pneumonia using CXRs remains an area of active research and is of significant importance.
Junior residents are often the first healthcare professionals to interpret CXRs for suspected COVID-19 patients in clinical practice. However, interpreting CXRs can be challenging, especially when the radiological report may impact patient disposition. This study presents an AI model capable of performing three-class classification for non-pneumonia, non-COVID-19 pneumonia, and COVID-19 pneumonia. The proposed model demonstrates superiority in AUC score, sensitivity, and specificity compared to five peer methods. Although all methods experience performance degradation when tested on the external test set, our model performs better, achieving an AUC score of 0.8594 (95% CI: 0.8594–0.8602), and outperforming the second-best method by a large margin [33].
This study provides compelling evidence for the potential of a trained AI model to enhance the diagnostic performance of junior residents in the field of chest radiography by providing probability outputs and heatmaps. Although previous studies have investigated the augmentation of radiologists’ performance using AI in distinguishing COVID-19 from other types of pneumonia on chest CTs [40], this study specifically focuses on the performance of junior residents, which may have significant implications for both diagnostic imaging and education. The participating JRs possessed varying levels of expertise, with JR1, JR2, and JR3 having approximately 6 months, 1 year, and more than 2 years of experience in interpreting CXRs, respectively. The findings suggest that the performance of the junior residents improved in proportion to their level of training. Furthermore, in addition to the quantitative improvements observed, all three junior residents reported increased confidence when using AI assistance, suggesting that AI may have positive qualitative outcomes as well.
In this study, the adoption of EfficientNet as the backbone for diagnosing COVID-19 pneumonia, and non-COVID-19 pneumonia from CXRs was proposed. Although the model achieved promising performance for the test set and improved decision-making for junior residents, several limitations were noted. Specifically, a significant decrease in performance was observed on the external test set, which is not unique to our method but common in medical image analysis due to domain shift problems [41,42,43]. The training process did not account for differences in data distributions between the TTSH dataset and the external test set, leading to a drop in performance. Medical datasets are often drawn from different domains, even within the same institution, which can pose challenges for deep learning models that assume similar data distributions between training and test sets [43]. Deep learning typically assumes that both the distributions of the training and test sets are similar [44]. If the distributions of both datasets are dissimilar, the performance may drop dramatically. In future work, domain adaption techniques will be explored to align the distributions of different datasets for automated disease diagnoses from CXRs. Furthermore, a purposeful 3-month hiatus was implemented, during which three JRs were tasked with interpreting the same dataset of 500 CXRs with and without AI assistance. Over this three-month period, the diagnostic abilities of the JRs may have improved, potentially influencing the accuracy enhancement achieved when using AI as an assistant.
In summary, our proposed method exhibited a high degree of accuracy in performing a three-class classification of non-pneumonia, non-COVID-19 pneumonia, and COVID-19 pneumonia. Furthermore, the method was observed to improve the diagnostic performance of junior residents in identifying COVID-19 pneumonia on CXRs, a potentially valuable asset in triage and education during the ongoing pandemic.

5. Conclusions

This study proposes an AI model that utilizes EfficientNet-b7 [29] as the backbone to differentiate non-pneumonia, non-COVID-19 pneumonia, and COVID-19 pneumonia from CXRs. The training dataset was collected from Tan Tock Seng Hospital and was annotated by experienced senior radiologists. This study showcases the innovative creation of an AI model for three-category CXR classification, emphasizing its ability to enhance the diagnostic precision of less experienced residents. Validated with external data, the model demonstrates practical relevance in real-world scenarios. Upon deployment, the AI model demonstrated its ability to assist junior residents in CXR interpretation and increase their confidence in diagnostic accuracy. Although the proposed AI model improved the performance of junior residents, a performance drop was observed on the external test in comparison to the results on the test set. It indicated that there has been a domain shift between the collected patient dataset (including the training, validation, and test set) and the external dataset. Domain shift is a common issue in practical applications of AI, especially in clinical practice. The clinical practice presents much more heterogeneous acquisition conditions [17] which may lead to the performance of the trained model on external test datasets (or a dataset it encounters when deployed) degrading.
To address the domain gap issue between the training dataset and the external test dataset, we plan to investigate and develop a test-time training domain adaptation model in future work. This model will aim to align the tested CXR in real time with the training data, which is expected to enhance the model’s performance after deployment. By aligning the test CXR with the training dataset, the test-time training domain adaptation model is expected to be more effective and mitigate performance degradation to a large extent. Moreover, it is expected to be more robust to the changing acquisition conditions during CXR screening.

Author Contributions

Conceptualization, X.X.; methodology, Y.F. and J.S.Z.T.; software, W.-X.C.; validation, Y.W., C.B.K., E.O.T.E. and H.I.T.W.J.; formal analysis, Z.W. and L.Z.; investigation, Y.C.; resources, Y.T.; data curation, X.L.; writing—original draft preparation, Y.F.; writing—review and editing, J.S.Z.T.; visualization, S.L.; supervision, C.H.T.; project administration, R.S.M.G.; funding acquisition, Y.L. and R.S.M.G. All authors have reviewed and approved the published version of the manuscript.

Funding

This research was supported by A*STAR through its AME Programmatic Funding Scheme Under Project grant number A20H4b0141.

Institutional Review Board Statement

This research obtained approval from the relevant organizations’ ethics committees (Approval code: 2017/00683-AMD0005) and adhered to the Health Insurance Portability and Accountability Act regulations. Owing to the study’s retrospective design and low associated risks, permission to waive consent was granted.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are not publicly available due to privacy restrictions.

Conflicts of Interest

All authors declared no conflict of interest.

References

  1. Lai, C.C.; Shih, T.P.; Ko, W.C.; Tang, H.J.; Hsueh, P.R. Severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) and coronavirus disease-2019 (COVID-19): The epidemic and the challenges. Int. J. Antimicrob. Agents 2020, 55, 105924. [Google Scholar] [CrossRef] [PubMed]
  2. Zu, Z.Y.; Jiang, M.D.; Xu, P.P.; Chen, W.; Ni, Q.Q.; Lu, G.M.; Zhang, L.J. Coronavirus disease 2019 (COVID-19): A perspective from China. Radiology 2020, 296, E15–E25. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Facilities and Services, National Centre for Infectious Diseases (NCID). Available online: https://www.ncid.sg/Facilities-Services/Pages/default.aspx (accessed on 21 October 2022).
  4. UPDATES ON COVID-19 (CORONAVIRUS DISEASE 2019) LOCAL SITUATION, Ministry of Health. Available online: https://www.moh.gov.sg/COVID-19/statistics (accessed on 26 February 2023).
  5. Kooraki, S.; Hosseiny, M.; Myers, L.; Gholamrezanezhad, A. Coronavirus (COVID-19) outbreak: What the department of radiology should know. J. Am. Coll. Radiol. 2020, 17, 447–451. [Google Scholar] [CrossRef]
  6. Zhang, R.; Tie, X.; Qi, Z.; Bevins, N.B.; Zhang, C.; Griner, D.; Song, T.K.; Nadig, J.D.; Schiebler, M.L.; Garrett, J.W.; et al. Diagnosis of COVID-19 Pneumonia Using Chest Radiography: Value of Artificial Intelligence. Radiology 2020, 24, 202944. [Google Scholar]
  7. Pereira, R.M.; Bertolini, D.; Teixeira, L.O.; Silla, C.N., Jr.; Costa, Y.M. COVID-19 identification in chest X-ray images on flat and hierarchical classification scenarios. Comput. Methods Programs Biomed. 2020, 8, 105532. [Google Scholar] [CrossRef]
  8. Rahimzadeh, M.; Attar, A. A modified deep convolutional neural network for detecting COVID-19 and pneumonia from chest X-ray images based on the concatenation of Xception and ResNet50V2. Inform. Med. Unlocked 2020, 26, 100360. [Google Scholar] [CrossRef]
  9. Khan, I.U.; Aslam, N.; Anwar, T.; Alsaif, H.S.; Chrouf, S.M.B.; Alzahrani, N.A.; Alamoudi, F.A.; Kamaleldin, M.M.A.; Awary, K.B. Using a deep learning model to explore the impact of clinical data on COVID-19 diagnosis using chest X-ray. Sensors 2022, 22, 669. [Google Scholar] [CrossRef] [PubMed]
  10. Ai, T.; Yang, Z.; Hou, H.; Zhan, C.; Chen, C.; Lv, W.; Tao, Q.; Sun, Z.; Xia, L. Correlation of chest CT and RT-PCR testing in coronavirus disease 2019 (COVID-19) in China: A report of 1014 cases. Radiology 2020, 26, 200642. [Google Scholar] [CrossRef] [Green Version]
  11. Stephanie, S.; Shum, T.; Clevel, H.; Challa, S.R.; Herring, A.; Jacobson, F.L.; Hatabu, H.; Byrne, S.C.; Shashi, K.; Araki, T.; et al. Determinants of Chest X-Ray Sensitivity for COVID-19: A Multi-Institutional Study in the United States. Radiol. Cardiothorac. Imaging 2020, 2, e200337. [Google Scholar] [CrossRef]
  12. Rubin, G.D.; Ryerson, C.J.; Haramati, L.B.; Sverzellati, N.; Kanne, J.P.; Raoof, S.; Schluger, N.W.; Volpi, A.; Yim, J.J.; Martin, I.B.; et al. The role of chest imaging in patient management during the COVID-19 pandemic: A multinational consensus statement from the Fleischner Society. Radiology 2020, 296, 172–180. [Google Scholar] [CrossRef] [Green Version]
  13. Hayden, G.E.; Wrenn, K.W. Chest radiograph vs. computed tomography scan in the evaluation for pneumonia. J. Emerg. Med. 2009, 36, 266–270. [Google Scholar] [CrossRef]
  14. Li, Y.; Xia, L. Coronavirus disease 2019 (COVID-19): Role of chest CT in diagnosis and management. Am. J. Roentgenol. 2020, 214, 1280–1286. [Google Scholar] [CrossRef] [PubMed]
  15. Gaur, L.; Bhatia, U.; Jhanjhi, N.Z.; Muhammad, G.; Masud, M. Medical image-based detection of COVID-19 using deep convolution neural networks. Multimed. Syst. 2021, 28, 1–10. [Google Scholar] [CrossRef]
  16. Wong, H.Y.; Lam, H.Y.; Fong, A.H.; Leung, S.T.; Chin, T.W.; Lo, C.S.; Lui, M.M.; Lee, J.C.; Chiu, K.W.; Chung, T.; et al. Frequency and distribution of chest radiographic findings in COVID-19 positive patients. Radiology 2020, 27, 201160. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Feng, Y.; Wang, Z.; Xu, X.; Wang, Y.; Fu, H.; Li, S.; Zhen, L.; Lei, X.; Cui, Y.; Ting, J.S.; et al. Contrastive domain adaptation with consistency match for automated pneumonia diagnosis. Med. Image Anal. 2023, 83, 102664. [Google Scholar] [CrossRef]
  18. Wang, Y.; Feng, Y.; Zhang, L.; Zhou, J.T.; Liu, Y.; Goh, R.S.; Zhen, L. Adversarial multimodal fusion with attention mechanism for skin lesion classification using clinical and dermoscopic images. Med. Image Anal. 2022, 81, 102535. [Google Scholar] [CrossRef] [PubMed]
  19. Feng, Y.; Xu, X.; Wang, Y.; Lei, X.; Teo, S.K.; Sim, J.Z.; Ting, Y.; Zhen, L.; Zhou, J.T.; Liu, Y.; et al. Deep supervised domain adaptation for pneumonia diagnosis from chest x-ray images. IEEE J. Biomed. Health Inform. 2021, 26, 1080–1090. [Google Scholar] [CrossRef]
  20. El-Rashidy, N.; Abdelrazik, S.; Abuhmed, T.; Amer, E.; Ali, F.; Hu, J.W.; El-Sappagh, S. Comprehensive survey of using machine learning in the COVID-19 pandemic. Diagnostics 2021, 11, 1155. [Google Scholar] [CrossRef]
  21. Hertel, R.; Benlamri, R. Deep learning techniques for COVID-19 diagnosis and prognosis based on radiological imaging. ACM Comput. Surv. 2023, 55, 1–39. [Google Scholar] [CrossRef]
  22. Sim, J.Z.; Ting, Y.H.; Tang, Y.; Feng, Y.; Lei, X.; Wang, X.; Chen, W.X.; Huang, S.; Wong, S.T.; Lu, Z.; et al. Diagnostic performance of a deep learning model deployed at a national COVID-19 screening facility for detection of pneumonia on frontal chest radiographs. Healthcare 2022, 10, 175. [Google Scholar] [CrossRef]
  23. Azad, A.K.; Ahmed, I.; Ahmed, M.U. In Search of an Efficient and Reliable Deep Learning Model for Identification of COVID-19 Infection from Chest X-ray Images. Diagnostics 2023, 13, 574. [Google Scholar] [CrossRef]
  24. Wang, L.; Lin, Z.Q.; Wong, A. Covid-net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef] [PubMed]
  25. Narin, A.; Kaya, C.; Pamuk, Z. Automatic detection of coronavirus disease (COVID-19) using x-ray images and deep convolutional neural networks. Pattern Anal. Appl. 2021, 24, 1207–1220. [Google Scholar] [CrossRef]
  26. Zhang, J.; Xie, Y.; Li, Y.; Shen, C.; Xia, Y. Covid-19 screening on chest x-ray images using deep learning based anomaly detection. arXiv 2020, arXiv:2003.12338. [Google Scholar]
  27. Kitamura, G.; Deible, C. Retraining an open-source pneumothorax detecting machine learning algorithm for improved performance to medical images. Clin. Imaging 2020, 61, 15–19. [Google Scholar] [CrossRef] [PubMed]
  28. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, UK, 2016; Volume 1, No. 2. [Google Scholar]
  29. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  30. Ahuja, S.; Panigrahi, B.K.; Dey, N.; Rajinikanth, V.; Gandhi, T.K. Deep transfer learning-based automated detection of COVID-19 from lung CT scan slices. Appl. Intell. 2021, 51, 571–585. [Google Scholar] [CrossRef]
  31. Panwar, H.; Gupta, P.K.; Siddiqui, M.K.; Morales-Menendez, R.; Singh, V. Application of deep learning for fast detection of COVID-19 in X-Rays using nCOVnet. Chaos Solitons Fractals 2020, 138, 109944. [Google Scholar] [CrossRef]
  32. Vaid, S.; Kalantar, R.; Bhandari, M. Deep learning COVID-19 detection bias: Accuracy through artificial intelligence. Int. Orthop. 2020, 44, 1539–1542. [Google Scholar] [CrossRef]
  33. Apostolopoulos, I.D.; Mpesiana, T.A. COVID-19: Automatic detection from X-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020, 43, 635–640. [Google Scholar] [CrossRef] [Green Version]
  34. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  35. Kang, H.; Xia, L.; Yan, F.; Wan, Z.; Shi, F.; Yuan, H.; Jiang, H.; Wu, D.; Sui, H.; Zhang, C.; et al. Diagnosis of coronavirus disease 2019 (COVID-19) with structured latent multi-view representation learning. IEEE Trans. Med. Imaging 2020, 39, 2606–2614. [Google Scholar] [CrossRef]
  36. Han, R.; Huang, L.; Jiang, H.; Dong, J.; Peng, H.; Zhang, D. Early clinical and CT manifestations of coronavirus disease 2019 (COVID-19) pneumonia. AJR Am. J. Roentgenol. 2020, 215, 338–343. [Google Scholar] [CrossRef] [PubMed]
  37. Revel, M.P.; Parkar, A.P.; Prosch, H.; Silva, M.; Sverzellati, N.; Gleeson, F.; Brady, A.; European Society of Radiology (ESR) and the European Society of Thoracic Imaging (ESTI). COVID-19 patients and the Radiology department—Advice from the European Society of Radiology (ESR) and the European Society of Thoracic Imaging (ESTI). Eur. Radiol. 2020, 30, 4903–4909. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Nair, A.; Rodrigues, J.C.; Hare, S.; Edey, A.; Devaraj, A.; Jacob, J.; Johnstone, A.; McStay, R.; Denton, E.; Robinson, G. A British Society of Thoracic Imaging statement: Considerations in designing local imaging diagnostic algorithms for the COVID-19 pandemic. Clin. Radiol. 2020, 75, 329–334. [Google Scholar] [CrossRef] [PubMed]
  39. Skulstad, H.; Cosyns, B.; Popescu, B.A.; Galderisi, M.; Salvo, G.D.; Donal, E.; Petersen, S.; Gimelli, A.; Haugaa, K.H.; Muraru, D.; et al. COVID-19 pandemic and cardiac imaging: EACVI recommendations on precautions, indications, prioritization, and protection for patients and healthcare personnel. Eur. Heart J.-Cardiovasc. Imaging 2020, 21, 592–598. [Google Scholar] [CrossRef]
  40. Bai, H.X.; Wang, R.; Xiong, Z.; Hsieh, B.; Chang, K.; Halsey, K.; Tran, T.M.L.; Choi, J.W.; Wang, D.-C.; Shi, L.-B.; et al. AI augmentation of radiologist performance in distinguishing COVID-19 from pneumonia of other etiology on chest CT. Radiology 2020, 296, 201491. [Google Scholar] [CrossRef]
  41. Sun, B.; Feng, J.; Saenko, K. Return of frustratingly easy domain adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; Volume 30. No. 1. [Google Scholar]
  42. Kamnitsas, K.; Baumgartner, C.; Ledig, C.; Newcombe, V.; Simpson, J.; Kane, A.; Menon, D.; Nori, A.; Criminisi, A.; Rueckert, D.; et al. Unsupervised domain adaptation in brain lesion segmentation with adversarial networks. In Proceedings of the 25th International Conference on Information Processing in Medical Imaging, IPMI 2017, Boone, NC, USA, 25–30 June 2017; Springer International Publishing: Cham, Switzerland, 2017; pp. 597–609. [Google Scholar]
  43. Varsavsky, T.; Orbes-Arteaga, M.; Sudre, C.H.; Graham, M.S.; Nachev, P.; Cardoso, M.J. Test-time unsupervised domain adaptation. In Proceedings of the 23rd International Conference on Medical Image Computing and Computer Assisted Intervention–MICCAI 2020, Lima, Peru, 4–8 October 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 428–436. [Google Scholar]
  44. Wang, X.; Liang, G.; Zhang, Y.; Blanton, H.; Bessinger, Z.; Jacobs, N. Inconsistent performance of deep learning models on mammogram classification. J. Am. Coll. Radiol. 2020, 17, 796–803. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The workflow for training the AI model and conducting AI-aided diagnosis. Stage (A) involves training, validating, and testing the AI model for diagnosing CXRs. In Stage (B), the trained AI model is deployed in the hospital, and the junior residents (JRs) can upload CXRs to obtain the predictions and heatmaps generated by the AI model. By using AI results, the JRs can make more informed decisions.
Figure 1. The workflow for training the AI model and conducting AI-aided diagnosis. Stage (A) involves training, validating, and testing the AI model for diagnosing CXRs. In Stage (B), the trained AI model is deployed in the hospital, and the junior residents (JRs) can upload CXRs to obtain the predictions and heatmaps generated by the AI model. By using AI results, the JRs can make more informed decisions.
Diagnostics 13 01397 g001
Figure 2. The architecture of EffcientNet-b7 for CXRs classification. The model takes CXRs as inputs and ultimately classifies them into three categories: non-pneumonia, non-COVID-19 pneumonia, and COVID-19 pneumonia.
Figure 2. The architecture of EffcientNet-b7 for CXRs classification. The model takes CXRs as inputs and ultimately classifies them into three categories: non-pneumonia, non-COVID-19 pneumonia, and COVID-19 pneumonia.
Diagnostics 13 01397 g002
Figure 3. The ROC curve of the peer methods (i.e., Ahuja’s method [30], nCOVnet [31], Apostolopoulos’s method [33], and CV19-Net [6]) and ours on the three classes.
Figure 3. The ROC curve of the peer methods (i.e., Ahuja’s method [30], nCOVnet [31], Apostolopoulos’s method [33], and CV19-Net [6]) and ours on the three classes.
Diagnostics 13 01397 g003
Figure 4. Examples of heatmaps and predictions generated by the AI model. The left-most image displays grayscale chest radiographs with superimposed blue circles indicating the suspicious area showing signs of possible COVID-19 infection based on the AI model’s heatmaps. The middle image shows the heatmaps overlaid on the original CXRs, with the highlights indicating the anatomical regions that contribute most to the final model predictions. The right-most image displays the predicted probabilities for each class.
Figure 4. Examples of heatmaps and predictions generated by the AI model. The left-most image displays grayscale chest radiographs with superimposed blue circles indicating the suspicious area showing signs of possible COVID-19 infection based on the AI model’s heatmaps. The middle image shows the heatmaps overlaid on the original CXRs, with the highlights indicating the anatomical regions that contribute most to the final model predictions. The right-most image displays the predicted probabilities for each class.
Diagnostics 13 01397 g004
Table 1. Statistics of training, validation, test, and external test sets.
Table 1. Statistics of training, validation, test, and external test sets.
COVID-19 PneumoniaNon-COVID-19 PneumoniaNon-PneumoniaTotal
Training set42539927123536
Validation set6157387505
Test set1211147751010
External test set7249379500
Table 2. Performance comparison of the proposed model with that of peer methods for the test set and the external test set in terms of AUC scores and 95% confidence interval (95% CI).
Table 2. Performance comparison of the proposed model with that of peer methods for the test set and the external test set in terms of AUC scores and 95% confidence interval (95% CI).
Test SetExternal Test Set
AUC95% CIAUC95% CI
Ahuja’s [30]0.89820.8968–0.89930.76800.7651–0.7704
nCOVnet [31]0.88760.8854–0.88970.68370.6012–0.6859
Vaid’s [32]0.90210.8996–0.90380.74020.7379–0.7425
Apostolopoulos’s [33]0.92790.9229–0.92940.81620.8145–0.8185
CV19-Net [6]0.93950.9361–0.94070.79870.7952–0.8032
Ours0.9520 *0.9479–0.95850.8588 *0.8570–0.8623
* Denotes statistically significant (p >0.05). AUC = the area under the receiver operating characteristic.
Table 3. Performance comparison for each class on the test set and the external test set in terms of AUC scores, Sensitivity, and Specificity.
Table 3. Performance comparison for each class on the test set and the external test set in terms of AUC scores, Sensitivity, and Specificity.
Test SetExternal Test Set
AUCSensitivitySpecificityAUCSensitivitySpecificity
Ahuja’s [30]COVID-19 pneumonia0.91850.89660.77680.73090.73610.5491
Non-COVID-19 pneumonia0.88860.84210.81360.77620.77550.6408
Non-pneumonia0.89640.84290.80000.77400.77840.6116
nCOVnet [31]COVID-19 pneumonia0.88970.87930.73120.64370.70830.5117
Non-COVID-19 pneumonia0.88170.86390.77390.72510.73470.5322
Non-pneumonia0.88820.85960.78640.68600.72300.5124
Vaid’s [32]COVID-19 pneumonia0.90880.84480.80640.71540.75000.6005
Non-COVID-19 pneumonia0.90240.84210.85000.73870.67350.5854
Non-pneumonia0.90100.86130.80870.74510.77040.5785
Apostolopoulos’s [33]COVID-19 pneumonia0.92840.87930.84970.78560.79170.6519
Non-COVID-19 pneumonia0.92340.87720.81820.79380.73470.7251
Non-pneumonia0.92850.85860.81740.82500.78630.7438
CV19-Net [6]COVID-19 pneumonia0.93270.89660.83600.77870.81940.5958
Non-COVID-19 pneumonia0.95650.89470.80910.78820.79590.5987
Non-pneumonia0.93800.92410.82610.80380.84700.6033
OursCOVID-19 pneumonia0.94900.93100.85190.81960.83330.7243
Non-COVID-19 pneumonia0.95410.91230.85000.83480.87760.7073
Non-pneumonia0.95220.93380.82610.86940.89180.6446
Table 4. Performance comparison of JRs and JRs+AI for the external test set in terms of AUC score, 95% confidence interval (95% CI), and Cohen’s kappa score.
Table 4. Performance comparison of JRs and JRs+AI for the external test set in terms of AUC score, 95% confidence interval (95% CI), and Cohen’s kappa score.
Expertise LevelJR1 (∼6 Months)JR2 (∼1 Year)JR3 (>2 Year)
w/o AI+AIw/o AI+AIw/o AI+AI
AUC0.78130.8482 *0.82140.8511 *0.86570.8609
95% CI0.7785–0.78270.8452–0.85110.8197–0.82320.8493–0.85260.8633–0.86760.8585–0.8624
Cohen’s kappa score 10.55740.46510.7400
* Denotes statistically significant (p >0.05). 1 0.41–0.60 moderate agreement, 0.61–0.80 substantial agreement.
Table 5. Performance comparison for each class on the test set and the external test set in terms of AUC scores, Sensitivity, and Specificity.
Table 5. Performance comparison for each class on the test set and the external test set in terms of AUC scores, Sensitivity, and Specificity.
JRsJRs+AI
AUCSensitivitySpecificityAUCSensitivitySpecificity
JR1 ∼ 6 monthsCOVID-19 pneumonia0.65240.38890.91590.74240.62500.8598
Non-COVID-19 pneumonia0.70260.67350.73170.68480.46940.9002
Non-pneumonia0.81210.71500.90910.88780.84170.9339
JR2 ∼ 1 yearCOVID-19 pneumonia0.70790.50000.91590.72390.58330.8645
Non-COVID-19 pneumonia0.68680.55100.82260.66040.46940.8514
Non- pneumonia0.85810.81530.90080.89810.81270.9835
JR3 > 2 yearsCOVID-19 pneumonia0.76810.62500.91120.75420.59720.9112
Non-COVID-19 pneumonia0.75180.61220.89140.76930.59180.9468
Non- pneumonia0.89680.86810.92560.89020.92080.8595
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Feng, Y.; Sim Zheng Ting, J.; Xu, X.; Bee Kun, C.; Ong Tien En, E.; Irawan Tan Wee Jun, H.; Ting, Y.; Lei, X.; Chen, W.-X.; Wang, Y.; et al. Deep Neural Network Augments Performance of Junior Residents in Diagnosing COVID-19 Pneumonia on Chest Radiographs. Diagnostics 2023, 13, 1397. https://doi.org/10.3390/diagnostics13081397

AMA Style

Feng Y, Sim Zheng Ting J, Xu X, Bee Kun C, Ong Tien En E, Irawan Tan Wee Jun H, Ting Y, Lei X, Chen W-X, Wang Y, et al. Deep Neural Network Augments Performance of Junior Residents in Diagnosing COVID-19 Pneumonia on Chest Radiographs. Diagnostics. 2023; 13(8):1397. https://doi.org/10.3390/diagnostics13081397

Chicago/Turabian Style

Feng, Yangqin, Jordan Sim Zheng Ting, Xinxing Xu, Chew Bee Kun, Edward Ong Tien En, Hendra Irawan Tan Wee Jun, Yonghan Ting, Xiaofeng Lei, Wen-Xiang Chen, Yan Wang, and et al. 2023. "Deep Neural Network Augments Performance of Junior Residents in Diagnosing COVID-19 Pneumonia on Chest Radiographs" Diagnostics 13, no. 8: 1397. https://doi.org/10.3390/diagnostics13081397

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop