Next Article in Journal
Omega-3 Polyunsaturated Fatty Acids (n-3 PUFAs) for Immunomodulation in COVID-19 Related Acute Respiratory Distress Syndrome (ARDS)
Next Article in Special Issue
Effects of Multidisciplinary Rehabilitation Program in Patients with Long COVID-19: Post-COVID-19 Rehabilitation (PCR SIRIO 8) Study
Previous Article in Journal
Single-Session Impact of High-Power Laser with Moses Technology for Lower Pole Stones in Retrograde Intrarenal Surgery: Retrospective Study
Previous Article in Special Issue
Clinicopathological Characteristics of Primary Pulmonary Hodgkin Lymphoma (PPHL): Two Institutional Experiences with Comprehensive Literature Review of 115 PPHL Cases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Diagnostic Accuracy of the Artificial Intelligence Methods in Medical Imaging for Pulmonary Tuberculosis: A Systematic Review and Meta-Analysis

1
Department of Respiratory and Critical Care Medicine, West China Medical School/West China Hospital, Sichuan University, Chengdu 610041, China
2
Department of Laboratory Medicine, West China Medical School/West China Hospital, Sichuan University, Chengdu 610041, China
*
Authors to whom correspondence should be addressed.
J. Clin. Med. 2023, 12(1), 303; https://doi.org/10.3390/jcm12010303
Submission received: 23 November 2022 / Revised: 21 December 2022 / Accepted: 24 December 2022 / Published: 30 December 2022

Abstract

:
Tuberculosis (TB) remains one of the leading causes of death among infectious diseases worldwide. Early screening and diagnosis of pulmonary tuberculosis (PTB) is crucial in TB control, and tend to benefit from artificial intelligence. Here, we aimed to evaluate the diagnostic efficacy of a variety of artificial intelligence methods in medical imaging for PTB. We searched MEDLINE and Embase with the OVID platform to identify trials published update to November 2022 that evaluated the effectiveness of artificial-intelligence-based software in medical imaging of patients with PTB. After data extraction, the quality of studies was assessed using quality assessment of diagnostic accuracy studies 2 (QUADAS-2). Pooled sensitivity and specificity were estimated using a bivariate random-effects model. In total, 3987 references were initially identified and 61 studies were finally included, covering a wide range of 124,959 individuals. The pooled sensitivity and the specificity were 91% (95% confidence interval (CI), 89–93%) and 65% (54–75%), respectively, in clinical trials, and 94% (89–96%) and 95% (91–97%), respectively, in model-development studies. These findings have demonstrated that artificial-intelligence-based software could serve as an accurate tool to diagnose PTB in medical imaging. However, standardized reporting guidance regarding AI-specific trials and multicenter clinical trials is urgently needed to truly transform this cutting-edge technology into clinical practice.

1. Introduction

Tuberculosis (TB) is one of the major communicable diseases that seriously endanger human health primarily in developing countries [1], and at least 5.8 million people were estimated to have contracted tuberculosis in 2020. However, around one-sixth of people with active tuberculosis are left undetected or not officially reported each year, which may delay the progress of elimination of this disease before 2035 [2]. Timely diagnosis and treatment could benefit a wide range of tuberculosis patients and minimize the transmission of pathogen in the whole population.
Mycobacterium tuberculosis culture on solid and/or liquid media is still the golden standard for diagnosis. However, the efficiency of culture-based diagnosis in clinical practice is diminished due to long turnaround times and lack of laboratory infrastructure, especially in resource-limited countries. To solve this, the Xpert MTB/RIF assay has emerged as a maturely implemented tool in many countries haunted greatly by TB disease, which is a semiautomated rapid molecular method allowing for rapid diagnosis based upon detection of Mycobacterium tuberculosis DNA and resistance to rifampicin [3], but the application of such rapid tests remains far too limited, with only 1.9 million (33%) people having taken it as an initial diagnostic test in 2022. Simultaneously, the World Health Organization (WHO) has recommended using chest X-ray (CXR) images as a screening technique to better target individuals needing a further microbiological test, which has been proved to be relatively easy to operate, low-cost and highly sensitive [4]. However, an accurate diagnosis with CXRs extremely depends on the clinical experience of radiologists, which poses a huge challenge in the aforementioned countries. As such, there has been increasing interest in using artificial-intelligence-based (AI-based) software in medical imaging for pulmonary tuberculosis (PTB) detection, achieving diagnostic accuracy improvement and cost reduction at the same time. Currently, more than 40 AI-based software programs certified for CXR or computed tomography (CT) examination are available, among which only five are certified for CXR tuberculosis detection [5]. In 2021, Creswell and colleagues conducted a study that tested the five certified software programs (CAD4TB (v6), InferRead®DR (v2), Lunit INSIGHT CXR (v4.9.0), JF CXR-1 (v2), and qXR (v3)) with cohorts in Bangladesh and found that AI-based software significantly outperformed radiologists in TB detection [6]. However, poor reporting and wide variations in design and methodology limit the reliable interpretation of reported diagnostic accuracy [7]. Furthermore, systematic reviews [8,9] of the diagnostic accuracy of this software also identified several limitations in the available evidence, and uncertainty remains regarding its performance in PTB diagnosis.
Hence, we conducted a systematic review and meta-analysis to synthesize evidence of the accuracy of AI-based software in medical imaging for PTB and to provide new insights for future research.

2. Materials and Methods

2.1. Data Source and Search Strategy

A MEDLINE and Embase search through the OVID platform was performed on update to November 2022 without any restriction of country. The search terms were built as follows: ‘artificial intelligence’ (deep learning, machine learning, computer assisted, or cnn), ‘imaging’ (radiography, computed tomography, CT, photograph, or X-ray), ‘diagnostic accuracy metrics’ (sensitivity or specificity), and ‘pulmonary tuberculosis’ (Tuberculosis or tb). The full search strategy is laid out in Supplementary Materials File S2. This systematic review was registered in PROSPERO with the number CRD42022379114 and followed the preferred reporting items for systematic reviews and meta-analyses (PRISMA) guidelines (Supplementary Materials File S3).

2.2. Study Selection

Two researchers independently assessed the candidate studies for inclusion via screening of titles and abstracts, followed by the full text. Any discrepancy between the two researchers was resolved by a third researcher to achieve a consensus. We included all published studies that used AI-based software to analyze medical imaging in PTB diagnosis. Studies that met the following criteria were included in the final group: (1) Any study that analyzed medical imaging for PTB diagnosis with AI-based software; (2) Studies that provided raw diagnostic accuracy data, sensitivity, specificity, area under curve (AUC), accuracy, negative predictive values (NPVs), or positive predictive values (PPVs). Studies were excluded when they met the following criteria: (1) Case reports, conference reports, reviews, meta-analyses, abstracts without full articles, commentaries/editorials, mathematical modeling studies, and economic analyses; (2) Studies that investigated the accuracy of image segmentation or disease prediction; (3) Triage studies; (4) Studies without outcomes or separate data; (5) Studies that failed to report the source of the included population.

2.3. Data Extraction

Two researchers independently extracted demographic and diagnostic-accuracy data using a standardized extraction form from the included studies. When disagreements could not be resolved, we consulted with a third researcher. We extracted data that included study characteristics (first author name, country, year, study design, patient selection methods), demographic information (gender, age, human immunodeficiency virus (HIV) status, drug resistance, history of TB, treatment, imaging modality), AI-based software descriptions (type of artificial intelligence, model, data set, validation methods, threshold score), reference standards, and diagnostic accuracy measures (true and false positives and negatives (TP, FP, FN, TN), AUC, accuracy, sensitivity, specificity, PPV, NPV, and other reported metrics). If there were more than one reported accuracy data set for the same software, with other conditions consistent except for the threshold, the data set with the highest summed sensitivity and specificity would be extracted.

2.4. Quality Assessment

The risk of bias and applicability concerns of the included studies were assessed by two researchers separately, with a revised tool developed for diagnostic studies: QUADAS-2. Any disagreement between the two researchers was resolved through discussion with a third researcher.

2.5. Data Synthesis and Analysis

Data from development studies and clinical studies were analyzed separately. We first obtained the accuracy data that corresponded to TP, FP, FN, and TN in each included study and calculated the estimated pooled sensitivity, specificity, and AUC associated with the 95% CI, using bivariate random-effects models. Additionally, forest plots of sensitivity and specificity were generated for each study. We also used the model to create a summary receiver operating characteristic (SROC) curve. The I2 index was used to assess the heterogeneity between the studies. Values greater than 50% were indicative of substantial heterogeneity [10]. We subsequently chose different study designs, software, reference standards, and AI types as potential sources of heterogeneity, using subgroup analyses to explore the results. A sensitivity analysis was also performed to assess the robustness of the results and identify possible sources of heterogeneity. According to the PRISMA-DTA statement, neither a systematic review nor a meta-analysis of diagnostic accuracy studies is required to assess publication bias [11]. Analyses were conducted in Review Manager version 5.7 and Stata version 17.0 (Stata Corp., College Station, TX, USA), with the midas and metaninf command packages.

3. Results

3.1. Identification of Studies and Study Characteristics

A total of 3987 articles were identified, of which 404 duplicates and 2628 articles were excluded based on screening of titles and abstracts initially. We then excluded 894 studies upon viewing the full-text articles. Finally, 61 studies (23 clinical and 38 development studies) were included in our descriptive analysis (Table 1) [6,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71]. Due to missing information about the raw diagnostic data from the development studies, we only included 13 development studies, with 18 test evaluation results, in the quantitative analysis (Figure 1) [38,42,43,47,51,54,63,64,66,67,68,69,70].
A total of 50 trials, described in 23 clinical studies, were included in the review, with 124,959 people reporting the diagnostic accuracy of the software used for CXR. No study provided prespecified sample-size calculations. In total, twelve studies [6,13,14,19,21,23,25,26,27,28,32,33] used prospectively collected data, and nine studies [6,17,19,20,21,23,25,26,29] used deep-learning-based versions. Additionally, twelve studies [6,12,14,17,18,19,20,25,26,28,29,30] compared software performance with human readers. Reference standards varied greatly; six studies [14,15,24,25,29,31] compared diagnostic performance with human readers, and fourteen studies [6,12,16,17,19,20,21,22,23,26,27,30,32,33] used microbiological references, while three studies [13,18,28] used both. Notably, some studies evaluated the diagnostic accuracy of AI-based software for special populations. Two studies were conducted on diabetic populations [27,28] and one study included only people from prison [14]. Only fourteen studies [6,13,16,17,20,21,22,23,26,27,28,31,32,33] included their own study populations, and the rest were collected from other studies.
Within the model-development studies, thirty reported diagnostic accuracy for PTB identification with deep-learning-based algorithms, compared with eight studies [34,35,43,50,51,52,53,67] that used machine-learning models. Altogether, twenty-seven out of thirty-eight of the available studies were based on public data sets. Several data sets (Montgomery (NIH), Shenzhen (NIH), and Belarus) were analyzed in most studies, but dataset demographic details were not described in most of the studies. Only one article explicitly described the use of semiautomatic lesion delineation for training data. To validate model performance, nine studies [44,46,48,49,59,60,61,68,70] validated algorithms for external data, while the remaining only implemented internal validation. Considering the economics of practical use, thirty-two out of the thirty-eight studies used CXRs as a diagnostic tool, with CT remaining to be further developed. In addition, eleven studies [36,39,40,41,42,48,52,53,57,60,70] made all of the code used in their implementation freely available to the public. As an important step in the radiomic pipeline, feature extraction played a decisive role in the whole process. Hogeweg, L. et al. [53] combined the results of shape analysis, texture analysis, and focal lesion detection into one combined TB score.

3.2. Quality Assessment of Studies

The overall results of the methodological-quality assessment of the included clinical and development studies are summarized respectively in Figure 2 and Figure 3. For clinical studies, the main sources of bias included index tests, flow, and timing. Most development studies were classified as high-risk, particularly with deficiencies in their methods of patient selection, the reference standards used, and their index tests.
For the patient-selection domain, a high or unclear risk of bias was observed in 84% (thirty-two out of thirty-eight) of the development studies, which was mainly related to missing information in the CXR/CT databases. For the index test, a prespecified threshold was reported only in 30% (seven out of twenty-three) of the clinical studies, and 18% (seven out of thirty-eight) of the development studies had a prespecified threshold, while the other studies had a high risk of bias, since the threshold was determined after the analysis in each. For the reference standard domain, a high or unclear risk of bias was seen in 76% (twenty-nine out of thirty-eight) of the development studies, with regards to assessment by radiologists as the reference standard. For flow and timing, there was a high or unclear risk of bias in 39% (nine out of twenty-three) of the clinical studies and 50% (nineteen out of thirty-eight) of the development studies due to the inconsistency of the reference standards and a lack of inclusion of all patients.

3.3. Diagnostic Accuracy Reported in AI-Based Software Assay for PTB

We found that only 13 development studies reported TP, FP, FN, and TN for index tests. Of all the 38 articles that included accuracy assessments, the sensitivity ranged from 0.580 to 0.993 and the specificity from 0.570 to 0.996. It is worth noting that CT showed a higher sensitivity in diagnosis with AI (0.750–0.993 of CT vs. 0.580–0.993 of CXR). The reported performance is summarized in Figure 4. The pooled sensitivity of all included studies was 94% (95% CI 89–96%), with I2 = 93.22 (95% CI 91.07–95.37), and the pooled specificity was 95% (95% CI 91–97%), with I2 = 97.52 (95% CI 96.94–98.09). After excluding the CT-based study, we obtained pooled sensitivity and specificity values of 93% (95% CI 87–96%) and 94% (95% CI 90–97%), respectively.
In total, 23 clinical studies, including 124,959 patients, evaluated the diagnostic efficacy of AI programs for PTB. The sensitivity ranged from 0.487 to 1.00, and the pooled sensitivity was 91% (95% CI 89–93%), with I2 = 93.05 (95% CI 91.74–94.36). The specificity ranged from 0.063 to 0.997, and the pooled specificity was 65% (95% CI 54–75%), with I2 = 99.87 (95% CI 99.86–99.88) (Figure 5).
There was significant heterogeneity in both sensitivity and specificity. We also constructed SROC curves and calculated the AUC for the included studies. The overall diagnostic performance of the clinical studies and the development studies was comparable [AUC 0.91 (95% CI 0.89–0.94) and 0.98 (95% CI 0.97–0.99), respectively] (Supplementary Materials File S1).

3.4. Subgroup and Sensitivity Analyses

Considering the variability of the methods and models tested in the development studies, we only performed a subgroup analysis in the clinical studies, based on predefined parameters, including study design, software, reference standard, and AI type. Some studies were excluded from the relevant subgroup analyses due to missing information or not being categorized into specific groups.
Compared to different study designs, the pooled specificity was 48% (95% CI 34–62%, I2 = 99.87; 99.86–99.88) in the prospective assay versus 75% (95% CI 53–89%, I2 = 99.94; 99.93–99.94) in the nonprospective assay. When Xpert MTB/RIF was used as the reference standard, the pooled specificity of the Xpert MTB/RIF assay [36% (95% CI 24–50%, I2 = 99.93; 99.93–99.94)] was much lower than that of the studies that used human readers [90% (95% CI 80–95%, I2 = 98.70; 98.32–99.08)]. Furthermore, the sensitivity and the specificity of various AI-based software (CAD4TB, qXR, Lunit INSIGHT CXR) evidently differed. The results of the subgroup analyses are summarized in detail in Table 2. There was still a substantial level of heterogeneity among each subgroup analysis.
We subsequently performed sensitivity analyses on the clinical and development studies, respectively. Results of our sensitivity analyses are provided in Supplementary Materials File S1. In the clinical studies, we found three articles that had great effects on the overall results. After removal of the corresponding articles, we obtained a still-high heterogeneity (I2 = 92.97, 91.55–94.39 for sensitivity, I2 = 99.83, 99.82–99.84 for specificity).

4. Discussion

This study sought to (1) evaluate the diagnostic efficacy of AI-based software for PTB and (2) describe the study characteristics, and evaluate the study methodology and the quality of reporting of AI-based software for PTB diagnosis, as well as providing some advice for future software development and clinical applications. Meta-analysis demonstrated that AI-based software has high accuracy in both clinical applications and development studies, indicating that it can assist the physicians in improving the accuracy of PTB diagnosis. However, due to the high heterogeneity and variability between studies, relevant results must be treated with caution when the result of AI-based software is used as a reference standard.
In this systematic review and meta-analysis, we included 23 clinical studies and 38 development studies of PTB diagnosis. Since some missing data were reported, the final count was 13 development studies and 23 clinical studies eligible for quantitative synthesis. Our results show that AI-based software has an excellent ability to diagnose PTB in medical imaging, with pooled sensitivities greater than 0.9 [clinical studies: 91% (95% CI 89%–93%); development studies: 94% (95% CI 89%–96%)]. Additionally, the pooled specificity of the software in the clinical studies was only modest [65% (95% CI 54%–75%)], while that in the development studies was relatively high [95% (95% CI 91%–97%)], which may have been caused by the application of the same test-data set for diagnostic performance assessment. However, a high level of heterogeneity was observed in all the results. Subgroup analysis revealed that nonprospective studies had significantly higher specificity and lower sensitivity than prospective studies had, which might have been due to the inclusion of identified PTB patients in the nonprospective studies. Additionally, studies that used Xpert MTB/RIF as a reference standard had much lower specificity compared to studies that used human readers, possibly because human readers were weaker than Xpert MTB/RIF in correctly identifying negative patients. Furthermore, all commercially available software (CAD4TB, Lunit INSIGHT CXR, and qXR) showed its advantages in improvement of diagnostic accuracy, but we found evident differences in sensitivity and specificity among various AI-based software. The level of heterogeneity between the subgroups remained high, suggesting that study design, software type, AI type, and different reference standards might not be source of heterogeneity. Our follow-up sensitivity analysis indicated that different types of medical imaging might be the sources of heterogeneity, as CT could offer enhanced sensitivity [72].
A number of methodological limitations in the existing evidence were identified, as were study-level factors associated with the reported accuracy, which should all be taken into consideration.
In development studies, most of the current AI-based software was developed for CXR, and only six studies were applied to CT. Because of the deficiency of accuracy data, we performed no subgroup analysis for CT versus CXR. In addition, specific accuracy results, threshold establishment, and inclusion criteria may not have been described well enough to allow emulation for further comparison and may cause greater clinical and methodological heterogeneity. A large proportion of the articles used human readers as the reference standard, meaning systematic overestimation of the diagnostic accuracy of the software. Furthermore, the lack of external validation made it very difficult to formally evaluate algorithm performance. Although most of the experiments used publicly available data sets for model training, few experiments fully disclosed their model details and codes. In addition, almost all of the development articles used manual-lesion-depiction data sets. Semiautomated approaches are known to have greater advantages in lesion delineation, as has been demonstrated with other lung diseases [73], so we encourage more studies in the future to adopt this approach. Several aspects mentioned above lead to the inability to guarantee reproducibility of these experiments. Much of the existing work focuses on multiparametric classification models, ignoring the influence of individual features. Accumulating evidence has confirmed the important role of individual features in discrimination of benign and malignant lung lesions [74,75]; this has great potential for improvement of accuracy and disease identification, and is also informative for research of automated classification models for PTB.
All of the clinical studies evaluated commercially available software developed for CXR. A total of 11 software types were tested, but the version and threshold reported varied among studies. There were varying methodologies of threshold determination and population inclusion, potentially resulting in a high level of heterogeneity. It is worth noting that 13 articles also compared the diagnostic accuracy of AI-based software with human clinicians, which would provide a more objective criterion allowing for a better comparison of models between studies.
Our study had several limitations. Although we searched the relevant literature as comprehensively as possible, some of the literature might have been missed. In addition, some studies failed to report demographic information in detail, and the corresponding subgroup analysis could not be performed. Furthermore, the limited number of studies included for different versions of the software allowed for no further analysis. When AI-based software was used to diagnose PTB, there was significant heterogeneity among studies, so it is difficult to determine whether the software is clinically applicable. Lastly, because current clinical software requires the inclusion of patients over 15 years of age, the diagnostic efficiency for children needs to be further determined.
To improve the future clinical applicability of AI-based software, we recommend that studies include detailed reporting of demographic information, and hope that existing reporting guidelines for diagnostic accuracy studies (STARD) [76] and prediction models (TRIPOD) [77] can be improved as soon as possible to conduct AI-specific amendments. In addition, some model training and validations were performed on CXRs from data sets or sites, potentially resulting in an overestimation of diagnosis power. As such, we suggest that different data sets should be used for model training and testing. Moreover, research teams can collaborate with multiple clinical centers for clinical trials and external validation to make results superior and investigate the stability and heterogeneity of their performance in clinical scenarios. What is more, we appealed to a large number of open, multi-source, and anonymous databases, along with detailed reporting of all of the information needed, such as reference standard, age, HIV status, etc., to fulfill the need for an adequate amount of data with high quality. At the same time, we recommend that development studies make their model details and all of the code used for their experiments freely available to the public to make it possible to reproduce these studies. It is also noteworthy that the diagnostic accuracy of AI-based software should be evaluated against a microbiological reference standard. Lastly, we found a lack of use of AI-based software in CT, and more studies may be needed to explore its superiority in early diagnosis of PTB. In addition, the influence of parameters such as intensity quantization, on imaging and final diagnosis in particular, could be considered.

5. Conclusions

In summary, there were relatively high pooled sensitivity and specificity values of AI-based software, which indicates that AI-based software has potential to facilitate diagnosis of PTB in medical imaging, especially in large-scale screening. Heterogeneity was significantly high and extensive variation in reporting, design, and methodology was observed. Thus, standardized reporting guidance around AI-specific trials and multicenter clinical trials is urgently needed to further confirm their stability and heterogeneity in various populations and settings. In the future, we expect more AI-based software with high accuracy to be comprehensively applied for early clinical detection of PTB.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jcm12010303/s1, Supplementary Materials File S1: Figure S1 Quality assessment (QUADAS 2) summary of clinical studies: risk of bias & applicability concerns; Figure S2 Quality assessment (QUADAS 2) summary of development studies: Risk of bias and applicability concerns; Figure S3 Sensitivity analysis of clinical studies; Figure S4 Sensitivity analysis of development studies; Figure S5 Summary receiver operating characteristic (SOC) curve of clinical studies; Figure S6 Summary receiver operating characteristic (SROC) curve of development studies; Table S1 Demographics of clinical studies; Table S2 Accuracy information of clinical studies included; Table S3 Accuracy measures reported by development studies; Table S4 Accuracy measures reported by development studies; Supplementary Materials File S2: Search strategy; Supplementary Materials File S3: PRISMA checklist [78].

Author Contributions

Conceptualization and design, C.W. and B.Y.; data curation and data analysis, Y.Z., Y.W. and W.Z.; manuscript editing and manuscript review, C.W., B.Y., Y.Z., Y.W. and W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (82100119), Chinese Postdoctoral Science Foundation (2022T150451, 2021M692309), the Science and Technology Project of Sichuan (2020YFG0473, 2022ZDZX0018), National College Students’ innovation and entrepreneurship training program of Sichuan University (20231261L).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pai, M.; Behr, M.A.; Dowdy, D.; Dheda, K.; Divangahi, M.; Boehme, C.C.; Ginsberg, A.; Swaminathan, S.; Spigelman, M.; Getahun, H.; et al. Tuberculosis. Nat. Rev. Dis. Prim. 2016, 2, 16076. [Google Scholar] [CrossRef] [PubMed]
  2. World Health Organization. Global Tuberculosis Report 2021. Available online: https://www.who.int/teams/global-tuberculosis-programme/tb-reports (accessed on 10 November 2022).
  3. Chen, X.; Hu, T.Y. Strategies for advanced personalized tuberculosis diagnosis: Current technologies and clinical approaches. Precis. Clin.Med. 2021, 2, 35–44. [Google Scholar] [CrossRef] [PubMed]
  4. Hoog, A.H.V.; Meme, H.K.; Laserson, K.F.; Agaya, J.A.; Muchiri, B.G.; Githui, W.A.; Odeny, L.O.; Marston, B.J.; Borgdorff, M.W. Screening strategies for tuberculosis prevalence surveys: The value of chest radiography and symptoms. PLoS ONE 2012, 7, e38691. [Google Scholar] [CrossRef]
  5. Diagnostic Image Analysis Group. AI for radiology: An implementation guide 2020. Available online: https://grand-challenge.org/aiforradiology/ (accessed on 10 November 2022).
  6. Qin, Z.Z.; Ahmed, S.; Sarker, M.S.; Paul, K.; Adel, A.S.S.; Naheyan, T.; Barrett, R.; Banu, S.; Creswell, J. Tuberculosis detection from chest X-rays for triaging in a high tuberculosis-burden setting: An evaluation of five artificial intelligence algorithms. Lancet Digit. Health 2021, 3, e543–e554. [Google Scholar] [CrossRef]
  7. Liu, X.; Faes, L.; Kale, A.U.; Wagner, S.K.; Fu, D.J.; Bruynseels, A.; Mahendiran, T.; Moraes, G.; Shamdas, M.; Kern, C.; et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: A systematic review and meta-analysis. Lancet Digit. Health 2019, 1, e271–e297. [Google Scholar] [CrossRef] [PubMed]
  8. Harris, M.; Qi, A.; JeaGal, L.; Torabi, N.; Menzies, D.; Korobitsyn, A.; Pai, M.; Nathavitharana, R.R.; Khan, F.A. A systematic review of the diagnostic accuracy of artificial intelligence-based computer programs to analyze chest X-rays for pulmonary tuberculosis. PLoS ONE 2019, 14, e0221339. [Google Scholar] [CrossRef]
  9. Pande, T.; Cohen, C.; Pai, M.; Khan, F.A. Computer-aided detection of pulmonary tuberculosis on digital chest radiographs: A systematic review. Int. J. Tuberc. Lung Dis. 2016, 20, 1226–1230. [Google Scholar] [CrossRef]
  10. Puhan, M.A.; Gimeno-Santos, E.; Cates, C.J.; Troosters, T. Pulmonary rehabilitation following exacerbations of chronic obstructive pulmonary disease. Cochrane Database Syst. Rev. 2016, 2019, CD005305. [Google Scholar] [CrossRef]
  11. McInnes, M.D.F.; Moher, D.; Thombs, B.D.; McGrath, T.A.; Bossuyt, P.M.; Clifford, T.; Cohen, J.F.; Deeks, J.J.; Gatsonis, C.; Hooft, L.; et al. Preferred reporting items for a systematic review and meta-analysis of diagnostic test accuracy studies. The PRISMA-DTA Statement. JAMA 2018, 319, 388–396. [Google Scholar] [CrossRef]
  12. Maduskar, P.; Muyoyeta, M.; Ayles, H.; Hogeweg, L.; Peters-Bax, L.; Van Ginneken, B. Detection of tuberculosis using digital chest radiography: Automated reading vs. interpretation by clinical officers. Int. J. Tuberc. Lung Dis. 2013, 17, 1613–1620. [Google Scholar] [CrossRef]
  13. Muyoyeta, M.; Maduskar, P.; Moyo, M.; Kasese, N.; Milimo, D.; Spooner, R.; Kapata, N.; Hogeweg, L.; Van Ginneken, B.; Ayles, H. The sensitivity and specificity of using a computer aided diagnosis program for automatically scoring chest X-rays of presumptive TB patients compared with Xpert MTB/RIF in Lusaka Zambia. PLoS ONE 2014, 9, e93757. [Google Scholar] [CrossRef] [PubMed]
  14. Steiner, A.; Mangu, C.; Hombergh, J.V.D.; van Deutekom, H.; van Ginneken, B.; Clowes, P.; Mhimbira, F.; Mfinanga, S.; Rachow, A.; Reither, K.; et al. Screening for pulmonary tuberculosis in a Tanzanian prison and computer-aided interpretation of chest X-rays. Public Health Action 2015, 5, 249–254. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Melendez, J.; Hogeweg, L.; Sánchez, C.I.; Philipsen, R.H.H.M.; Aldridge, R.; Hayward, A.C.; Abubakar, I.; van Ginneken, B.; Story, A. Accuracy of an automated system for tuberculosis detection on chest radiographs in high-risk screening. Int. J. Tuberc. Lung Dis. 2018, 22, 567–571. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Zaidi, S.M.A.; Habib, S.S.; Van Ginneken, B.; Ferrand, R.A.; Creswell, J.; Khowaja, S.; Khan, A. Evaluation of the diagnostic accuracy of computer-aided detection of tuberculosis on chest radiography among private sector patients in Pakistan. Sci. Rep. 2018, 8, 12339. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Qin, Z.Z.; Sander, M.S.; Rai, B.; Titahong, C.N.; Sudrungrot, S.; Laah, S.N.; Adhikari, L.M.; Carter, E.J.; Puri, L.; Codlin, A.J.; et al. Using artificial intelligence to read chest radiographs for tuberculosis detection: A multi-site evaluation of the diagnostic accuracy of three deep learning systems. Sci. Rep. 2019, 9, 15000. [Google Scholar] [CrossRef] [Green Version]
  18. Philipsen, R.H.H.M.; Sánchez, C.I.; Melendez, J.; Lew, W.J.; Van Ginneken, B. Automated chest X-ray reading for tuberculosis in the Philippines to improve case detection: A cohort study. Int. J. Tuberc. Lung Dis. 2019, 23, 805–810. [Google Scholar] [CrossRef]
  19. Murphy, K.; Habib, S.S.; Zaidi, S.M.A.; Khowaja, S.; Khan, A.; Melendez, J.; Scholten, E.T.; Amad, F.; Schalekamp, S.; Verhagen, M.; et al. Computer aided detection of tuberculosis on chest radiographs: An evaluation of the CAD4TB v6 system. Sci. Rep. 2020, 10, 5492. [Google Scholar] [CrossRef] [Green Version]
  20. Nash, M.; Kadavigere, R.; Andrade, J.; Sukumar, C.A.; Chawla, K.; Shenoy, V.P.; Pande, T.; Huddart, S.; Pai, M.; Saravu, K. Deep learning, computer-aided radiography reading for tuberculosis: A diagnostic accuracy study from a tertiary hospital in India. Sci. Rep. 2020, 10, 210. [Google Scholar] [CrossRef] [Green Version]
  21. Soares, T.R.; de Oliveira, R.D.; Liu, Y.E.; Santos, A.D.S.; dos Santos, P.C.P.; Monte, L.R.S.; de Oliveira, L.M.; Park, C.M.; Hwang, E.J.; Andrews, J.R.; et al. Evaluation of chest X-ray with automated interpretation algorithms for mass tuberculosis screening in prisons: A cross-sectional study. Lancet Reg. Health-Am. 2023, 17, 100388. [Google Scholar] [CrossRef]
  22. Breuninger, M.; Van Ginneken, B.; Philipsen, R.H.H.M.; Mhimbira, F.; Hella, J.J.; Lwilla, F.; Hombergh, J.V.D.; Ross, A.; Jugheli, L.; Wagner, D.; et al. Diagnostic accuracy of computer-aided detection of pulmonary tuberculosis in chest radiographs: A validation study from Sub-Saharan Africa. PLoS ONE 2014, 9, e106381. [Google Scholar] [CrossRef]
  23. Khan, F.A.; Majidulla, A.; Tavaziva, G.; Nazish, A.; Abidi, S.K.; Benedetti, A.; Menzies, D.; Johnston, J.C.; Khan, A.J.; Saeed, S. Chest X-ray analysis with deep learning-based software as a triage test for pulmonary tuberculosis: A prospective study of diagnostic accuracy for culture-confirmed disease. Lancet Digit. Health 2020, 2, e573–e581. [Google Scholar] [CrossRef] [PubMed]
  24. Young, C.; Barker, S.; Ehrlich, R.; Kistnasamy, B.; Yassi, A. Computer-aided detection for tuberculosis and silicosis in chest radiographs of gold miners of South Africa. Int. J. Tuberc. Lung Dis. 2020, 24, 444–451. [Google Scholar] [CrossRef] [PubMed]
  25. Liao, Q.; Feng, H.; Li, Y.; Lai, X.; Pan, J.; Zhou, F.; Zhou, L.; Chen, L. Evaluation of an artificial intelligence (AI) system to detect tuberculosis on chest X-ray at a pilot active screening project in Guangdong, China in 2019. J. X-ray Sci. Technol. 2022, 30, 221–230. [Google Scholar] [CrossRef] [PubMed]
  26. Codlin, A.J.; Dao, T.P.; Vo, L.N.Q.; Forse, R.J.; Van Truong, V.; Dang, H.M.; Nguyen, L.H.; Nguyen, H.B.; Nguyen, N.V.; Sidney-Annerstedt, K.; et al. Independent evaluation of 12 artificial intelligence solutions for the detection of tuberculosis. Sci. Rep. 2021, 11, 23895. [Google Scholar] [CrossRef]
  27. Habib, S.S.; Rafiq, S.; Zaidi, S.M.A.; Ferrand, R.A.; Creswell, J.; Van Ginneken, B.; Jamal, W.Z.; Azeemi, K.S.; Khowaja, S.; Khan, A. Evaluation of computer aided detection of tuberculosis on chest radiography among people with diabetes in Karachi Pakistan. Sci. Rep. 2020, 10, 6276. [Google Scholar] [CrossRef] [Green Version]
  28. Koesoemadinata, R.C.; Kranzer, K.; Livia, R.; Susilawati, N.; Annisa, J.; Soetedjo, N.N.M.; Ruslami, R.; Philipsen, R.; van Ginneken, B.; Soetikno, R.D.; et al. Computer-assisted chest radiography reading for tuberculosis screening in people living with diabetes mellitus. Int. J. Tuberc. Lung Dis. 2018, 22, 1088–1094. [Google Scholar] [CrossRef]
  29. Lee, J.H.; Park, S.; Hwang, E.J.; Goo, J.M.; Lee, W.Y.; Lee, S.; Kim, H.; Andrews, J.R.; Park, C.M. Deep learning–based automated detection algorithm for active pulmonary tuberculosis on chest radiographs: Diagnostic performance in systematic screening of asymptomatic individuals. Eur. Radiol. 2020, 31, 1069–1080. [Google Scholar] [CrossRef]
  30. Gelaw, S.M.; Kik, S.V.; Ruhwald, M.; Ongarello, S.; Egzertegegne, T.S.; Gorbacheva, O.; Gilpin, C.; Marano, N.; Lee, S.; Phares, C.R. Diagnostic accuracy of three computer-aided detection systems for detecting pulmonary tuberculosis on chest radiography when used for screening: Analysis of an international, multicenter migrants screening study. medRxiv 2022. [Google Scholar] [CrossRef]
  31. Ehrlich, R.; Barker, S.; Naude, J.T.W.; Rees, D.; Kistnasamy, B.; Naidoo, J.; Yassi, A. Accuracy of computer-aided detection of occupational lung disease: Silicosis and pulmonary tuberculosis in Ex-Miners from the South African gold mines. Int. J. Environ. Res. Public Health 2022, 19, 12402. [Google Scholar] [CrossRef]
  32. Kagujje, M.; Kerkhoff, A.D.; Nteeni, M.; Dunn, I.; Mateyo, K.; Muyoyeta, M. The performance of computer-aided detection digital chest X-ray reading technologies for triage of active tuberculosis among persons with a history of previous tuberculosis. Clin. Infect. Dis. 2022, ciac679. [Google Scholar] [CrossRef]
  33. Tavaziva, G.; Majidulla, A.; Nazish, A.; Saeed, S.; Benedetti, A.; Khan, A.J.; Khan, F.A. Diagnostic accuracy of a commercially available, deep learning-based chest X-ray interpretation software for detecting culture-confirmed pulmonary tuberculosis. Int. J. Infect. Dis. 2022, 122, 15–20. [Google Scholar] [CrossRef] [PubMed]
  34. Shen, R.; Cheng, I.; Basu, A. A hybrid knowledge-guided detection technique for screening of infectious pulmonary tuberculosis from chest radiographs. IEEE Trans. Biomed. Eng. 2010, 57, 2646–2656. [Google Scholar] [CrossRef] [PubMed]
  35. Melendez, J.; van Ginneken, B.; Maduskar, P.; Philipsen, R.H.H.M.; Reither, K.; Breuninger, M.; Adetifa, I.M.O.; Maane, R.; Ayles, H.; Sanchez, C.I. A novel multiple-instance learning-based approach to computer-aided detection of tuberculosis on chest X-rays. IEEE Trans. Med. Imaging 2015, 34, 179–192. [Google Scholar] [CrossRef] [PubMed]
  36. Pasa, F.; Golkov, V.; Pfeiffer, F.; Cremers, D. Efficient deep network architectures for fast chest X-ray tuberculosis screening and visualization. Sci. Rep. 2019, 9, 6268. [Google Scholar] [CrossRef] [Green Version]
  37. Xie, Y.; Wu, Z.; Han, X.; Wang, H.; Wu, Y.; Cui, L.; Feng, J.; Zhu, Z.; Chen, Z. Computer-aided system for the detection of multicategory pulmonary tuberculosis in radiographs. J. Healthc. Eng. 2020, 2020, 9205082. [Google Scholar] [CrossRef]
  38. Ma, L.; Wang, Y.; Guo, L.; Zhang, Y.; Wang, P.; Pei, X.; Qian, L.; Jaeger, S.; Ke, X.; Yin, X.; et al. Developing and verifying automatic detection of active pulmonary tuberculosis from multi-slice spiral CT images based on deep learning. J. X-ray Sci. Technol. 2020, 28, 939–951. [Google Scholar] [CrossRef]
  39. Rajpurkar, P.; O’Connell, C.; Schechter, A.; Asnani, N.; Li, J.; Kiani, A.; Ball, R.L.; Mendelson, M.; Maartens, G.; van Hoving, D.J.; et al. CheXaid: Deep learning assistance for physician diagnosis of tuberculosis using chest X-rays in patients with HIV. npj Digit. Med. 2020, 3, 115. [Google Scholar] [CrossRef]
  40. Oloko-Oba, M.; Viriri, S. Ensemble of EfficientNets for the diagnosis of tuberculosis. Comput. Intell. Neurosci. 2021, 2021, 9790894. [Google Scholar] [CrossRef]
  41. Mamalakis, M.; Swift, A.J.; Vorselaars, B.; Ray, S.; Weeks, S.; Ding, W.; Clayton, R.H.; Mackenzie, L.S.; Banerjee, A. DenResCov-19: A deep transfer learning network for robust automatic classification of COVID-19, pneumonia, and tuberculosis from X-rays. Comput. Med. Imaging Graph. 2021, 94, 102008. [Google Scholar] [CrossRef]
  42. Rajakumar, M.; Sonia, R.; Maheswari, B.U.; Karuppiah, S. Tuberculosis detection in chest X-ray using Mayfly-algorithm optimized dual-deep-learning features. J. X-ray Sci. Technol. 2021, 29, 961–974. [Google Scholar] [CrossRef]
  43. Sharma, A.; Sharma, A.; Malhotra, R.; Singh, P.; Chakrabortty, R.K.; Mahajan, S.; Pandit, A.K. An accurate artificial intelligence system for the detection of pulmonary and extra pulmonary tuberculosis. Tuberculosis 2021, 131, 102143. [Google Scholar] [CrossRef] [PubMed]
  44. Wang, L.; Ding, W.; Mo, Y.; Shi, D.; Zhang, S.; Zhong, L.; Wang, K.; Wang, J.; Huang, C.; Ye, Z.; et al. Distinguishing nontuberculous mycobacteria from Mycobacterium tuberculosis lung disease from CT images using a deep learning framework. Eur. J. Nucl. Med. 2021, 48, 4293–4306. [Google Scholar] [CrossRef] [PubMed]
  45. Showkatian, E.; Salehi, M.; Ghaffari, H.; Reiazi, R.; Sadighi, N. Deep learning-based automatic detection of tuberculosis disease in chest X-ray images. Pol. J. Radiol. 2022, 87, 118–124. [Google Scholar] [CrossRef] [PubMed]
  46. Zhou, W.; Cheng, G.; Zhang, Z.; Zhu, L.; Jaeger, S.; Lure, F.Y.M.; Guo, L. Deep learning-based pulmonary tuberculosis automated detection on chest radiography: Large-scale independent testing. Quant. Imaging Med. Surg. 2022, 12, 2344. [Google Scholar] [CrossRef] [PubMed]
  47. Rajaraman, S.; Zamzmi, G.; Folio, L.; Alderson, P.; Antani, S. Chest X-ray bone suppression for improving classification of tuberculosis-consistent findings. Diagnostics 2021, 11, 840. [Google Scholar] [CrossRef] [PubMed]
  48. Yan, C.; Wang, L.; Lin, J.; Xu, J.; Zhang, T.; Qi, J.; Li, X.; Ni, W.; Wu, G.; Huang, J.; et al. A fully automatic artificial intelligence–based CT image analysis system for accurate detection, diagnosis, and quantitative severity evaluation of pulmonary tuberculosis. Eur. Radiol. 2021, 32, 2188–2199. [Google Scholar] [CrossRef]
  49. Zhang, K.; Qi, S.; Cai, J.; Zhao, D.; Yu, T.; Yue, Y.; Yao, Y.; Qian, W. Content-based image retrieval with a Convolutional Siamese Neural Network: Distinguishing lung cancer and tuberculosis in CT images. Comput. Biol. Med. 2021, 140, 105096. [Google Scholar] [CrossRef]
  50. Arzhaeva, Y.; Hogeweg, L.; De Jong, P.A.; Viergever, M.A.; Van Ginneken, B. Global and local multi-valued dissimilarity-based classification: Application to computer-aided detection of tuberculosis. Med. Image Comput. Comput. Assist. Interv. 2009, 12 Pt 2, 724–731. [Google Scholar] [CrossRef]
  51. Jaeger, S.; Karargyris, A.; Candemir, S.; Folio, L.; Siegelman, J.; Callaghan, F.; Xue, Z.; Palaniappan, K.; Singh, R.K.; Antani, S.; et al. Automatic tuberculosis screening using chest radiographs. IEEE Trans. Med. Imaging 2014, 33, 233–245. [Google Scholar] [CrossRef]
  52. Chauhan, A.; Chauhan, D.; Rout, C. Role of Gist and PHOG features in computer-aided diagnosis of tuberculosis without segmentation. PLoS ONE 2014, 9, e112980. [Google Scholar] [CrossRef]
  53. Hogeweg, L.; Sanchez, C.I.; Maduskar, P.; Philipsen, R.; Story, A.; Dawson, R.; Theron, G.; Dheda, K.; Peters-Bax, L.; van Ginneken, B. Automatic detection of tuberculosis in chest radiographs using a combination of textural, focal, and shape abnormality analysis. IEEE Trans. Med. Imaging 2015, 34, 2429–2442. [Google Scholar] [CrossRef] [PubMed]
  54. Lakhani, P.; Sundaram, B. Deep learning at chest radiography: Automated classification of pulmonary tuberculosis by using Convolutional Neural Networks. Radiology 2017, 284, 574–582. [Google Scholar] [CrossRef] [PubMed]
  55. Han, D.; He, T.; Yu, Y.; Guo, Y.; Chen, Y.; Duan, H.; Yu, N. Diagnosis of active pulmonary tuberculosis and community acquired pneumonia using Convolution Neural Network based on transfer learning. Acad. Radiol. 2022, 29, 1486–1492. [Google Scholar] [CrossRef] [PubMed]
  56. An, L.; Peng, K.; Yang, X.; Huang, P.; Luo, Y.; Feng, P.; Wei, B. E-TBNet: Light Deep Neural Network for automatic detection of tuberculosis with X-ray DR Imaging. Sensors 2022, 22, 821. [Google Scholar] [CrossRef] [PubMed]
  57. Lee, S.; Yim, J.-J.; Kwak, N.; Lee, Y.J.; Lee, J.-K.; Lee, J.Y.; Kim, J.S.; Kang, Y.A.; Jeon, D.; Jang, M.-J.; et al. Deep learning to determine the activity of pulmonary tuberculosis on chest radiographs. Radiology 2021, 301, 435–442. [Google Scholar] [CrossRef] [PubMed]
  58. Khatibi, T.; Shahsavari, A.; Farahani, A. Proposing a novel multi-instance learning model for tuberculosis recognition from chest X-ray images based on CNNs, complex networks and stacked ensemble. Phys. Eng. Sci. Med. 2021, 44, 291–311. [Google Scholar] [CrossRef]
  59. Kim, T.K.; Yi, P.H.; Hager, G.D.; Lin, C.T. Refining dataset curation methods for deep learning-based automated tuberculosis screening. J. Thorac. Dis. 2020, 12, 5078–5085. [Google Scholar] [CrossRef] [PubMed]
  60. Feng, B.; Chen, X.; Chen, Y.; Lu, S.; Liu, K.; Li, K.; Liu, Z.; Hao, Y.; Li, Z.; Zhu, Z.; et al. Solitary solid pulmonary nodules: A CT-based deep learning nomogram helps differentiate tuberculosis granulomas from lung adenocarcinomas. Eur. Radiol. 2020, 30, 6497–6507. [Google Scholar] [CrossRef]
  61. Hwang, E.J.; Park, S.; Jin, K.-N.; Kim, J.I.; Choi, S.Y.; Lee, J.H.; Goo, J.M.; Aum, J.; Yim, J.-J.; Park, C.M.; et al. Development and validation of a deep learning–based automatic detection algorithm for active pulmonary tuberculosis on chest radiographs. Clin. Infect. Dis. 2019, 69, 739–747. [Google Scholar] [CrossRef] [Green Version]
  62. Heo, S.-J.; Kim, Y.; Yun, S.; Lim, S.-S.; Kim, J.; Nam, C.-M.; Park, E.-C.; Jung, I.; Yoon, J.-H. Deep learning algorithms with demographic information help to detect tuberculosis in chest radiographs in annual workers’ health examination data. Int. J. Environ. Res. Public Health 2019, 16, 250. [Google Scholar] [CrossRef]
  63. Aguiar, F.S.; Torres, R.C.; Pinto, J.V.F.; Kritski, A.L.; Seixas, J.M.; Mello, F.C.Q. Development of two artificial neural network models to support the diagnosis of pulmonary tuberculosis in hospitalized patients in Rio de Janeiro, Brazil. Med. Biol. Eng. Comput. 2016, 54, 1751–1759. [Google Scholar] [CrossRef] [PubMed]
  64. Faruk, O.; Ahmed, E.; Ahmed, S.; Tabassum, A.; Tazin, T.; Bourouis, S.; Khan, M.M. A novel and robust approach to detect tuberculosis using transfer learning. J. Healthc. Eng. 2021, 2021, 1002799. [Google Scholar] [CrossRef] [PubMed]
  65. Karki, M.; Kantipudi, K.; Yu, H.; Yang, F.; Kassim, Y.M.; Yaniv, Z.; Jaeger, S. Identifying drug-resistant tuberculosis in chest radiographs: Evaluation of CNN architectures and training strategies. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Mexico City, Mexico, 1–5 November 2021; pp. 2964–2967. [Google Scholar] [CrossRef]
  66. Dasanayaka, C.; Dissanayake, M.B. Deep learning methods for screening pulmonary tuberculosis using chest X-rays. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2021, 9, 39–49. [Google Scholar] [CrossRef]
  67. Govindarajan, S.; Swaminathan, R. Extreme learning machine based differentiation of pulmonary tuberculosis in chest radiographs using integrated local feature descriptors. Comput. Methods Programs Biomed. 2021, 204, 106058. [Google Scholar] [CrossRef] [PubMed]
  68. Acharya, V.; Dhiman, G.; Prakasha, K.; Bahadur, P.; Choraria, A.; Prabhu, S.; Chadaga, K.; Viriyasitavat, W.; Kautish, S. AI-assisted tuberculosis detection and classification from chest X-rays using a deep learning normalization-free network model. Comput. Intell. Neurosci. 2022, 2022, 2399428. [Google Scholar] [CrossRef] [PubMed]
  69. Kadry, S.; Srivastava, G.; Rajinikanth, V.; Rho, S.; Kim, Y. Tuberculosis detection in chest radiographs using spotted hyena algorithm optimized deep and handcrafted features. Comput. Intell. Neurosci. 2022, 2022, 9263379. [Google Scholar] [CrossRef] [PubMed]
  70. Kazemzadeh, S.; Yu, J.; Jamshy, S.; Pilgrim, R.; Nabulsi, Z.; Chen, C.; Beladia, N.; Lau, C.; McKinney, S.M.; Hughes, T.; et al. Deep learning detection of active pulmonary tuberculosis at chest radiography matched the clinical performance of radiologists. Radiology 2023, 306, 124–137. [Google Scholar] [CrossRef] [PubMed]
  71. Margarat, G.S.; Hemalatha, G.; Mishra, A.; Shaheen, H.; Maheswari, K.; Tamijeselvan, S.; Kumar, U.P.; Banupriya, V.; Ferede, A.W. Early diagnosis of tuberculosis using deep learning approach for iot based healthcare applications. Comput. Intell. Neurosci. 2022, 2022, 3357508. [Google Scholar] [CrossRef]
  72. Skoura, E.; Zumla, A.; Bomanji, J. Imaging in tuberculosis. Int. J. Infect. Dis. 2015, 32, 87–93. [Google Scholar] [CrossRef]
  73. Owens, C.A.; Peterson, C.; Tang, C.; Koay, E.J.; Yu, W.; Mackin, D.S.; Li, J.; Salehpour, M.R.; Fuentes, D.T.; Court, L.; et al. Lung tumor segmentation methods: Impact on the uncertainty of radiomics features for non-small cell lung cancer. PLoS ONE 2018, 13, e0205003. [Google Scholar] [CrossRef] [Green Version]
  74. Bianconi, F.; Palumbo, I.; Fravolini, M.L.; Rondini, M.; Minestrini, M.; Pascoletti, G.; Nuvoli, S.; Spanu, A.; Scialpi, M.; Aristei, C.; et al. Form factors as potential imaging biomarkers to differentiate benign vs. malignant lung lesions on CT scans. Sensors 2022, 22, 5044. [Google Scholar] [CrossRef] [PubMed]
  75. Peikert, T.; Duan, F.; Rajagopalan, S.; Karwoski, R.A.; Clay, R.; Robb, R.A.; Qin, Z.; Sicks, J.; Bartholmai, B.J.; Maldonado, F. Novel high-resolution computed tomography-based radiomic classifier for screen-identified pulmonary nodules in the national lung screening trial. PLoS ONE 2018, 13, e0196910. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  76. Bossuyt, P.M.; Reitsma, J.B.; Bruns, D.E.; Gatsonis, C.A.; Glasziou, P.P.; Irwig, L.; Lijmer, J.G.; Moher, D.; Rennie, D.; de Vet, H.C.W.; et al. STARD 2015: An updated list of essential items for reporting diagnostic accuracy studies. Clin. Chem. 2015, 61, 1446–1452. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  77. Collins, G.S.; Reitsma, J.B.; Altman, D.G.; Moons, K.G.M. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): The TRIPOD statement. BMJ 2014, 350, g7594. [Google Scholar] [CrossRef] [Green Version]
  78. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef]
Figure 1. Study flow diagram. Computer-aided detection (CAD).
Figure 1. Study flow diagram. Computer-aided detection (CAD).
Jcm 12 00303 g001
Figure 2. Quality assessment (QUADAS) graph of clinical studies.
Figure 2. Quality assessment (QUADAS) graph of clinical studies.
Jcm 12 00303 g002
Figure 3. Quality assessment (QUADAS) graph of development studies.
Figure 3. Quality assessment (QUADAS) graph of development studies.
Jcm 12 00303 g003
Figure 4. Forest plot of development-study sensitivity and specificity for PTB [38,42,43,47,51,54,63,64,66,67,68,69,70].
Figure 4. Forest plot of development-study sensitivity and specificity for PTB [38,42,43,47,51,54,63,64,66,67,68,69,70].
Jcm 12 00303 g004
Figure 5. Forest plot of clinical-study sensitivity and specificity for PTB [6,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33].
Figure 5. Forest plot of clinical-study sensitivity and specificity for PTB [6,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33].
Jcm 12 00303 g005
Table 1. Methods of studies included in descriptive analysis.
Table 1. Methods of studies included in descriptive analysis.
First Author, YearImaging ModalityComputer Software/ModelReference StandardAccuracy Measures
Maduskar et al., 2013 [12]CXRCAD4TB (v 1.08)AFB smear, MTB cultureTP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Muyoyeta et al., 2014 [13]CXRCAD4TB (v 1.08)Xpert MTB/RIF, human readerTP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Steiner et al., 2015 [14]CXRCAD4TB (v 3.07)Human readerAUC
Melendez et al., 2018 [15]CXRCAD4TB (v 5)Human readerTP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Zaidi et al., 2018 [16]CXRCAD4TB (v 3.07)Xpert MTB/RIFTP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Qin et al., 2019 [17]CXRCAD4TB (v 6),
qXR (v 2),
Lunit INSIGHT CXR (v 4.7.2)
Xpert MTB/RIFTP, FP, TN, FN, AUC, ACC, Sn, Sp
Philipsen et al., 2019 [18]CXRCAD4TB (v 5)Xpert MTB/RIF, human readerTP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Murphy et al., 2020 [19]CXRCAD4TB (v 6)Xpert MTB/RIFTP, FP, TN, FN, AUC, Sn, Sp
Nash et al., 2020 [20]CXRqXR (v 2)AFB smear, Xpert MTB/RIF or MTB cultureAUC, Sn, Sp
Soares et al., 2023 [21]CXRCAD4TB (v 6),
Lunit INSIGHT CXR (v 3.1.0.0),
qXR (v 3)
Xpert MTB/RIF, MTB cultureAUC, Sn, Sp, PPV, NPV
Qin et al., 2021 [6]CXRCAD4TB (v 7),
InferRead DR (v 2),
Lunit INSIGHT CXR (v 4.9.0),
JF CXR-1 (v 2),
qXR, (v 3)
Xpert MTB/RIFAUC, Sn, Sp
Breuninger et al., 2014 [22]CXRCAD4TB (v 3.07)AFB smear, MTB cultureSn, Sp, PPV, NPV
Khan et al., 2020 [23]CXRqXR (v 2),
CAD4TB (v 6)
MTB cultureACC, Sn, Sp, PPV, NPV
Young et al., 2020 [24]CXRNot namedHuman readerAUC, Sn, Sp
Liao et al., 2022 [25]CXRJF CXR-1 (v 2)Human readerTP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Codlin et al., 2021 [26]CXRqXR (v 3),
CAD4TB (v 7),
Genki (v 2),
Lunit INSIGHT CXR (v 3.1.0.0),
JF CXR-1 (v 3.0),
InferRead DR Chest (v 1.0.0.0),
ChestEye (v 1),
T-Xnet (v 1),
XrayAME (v 1),
COTO (v 1),
SemanticMD (v 1),
Dr CADx (v 0.1)
Xpert MTB/RIFTP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Habib et al., 2020 [27]CXRCAD4TB (v 3.07)Xpert MTB/RIFAUC, Sn, Sp, PPV, NPV
Koesoemadinata et al., 2018 [28]CXRCAD4TB (v 5)Composite reference standard(s)AUC, Sn, Sp
Lee et al., 2020 [29]CXRLunit INSIGHT CXR (v 4.7.2)MTB culture, AFB smear, TB polymerase chain reaction, human readerTP, FP, TN, FN, AUC, ACC, Sn, Sp, PPV, NPV
Gelaw et al., 2022 [30]CXRCAD4TB (v 6),
Lunit INSIGHT CXR (v 4.9.0),
qXR (v 2)
Xpert MTB/RIF, Mycobacterium tuberculosis (MTB) cultureTP, FP, TN, FN, Sn, Sp
Ehrlich et al., 2022 [31]CXRCAD4TB (v 7)Human readerTP, FP, TN, FN, AUC, Sn, Sp
Kagujje et al., 2022 [32]CXRCAD4TB (v 7),
qXR (v 3)
Xpert MTB/RIFTP, FP, TN, FN, AUC, Sn, Sp
Tavaziva et al., 2022 [33]CXRLunit INSIGHT CXR (v 4.9.0)Xpert MTB/RIF, Mycobacterium tuberculosis (MTB) cultureTP, FP, TN, FN, AUC, ACC, Sn, Sp
Shen et al., 2010 [34]CXRBayesian classifierHuman readerACC
Melendez et al., 2015 [35]CXRsi-miSVM+PEDDHuman readerAUC
Pasa et al., 2019 [36]CXRCNNHuman readerAUC, ACC
Xie et al., 2020 [37]CXRRCNNHuman readerAUC, ACC, Sn, Sp
Ma et al., 2020 [38]CTU-NetSputum smearAUC, ACC, Sn, Sp, PPV, NPV
Rajpurkar et al., 2020 [39]CXRDenseNetXpert MTB/RIF, MTB cultureACC, Sn, Sp
Oloko-Oba et al., 2021 [40]CXREfficientNetsHuman readerAUC, ACC, Sn, Sp
Mamalakis et al., 2021 [41]CXRDenseNet-121,
ResNet-50
Human readerAUC, F1, precision, recall
Rajakumar et al., 2021 [42]CXRVGG16,
VGG19,
KNN
Human readerACC, Sn, Sp, NPV
Sharma et al., 2021 [43]CXRTree,
SVM,
Naïve Bayes
Composite reference standard(s)AUC, F1, CA, precision, recall
Wang et al., 2021 [44]CT3D-ResNetAFB smear, MTB cultureAUC, Sn, Sp, ACC, F1
Showkatian et al., 2022 [45]CXRConvNetHuman readerAUC, ACC, F1, precision, recall
Zhou et al., 2022 [46]CXRResNetHuman readerAUC, ACC, Sn, Sp, PPV, NPV
Rajaraman et al., 2021 [47]CXRImageNet,
VGG-16
Human readerAUC, ACC, Sn, Sp, F1, precision
Yan et al., 2021 [48]CTSeNet-ResNet-18Human readerACC, precision, recall
Zhang et al., 2021 [49]CTCBIR-CSNNComposite reference standard(s)AUC, ACC
Arzhaeva et al., 2009 [50]CXRMVDBHuman readerAUC
Jaeger et al., 2014 [51]CXRSVMHuman readerAUC, ACC
Chauhan et al., 2014 [52]CXRSVMHuman readerAUC, ACC, Sn, Sp, F1, precision
Hogeweg et al., 2015 [53]CXRRF50,
GB50,
LDA,
KNN13
MTB culture, human readerAUC
Lakhani et al., 2017 [54]CXRAlexNet,
GoogLeNet
Human readerAUC, ACC, Sn, Sp
Han et al., 2021 [55]CXRVGG16Human readerAUC, Sn, Sp
An et al., 2022 [56]CXRE-TBNet (ResNet)Human readerACC, Sn, Sp, NPV, ppv, F1
Lee et al., 2021 [57]CXREfficientNetXpert MTB/RIF, MTB culture, human readerAUC
Khatibi et al., 2021 [58]CXRCNN,
CCNSE
Human readerAUC, ACC
Kim et al., 2020 [59]CXRDCNNHuman readerAUC, Sn, Sp, NPV, PPV, F1
Feng et al., 2020 [60]CTCNNComposite reference standard(s)AUC, ACC, Sn, Sp
Hwang et al., 2019 [61]CXRCNNHuman readerAUC, Sn, Sp
Heo et al., 2019 [62]CXRI-CNN(VGG19),
D-CNN(VGG19)
Human readerAUC
Aguiar et al., 2016 [63]CXRMLPHuman readerAUC, Sn, Sp, PPV, NPV
Faruk et al., 2021 [64]CTXception,
InceptionV3,
InceptionResNetV2,
MobileNetV2
Human readerSn, precision, recall, F1
Karki et al., 2021 [65]CXRInceptionV3,
Xception
Human readerAUC
Dasanayaka et al., 2021 [66]CXRVGG16,
InceptionV3,
Ensemble
Human readerACC, Sn, Sp
Govindarajan et al., 2021 [67]CXRELM,
OSELM
Human readerSn, Sp, precision, F1
Acharya et al., 2022 [68]CXRImageNet fine-tuned normalization-free networksHuman readerSn, Sp, AUC, ACC, precision, recall
Kadry et al., 2022 [69]CXRVGG16, Fine TreeXpert MTB/RIF, Mycobacterium tuberculosis (MTB) culture, human readerSn, Sp, ACC, NPV
Kazemzadeh et al., 2023 [70]CXRNRHuman readerSn, Sp, AUC
Margarat et al., 2022 [71]CXRDBN-AMBOHuman readerSp, ACC, precision, recall, NPV
Abbreviations: CXR, chest X-ray; CT, computed tomography; CAD, computer-aided detection; CNN, convolutional neural networks; RCNN, regions with CNN features; KNN, K-nearest neighbor; VGG, visual geometry group; SVM, support vector machine; HIV, human immunodeficiency virus; DLAD, deep-learning-based automatic detection; AFB, acid-fast bacilli; MTB, Mycobacterium tuberculosis; TP, true positive; FP, false positive; TN, true negative; FN, false negative; AUC, area under the receiver operating curve; ACC, accuracy; Sn, sensitivity; Sp, specificity; CA, cluster accuracy; DBN-AMBO, deep belief network with adaptive monarch butterfly optimization.
Table 2. Subgroup analysis based on different standards.
Table 2. Subgroup analysis based on different standards.
StudiesSensitivity (95%CI)Specificity (95%CI)DOR (95%CI)AUC (95%CI)
All (23)0.91(0.89–0.93)0.65(0.55–0.75)20(13–29)0.91(0.89–0.94)
Study Design
Prospective (12)0.91(0.87–0.94)0.48(0.34–0.62)9(4–20)0.85(0.82–0.88)
Nonprospective (11)0.87(0.78–0.93)0.75(0.53–0.89)20(5–84)0.90(0.87–0.92)
Software
CAD4TB (18)0.89(0.82–0.94)0.57(0.42–0.70)11(4–30)0.83(0.80–0.86)
qXR (8)0.79(0.61–0.90)0.55(0.24–0.83)5(1–38)0.77(0.73–0.80)
Lunit INSIGHT CXR (8)0.88(0.75–0.94)0.78(0.40–0.95)25(3–211)0.91(0.88–0.93)
Reference standard
Human reader (5)0.90(0.84–0.94)0.90(0.80–0.95)77(22–269)0.95(0.93–0.97)
Xpert MTB/RIF (9)0.90(0.85–0.93)0.36(0.24–0.50)5(2–12)0.79(0.75–0.82)
AI type
Deep learning (13)0.91(0.89–0.92)0.62(0.48–0.74)16(10–23)0.91(0.88–0.93)
Machine learning (9)0.93(0.85–0.97)0.61(0.46–0.75)21(11–42)0.87(0.83–0.89)
Abbreviation: DOR, diagnostic odds ratio; AUC, area under curve.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhan, Y.; Wang, Y.; Zhang, W.; Ying, B.; Wang, C. Diagnostic Accuracy of the Artificial Intelligence Methods in Medical Imaging for Pulmonary Tuberculosis: A Systematic Review and Meta-Analysis. J. Clin. Med. 2023, 12, 303. https://doi.org/10.3390/jcm12010303

AMA Style

Zhan Y, Wang Y, Zhang W, Ying B, Wang C. Diagnostic Accuracy of the Artificial Intelligence Methods in Medical Imaging for Pulmonary Tuberculosis: A Systematic Review and Meta-Analysis. Journal of Clinical Medicine. 2023; 12(1):303. https://doi.org/10.3390/jcm12010303

Chicago/Turabian Style

Zhan, Yuejuan, Yuqi Wang, Wendi Zhang, Binwu Ying, and Chengdi Wang. 2023. "Diagnostic Accuracy of the Artificial Intelligence Methods in Medical Imaging for Pulmonary Tuberculosis: A Systematic Review and Meta-Analysis" Journal of Clinical Medicine 12, no. 1: 303. https://doi.org/10.3390/jcm12010303

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop