Next Article in Journal
Quantitative Ultrasound for Hepatic Steatosis: A Systematic Review Highlighting the Diagnostic Performance of Ultrasound-Derived Fat Fraction
Previous Article in Journal
Interactive Effect of Depression upon the Relationship Between Chlamydia trachomatis and Depression
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Impact of AI Assistance in Pneumothorax Detection on Chest Radiographs Among Readers of Varying Experience

1
Department of Radiology, Kaohsiung Veterans General Hospital, No. 386, Dazhong 1st Rd., Zuoying Dist., Kaohsiung City 813414, Taiwan
2
Department of Radiation Oncology, Kaohsiung Veterans General Hospital, No. 386, Dazhong 1st Rd., Zuoying Dist., Kaohsiung City 813414, Taiwan
3
School of Occupational Therapy, College of Medicine, National Taiwan University, No. 1, Sec. 1, Ren’ai Rd., Zhongzheng Dist., Taipei City 100233, Taiwan
4
School of Medicine, College of Medicine, National Yang Ming Chiao Tung University, No. 155, Sec. 2, Linong St., Beitou Dist., Taipei City 112304, Taiwan
5
Institute of Clinical Medicine, College of Medicine, National Yang Ming Chiao Tung University, No. 155, Sec. 2, Linong St., Beitou Dist., Taipei City 112304, Taiwan
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(20), 2639; https://doi.org/10.3390/diagnostics15202639
Submission received: 3 September 2025 / Revised: 9 October 2025 / Accepted: 17 October 2025 / Published: 19 October 2025
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

Objectives: We aimed to investigate whether AI assistance could improve the performance of pneumothorax detection on chest radiographs (CXR) by readers with varying experience from radiologists to the frontline healthcare providers, and whether AI assistance could diminish the potential confounders for readers’ detecting pneumothorax. Methods: In this retrospective, single-center, blinded, multi-reader diagnostic accuracy study, 125 CXRs were prepared from radiological information system (March 2024 to August 2024) for test. The 18 readers were composed of six groups, each had 3 persons: board-certified radiologists (Group-1), senior radiology residents (Group-2), junior radiology residents (Group-3), postgraduate year residents (Group-4), senior radiographers (Group-5), and junior radiographers (Group-6). They read the CXR independently twice, without and with AI assistance, at an interval of one month. We used receiver operating characteristic curve for performance analysis and generalized estimating equation (GEE) model for confounding factor analysis. Results: AI software alone achieved a high area under curve of 0.965 (95% CI: 0.926, 0.995). With AI assistance, the performance in all groups significantly improved (p < 0.01) especially the junior readers (the frontline healthcare providers, Group-3, 4, 6) and diminished the difference among all groups except some related to Group-1. GEE model showed that AI assistance, reader’s experience, and projection type interfere with the readers’ performance (all p < 0.05). Conclusions: AI assistance could improve the performance of pneumothorax detection by varying experience of readers, especially the frontline healthcare providers. The influence of confounders, such as reader’s experience, also be diminished by AI assistance.

1. Introduction

Pneumothorax is one of the most common thoracic emergencies, which can be life-threatening and require prompt diagnosis for clinical management. Chest radiography (CXR) is regarded as the primary imaging modality for first-line assessment. However, interpreting CXRs depends heavily on the reader’s expertise and can be challenging due to subtle imaging findings or presence of artifacts [1,2].
In recent years, artificial intelligence (AI) systems have been introduced for image interpretation, showing an improvement in diagnostic accuracy [3,4,5,6]. Nevertheless, AI models also have their limitations. Common sources of false positives for detecting pneumothorax were described in previous studies [3,7,8], including rib edges, cardiac borders, pulmonary bullae, gastric gas, and linear artifacts such as thoracic tubes or skin folds.
Several studies have evaluated the performance of AI-assisted detection in pneumothorax for radiologists [9,10,11] and clinicians [12], but few have addressed its impact on radiographers and post-graduate year (PGY) residents, who are often at the frontline of assessment and are less experienced in interpreting emergency CXRs. Furthermore, the influence of confounding factors such as pneumothorax size and skinfold artifacts on AI-assisted interpretation remains insufficiently explored.
This study aimed to assess the diagnostic performance of an AI model for pneumothorax detection on chest radiographs across healthcare providers with varying levels of expertise. Additionally, we investigated the effects of pneumothorax severity and the presence of skinfold artifacts as potential confounding factors within this context.

2. Materials and Methods

The ethics committee of Kaohsiung Veterans General Hospital (approval number KSVGH-25-CT3-05) approved this study. The need for informed consent was waived. No authors had a conflict of interest to disclose.

2.1. Image Dataset

Inpatient CXRs from the radiology information system of Kaohsiung Veterans General Hospital from March 2024 to August 2024 were retrospectively chosen by a 25-year-experience thoracic radiologist to generate a dataset of 125 radiographs without or with different severity of pneumothorax. Artifacts such as skin folds were intentionally overrepresented to enrich the dataset and ensure adequate inclusion of diagnostically challenging cases. Enrichment strategy was used to retrospectively select CXR from inpatient department and portable CXR with key words of “pneumothorax” and “skin fold” form the RIS reports, respectively, to enroll sufficient and balanced cases number for each group. CXRs with uncertain medical records or lack of follow-up CXRs were excluded. The severity of pneumothorax was defined as small (<2 cm) and large (>2 cm) according to the British Thoracic Society (BTS) guidelines (2010) [13] (Figure 1).

2.2. AI Software

An FDA-approved AI assistance CXR software (Lunit Insight CXR ver 3.1.4.1; Lunit Inc., Seoul, Republic of Korea) was applied in this study. This software is designed to detect 10 common abnormalities (pneumothorax, mediastinal widening, pneumoperitoneum, nodule/mass, consolidation, pleural effusion, linear atelectasis, fibrosis, calcification and cardiomegaly) in CXRs. The AI software assists readers to interpret CXRs by marking the suspected abnormal areas, labeling the type of abnormality, and generating heat maps indicating the probability of each abnormality on a second capture of CXR, with likelihood percentages ranging from 1% to 100%. Its performance has been validated in previous studies [14,15].

2.3. Readers

Eighteen readers participated in this study. They were composed of six groups, each consisting of three participants: board-certified radiologists (Group 1), senior radiology residents (Group 2, in their 4th year of residency), junior radiology residents (Group 3, in their 1st to 2nd year of residency), postgraduate year (PGY) residents (Group 4), senior radiographers with more than 5 years of experience (Group 5), and junior radiographers with less than 1 year of experience (Group 6). For analysis of impact of years of experience, these 6 groups were further categorized into senior group (Group 1, 2, 5) and junior group (Group 3, 4, 6).

2.4. Scoring

The CXRs were presented in a randomly arranged order, and all readers were blinded to the clinical information. Each reader independently assessed the presence of pneumothorax using a 5-point Likert scale, rating the likelihood of pneumothorax from 0 to 4 (0 = definitely no, 1 = probably no, 2 = equivocal, 3 = probably yes, 4 = definitely yes), without AI assistance during the first reading session. After a one-month interval, the same process was repeated at another randomly arranged order with AI assistance, i.e., the CXR and heat map of abnormality were present simultaneously for interpretation.

2.5. Reference Standard

The reference standard was the judgment of a 25-year-experience board-certified thoracic radiologist, based on available serial CXR, CT scan and clinical chart records.

2.6. Statistical Analyses

We used R software (version 4.4.3) to perform all statistical analyses. Diagnostic accuracy for both AI model and readers was assessed using the Area Under the Receiver Operating Characteristic Curve (AUROC; AUC). Mean AUCs with and without AI assistance were compared using two-sided 95% confidence intervals (CIs). Pairwise AUC comparisons were performed using DeLong’s test with Holm’s correction for multiple comparisons, with statistical significance defined as p < 0.05. The secondary outcome was to analyze the probable influence of those potential confounding factors. A generalized estimating equations (GEE) model was used to account for the within-subject correlation arising from multiple diagnostic assessments performed on the same patient (i.e., repeated observations by different readers). The results were demonstrated as parameter estimates along with their corresponding standard errors (SE) and p-value. A p-value of less than 0.05 indicated a statistical significance. Three primary variables were included in the model: (1) the presence or absence of AI assistance, (2) reader experience level (senior vs. junior), and (3) the presence of skinfold artifacts (coded as 0 for absent and 1 for present). Two additional variables were entered as covariates to control for potential confounding: pneumothorax volume (coded as 0 for no pneumothorax, 1 for small, and 2 for large pneumothorax) and projection type (coded as 0 for PA, 1 for AP, and 2 for portable views). All main effects and interaction terms among the primary variables were included in the analysis.

3. Results

3.1. Data Characteristics

A total of 125 CXRs from 125 patients were selected. The demography and CXR characteristics details are listed in Table 1. There were 88 male and 37 female patients, with a mean age of 66.4 years. Among these, 73 (58.4%) had no pneumothorax and 52 (41.6%) had pneumothorax (30 small, 22 large). 61 were obtained in supine AP projection, 16 in standing AP and 48 were in standing PA views. Skinfold artifacts were present in 41 CXRs (Table 1).

3.2. Performance of AI Alone and Compared with Unassisted Readers

The AUC of AI alone was 0.965 (95% CI: 0.926, 0.995), which was significantly higher than (p < 0.05) all the readers’ group except for Group 1 (Figure 2; Table 2a,b).

3.3. Readers’ Performance Without AI Assistance

The AUC of each group without AI assistance was Group 1, 0.908 (95% CI: 0.874, 0.935); Group 2, 0.850 (95% CI: 0.810, 0.885); Group 3, 0.777 (95% CI: 0.731, 0.818); Group 4, 0.711 (95% CI: 0.662, 0.757); Group 5, 0.840 (95% CI: 0.798, 0.875); Group 6, 0.607 (95% CI: 0.556, 0.657) (Figure 3a–f). Significant differences were observed between Group 1 and Groups 3, 4, and 6; between Group 2 and Groups 4 and 6; between Group 3 and Group 6; between Group 4 and Group 6; and between Group 5 and Group 6 (p < 0.01) (Table 2a).

3.4. Readers’ Performance with AI Assistance

With AI assistance, the performance of all groups improved (Figure 3a–f and Figure 4a,b). The AUC of each group with AI assistance was of Group 1, 0.933 (95% CI, 0.902 to 0.95); Group 2, 0.889 (95% CI, 0.853 to 0.919); Group 3, 0.864 (95% CI, 0.825 to 0.818); Group 4, 0.832 (95% CI, 0.79 to 0.868); Group 5, 0.879 (95% CI, 0.841, 0.910); Group 6, 0.837 (95% CI, 0.796 to 0.873). No significant differences were found between Group 1 and Groups 2, 3, and 5, and no significant pairwise differences were observed among Groups 2 to 6 after AI assistance (Table 2b).
Take two supine chest radiographs for examples (Figure 5). Figure 5a is a case of pneumothorax presents with the deep sulcus sign. Without AI assistance, apart from the board-certified radiologist, all other groups were not definite certain about the presence of pneumothorax, and junior radiographers were even unable to detect it. However, with AI assistance, all readers were able to confidently identify pneumothorax, reaching a median scale of 4 in every subgroup. Figure 5b is a case of supine radiograph with skinfold artifacts which mimic pneumothorax for junior groups. Even with AI assistance, some junior radiographers still insisted on the presence of pneumothorax. These two cases, respectively, show that AI assistance improved the false negative readings (Figure 5a) and the false positive readings (Figure 5b) in detecting pneumothorax, especially for the junior group (Group 4, Group 6).

3.5. GEE Model of Pneumothorax Diagnostic Accuracy of Readers

3.5.1. Main Effects

Analysis of the main effects revealed several significant findings. First, AI assistance significantly improved diagnostic accuracy (estimate = 0.76, SE = 0.14, Wald = 31.2, p < 0.001). Second, reader experience was also a strong predictor, with senior readers outperforming junior readers (estimate = 0.94, SE = 0.13, Wald = 49.5, p < 0.001). Among covariates, diagnostic accuracy varied by pneumothorax severity relative to large pneumothorax. When no pneumothorax was present, accuracy was higher (ptx012 = 0: estimate = 1.56, SE = 0.13, Wald = 147.2, p < 0.001), whereas small pneumothorax was more difficult to identify correctly (ptx012 = 1: estimate = −1.16, SE = 0.11, Wald = 116.1, p < 0.001). For projection type of image (relative to the reference category PA), the AP view was not statistically significant (estimate = 0.11, SE = 0.16, p = 0.50), while portable view was associated with lower accuracy (estimate = −0.33, SE = 0.10, Wald = 9.9, p = 0.002). The presence of skinfold artifacts (SF) did not show a significant main effect in this model (sf = 1: estimate = −0.16, SE = 0.19, p = 0.40). See Table 3 for full results. (Table 3). To further examine the issue of skinfold artifacts (SF), we conducted a descriptive analysis of diagnostic outcome distributions stratified by SF status, Seniority, and AI assistance. Without AI assistance, the false positive rate (FPR) was 5% in the junior group and 1% in the senior group when SF is present, and 3% versus 1% without SF. The false negative rate (FNR) was 7% in the junior group and 5% in the senior group with SF, and 15% versus 9% without SF. With AI assistance, the FPR decreased to 2% in the junior group and 1% in the senior group with SF, and 2% versus 1% without SF. The FNR was 5% versus 4% with SF, and 10% versus 7% without SF. (Table 4).

3.5.2. Two-Way Interactions

A significant two-way interaction was observed between AI assistance and reader experience (estimate = −0.43, SE = 0.20, Wald = 4.7, p = 0.03). This negative interaction indicates that the improvement associated with AI assistance was greater for junior readers than for senior readers (i.e., AI narrows the senior–junior performance gap). Interactions involving skinfold artifacts were not significant (AI assistance × Skinfold artifacts: p = 0.90; Senior group × Skinfold artifacts: p = 0.29). Predicted accuracies illustrating these patterns across AI aid and skinfold status by reader experience are shown in Figure 6, with 95% confidence intervals.

3.5.3. Higher-Order Interactions

No higher-order interaction reached statistical significance. In particular, the three-way interaction among AI assistance, reader experience, and skinfold status was not significant (AI assistance × Senior group × Skinfold artifacts: p = 0.50), suggesting no additional modifying effect beyond the AI × experience interaction.

4. Discussion

Our results confirmed that the applied AI model is a reliable tool for detecting pneumothorax. With AI assistance, diagnostic variability among the six reader groups was markedly reduced, including, for the first time in the literature report, the radiographer and PGY residents. The improvement was particularly pronounced among junior readers, who are often at the first line in clinical workflow.
Previous studies have reported high diagnostic performance for AI algorithms for pneumothorax detection on CXR, with AUC values ranging from 0.88 to 0.98 [12,16,17,18,19,20,21]. A meta-analysis of 23 studies conducted in 2024 revealed relatively high sensitivity and specificity for pneumothorax detection using deep learning algorithms [21], with the pooled sensitivity and specificity being 87% (95% CI: 0.81, 0.92) and 95% (95% CI: 0.92, 0.97%), respectively. Our observed AUC of 0.965 (95% CI: 0.926, 0.995) for pneumothorax detection aligns with previously published validations of Lunit Insight CXR, such as Ahn et al. [22], who reported an AUC of 0.999 for AI stand-alone detection for pneumothorax, and van Beek et al. (2023) [23], who found an AUC of 0.954 in primary and emergency settings. These results indicate that the excellent diagnostic performance in our study is consistent with real-world validation studies.
The benefits of AI-assisted interpretation have been widely reported for radiologists [9,10,11] and clinicians [12], particularly for less experienced readers [11,12]. However, few studies have explored its impact on PGY residents and radiographers, who play a critical role as frontline providers.
In our results, readers’ diagnostic accuracy without AI assistance closely correlated with experience level. Notably, senior radiographers performed comparably to senior radiology residents, suggesting that accumulated clinical exposure may compensate for formal training differences. AI assistance effectively elevated junior radiographers’ performance to a level comparable with their senior counterparts. This improvement is clinically meaningful, as radiographers and PGY residents often serve as the earliest detectors of critical findings such as pneumothorax in urgent settings.
To enhance communication efficiency in emergency radiology, the early alert role of radiographers has been emphasized in several studies [24,25,26]. In our institution, radiographers are trained to assess image quality and to promptly recognize critical findings such as barotrauma or tube malposition. There was meta-analytic evidence supporting that radiographers can accurately interpret plain radiographs, with performance comparable to radiologists in accident and emergency settings [27,28]. However, most previous research focused on trauma or lung cancer screening, with limited data on pneumothorax detection in critical care and almost none on very junior radiographers, who are often responsible for portable CXRs. To our knowledge, this is the first study examining pneumothorax detection by junior radiographers and PGY residents, comparing them to senior readers while also evaluating the effect of AI assistance. AI support enabled junior radiographers to perform at a level comparable to senior groups, reinforcing their potential to act as first responders in identifying pneumothorax. Khalaf et al. (2025) [29] investigated the knowledge, expectations, and attitudes toward AI integration in medical imaging in Kuwait. While most respondents showed a positive outlook, the study identified a need for greater AI-oriented education and training. In line with our findings, AI could function not only as a diagnostic support tool but also as an educational resource. With proper integration into training curricula, AI systems may accelerate learning curves, improve diagnostic confidence, and empower radiographers and PGY residents to take more active roles in emergency and routine image interpretation.
According to our clinical experience, certain confounding factors might bother readers when interpreting CXRs for detecting pneumothorax. In fact, this retrospective study was inspired by the frequent consultation of CXRs that have skinfold artifacts, and therefore, we intentionally enrolled more such cases (56/125 cases). However, GEE analysis revealed that AI assistance, reader’s experience, pneumothorax amount, and portable view were the main effects, while the presence of skinfold artifacts did not seem to significantly influence the diagnostic accuracy. A descriptive analysis of diagnostic outcomes stratified by skinfold (SF) status, reader’s seniority, and AI assistance showed that without AI assistance, junior readers showed higher false positive rates (FPR) and false negative rates (FNR) than senior readers, regardless of the presence of skinfold artifacts. With AI support, both rates decreased, and the performance gap between groups narrowed, suggesting that AI had a particularly beneficial effect on less experienced readers. Interestingly, the presence of skinfold artifacts appeared to slightly reduce the FNR, possibly because the visible artifact prompted readers to more carefully assess for pneumothorax, thereby reducing false negatives in both groups. However, skinfold artifacts were not a statistically significant factor in the revised GEE model. This aligns with prior evidence suggesting that well-trained AI models are relatively resistant to common image artifacts. For instance, Hillis et al. (2022) [17] reported consistently high AUCs for pneumothorax detection in both cases with (AUC = 0.964) and without (AUC = 0.982) skinfold artifacts, indicating robust algorithmic performance. These results support the robustness of AI models to common imaging confounders and emphasize their potential as both supportive and educational tools, particularly in resource-limited or high-acuity environments where less experienced readers often interpret chest radiographs.
For interaction analysis, the results revealed a significant two-way interaction between AI assistance and reader experience, indicating that AI provided a greater diagnostic benefit for junior readers compared with senior readers. This finding suggests that AI support can effectively narrow the performance gap between less experienced and more experienced readers, particularly in detecting subtle or small pneumothoraxes. Such results align with previous research showing that AI-assisted image interpretation improves diagnostic accuracy mainly among less experienced clinicians, while providing limited additional benefit to experts who already perform at a high baseline level [12,22].
The study has several limitations. First, the dataset was intentionally enriched with challenging cases, meaning results may differ from those in a true screening population. Nevertheless, since this study was inspired by frequent clinical misinterpretations involving skinfold artifacts, a balanced case distribution was necessary to assess the statistical impact of these artifacts. Second, the study’s relatively small sample size and retrospective single-center design may limit generalizability. Third, all readers were from the same institution, which introduces the possibility of institutional bias. Future multicenter and prospective studies are warranted to validate these findings. Nonetheless, our inclusion of 18 readers across multiple professional levels provides valuable preliminary evidence to inform future large-scale designs and potential real-world implementation of AI-assisted interpretation within PACS systems.

5. Conclusions

We confirmed that AI-assisted reading improves pneumothorax detection accuracy on CXR across diverse experience levels and remains robust against confounding artifacts. This study extends AI validation to the frontline of clinical workflow, encompassing radiographers and PGY residents for the first time. We believe these extensions closely reflect the impact of AI-assisted CXR reading in real-world scenarios nowadays, and our findings suggest that AI not only enhances diagnostic accuracy but may also serve as a training and decision-support tool to strengthen emergency radiological care. Future prospective randomized controlled trials are recommended to assess the cost-effectiveness and scalability of integrating AI assistance into real-world frontline practice.

Author Contributions

Conceptualization, Y.-C.C. and M.-T.W.; Data curation, C.-W.H. and M.-T.W.; Formal analysis, Y.-J.J. and M.-T.W.; Funding acquisition, M.-T.W.; Investigation, C.-W.H. and Y.-L.W.; Methodology, Y.-C.C. and M.-T.W.; Project administration, M.-T.W.; Resources, M.-T.W.; Software, Y.-J.J.; Supervision, M.-T.W.; Validation, C.-W.H.; Visualization, C.-W.H. and M.-T.W.; Writing—original draft, C.-W.H.; Writing—review and editing, M.-T.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science and Technology Council, grant number NSTC 112-2314-B-A49-063-MY3, and the Kaohsiung Veterans General Hospital, grant number KSVGH-113-115.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Kaohsiung Veterans General Hospital (protocol code KSVGH-25-CT3-05 and date of approval was 13 January 2024).

Informed Consent Statement

Patient consent was waived due to a retrospective study design. Written informed consent was waived by the Institutional Review Board.

Data Availability Statement

The data presented in this study are available on request from the corresponding author; the data are owned by the hospital and cannot be publicly released without formal permission.

Acknowledgments

We thank all radiologists, radiographers, and PGY residents who participated in this study: Yung-Lin Chen, Yi-Chin Chang, Wei-Hsin Chiang, Chien-Yang Liao, Pin-Hsueh Wann, Yao-Hung Yeh, En-Chia Mai, Po-Yu Huang, Po-An Chen, Chun-Yi Chiu, Ying-Ling Chen, Shu-Ya Kuo, Wei Ting Li, Chien-Wen Chen, Yih-Lan Suen, Po-Chen Wang, Nai-Yuan Hsu, and Po-Hsun Huang.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study, in the collection, analysis, or interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
AUCArea Under Curve
CXRChest Radiograph
FNRFalse Negative Rate
FPRFalse Positive Rate
PGYPostgraduate Year Residents
PTXPneumothorax
ROCReceiver Operating Characteristic Curve
SFSkin fold

References

  1. Gefter, W.B.; Post, B.A.; Hatabu, H. Commonly Missed Findings on Chest Radiographs: Causes and Consequences. Chest 2023, 163, 650–661. [Google Scholar] [CrossRef] [PubMed]
  2. Kaviani, P.; Kalra, M.K.; Digumarthy, S.R.; Gupta, R.V.; Dasegowda, G.; Jagirdar, A.; Gupta, S.; Putha, P.; Mahajan, V.; Reddy, B.; et al. Frequency of Missed Findings on Chest Radiographs (CXRs) in an International, Multicenter Study: Application of AI to Reduce Missed Findings. Diagnostics 2022, 12, 2382. [Google Scholar] [CrossRef] [PubMed]
  3. Kao, C.Y.; Lin, C.Y.; Chao, C.C.; Huang, H.S.; Lee, H.Y.; Chang, C.M.; Sung, K.; Chen, T.R.; Chiang, P.C.; Huang, L.T.; et al. Automated Radiology Alert System for Pneumothorax Detection on Chest Radiographs Improves Efficiency and Diagnostic Performance. Diagnostics 2021, 11, 1182. [Google Scholar] [CrossRef]
  4. Taylor, A.G.; Mielke, C.; Mongan, J. Automated detection of moderate and large pneumothorax on frontal chest X-rays using deep convolutional neural networks: A retrospective study. PLoS Med. 2018, 15, e1002697. [Google Scholar] [CrossRef]
  5. Lind Plesner, L.; Muller, F.C.; Brejnebol, M.W.; Laustrup, L.C.; Rasmussen, F.; Nielsen, O.W.; Boesen, M.; Brun Andersen, M. Commercially Available Chest Radiograph AI Tools for Detecting Airspace Disease, Pneumothorax, and Pleural Effusion. Radiology 2023, 308, e231236. [Google Scholar] [CrossRef]
  6. Mones Alamri, A. Artificial Intelligence in Radiology: Current Applications and Future Prospects. Kurd. Stud. 2021, 9, 161–163. [Google Scholar] [CrossRef]
  7. Lee, S.; Kim, E.K.; Han, K.; Ryu, L.; Lee, E.H.; Shin, H.J. Factors for increasing positive predictive value of pneumothorax detection on chest radiographs using artificial intelligence. Sci. Rep. 2024, 14, 19624. [Google Scholar] [CrossRef]
  8. Rueckel, J.; Trappmann, L.; Schachtner, B.; Wesp, P.; Hoppe, B.F.; Fink, N.; Ricke, J.; Dinkel, J.; Ingrisch, M.; Sabel, B.O. Impact of Confounding Thoracic Tubes and Pleural Dehiscence Extent on Artificial Intelligence Pneumothorax Detection in Chest Radiographs. Investig. Radiol. 2020, 55, 792–798. [Google Scholar] [CrossRef]
  9. Jin, K.N.; Kim, E.Y.; Kim, Y.J.; Lee, G.P.; Kim, H.; Oh, S.; Kim, Y.S.; Han, J.H.; Cho, Y.J. Diagnostic effect of artificial intelligence solution for referable thoracic abnormalities on chest radiography: A multicenter respiratory outpatient diagnostic cohort study. Eur. Radiol. 2022, 32, 3469–3479. [Google Scholar] [CrossRef] [PubMed]
  10. Hunter, J.G.; Bera, K.; Shah, N.; Bukhari, S.M.A.; Marshall, C.; Caovan, D.; Rosipko, B.; Gupta, A. Real-World Performance of Pneumothorax-Detecting Artificial Intelligence Algorithm and its Impact on Radiologist Reporting Times. Acad. Radiol. 2024, 32, P1165–P1174. [Google Scholar] [CrossRef]
  11. Bennani, S.; Regnard, N.E.; Ventre, J.; Lassalle, L.; Nguyen, T.; Ducarouge, A.; Dargent, L.; Guillo, E.; Gouhier, E.; Zaimi, S.H.; et al. Using AI to Improve Radiologist Performance in Detection of Abnormalities on Chest Radiographs. Radiology 2023, 309, e230860. [Google Scholar] [CrossRef]
  12. Novak, A.; Ather, S.; Gill, A.; Aylward, P.; Maskell, G.; Cowell, G.W.; Espinosa Morgado, A.T.; Duggan, T.; Keevill, M.; Gamble, O.; et al. Evaluation of the impact of artificial intelligence-assisted image interpretation on the diagnostic performance of clinicians in identifying pneumothoraces on plain chest X-ray: A multi-case multi-reader study. Emerg. Med. J. 2024, 41, 602–609. [Google Scholar] [CrossRef] [PubMed]
  13. MacDuff, A.; Arnold, A.; Harvey, J. Management of spontaneous pneumothorax: British Thoracic Society Pleural Disease Guideline 2010. Thorax 2010, 65 (Suppl. 2), ii18–ii31. [Google Scholar] [CrossRef]
  14. Nam, J.G.; Kim, M.; Park, J.; Hwang, E.J.; Lee, J.H.; Hong, J.H.; Goo, J.M.; Park, C.M. Development and validation of a deep learning algorithm detecting 10 common abnormalities on chest radiographs. Eur. Respir. J. 2021, 57, 2003061. [Google Scholar] [CrossRef] [PubMed]
  15. Jang, S.; Song, H.; Shin, Y.J.; Kim, J.; Kim, J.; Lee, K.W.; Lee, S.S.; Lee, W.; Lee, S.; Lee, K.H. Deep Learning-based Automatic Detection Algorithm for Reducing Overlooked Lung Cancers on Chest Radiographs. Radiology 2020, 296, 652–661. [Google Scholar] [CrossRef]
  16. Chassagnon, G.; Soyer, P. Artificial Intelligence for the Detection of Pneumothorax on Chest Radiograph: Not yet the Panacea. Can. Assoc. Radiol. J. 2024, 75, 458–459. [Google Scholar] [CrossRef] [PubMed]
  17. Hillis, J.M.; Bizzo, B.C.; Mercaldo, S.; Chin, J.K.; Newbury-Chaet, I.; Digumarthy, S.R.; Gilman, M.D.; Muse, V.V.; Bottrell, G.; Seah, J.C.Y.; et al. Evaluation of an Artificial Intelligence Model for Detection of Pneumothorax and Tension Pneumothorax in Chest Radiographs. JAMA Netw. Open 2022, 5, e2247172. [Google Scholar] [CrossRef]
  18. Sugibayashi, T.; Walston, S.L.; Matsumoto, T.; Mitsuyama, Y.; Miki, Y.; Ueda, D. Deep learning for pneumothorax diagnosis: A systematic review and meta-analysis. Eur. Respir. Rev. 2023, 32, 220259. [Google Scholar] [CrossRef]
  19. Thian, Y.L.; Ng, D.; Hallinan, J.; Jagmohan, P.; Sia, S.Y.; Tan, C.H.; Ting, Y.H.; Kei, P.L.; Pulickal, G.G.; Tiong, V.T.Y.; et al. Deep Learning Systems for Pneumothorax Detection on Chest Radiographs: A Multicenter External Validation Study. Radiol. Artif. Intell. 2021, 3, e200190. [Google Scholar] [CrossRef]
  20. Iqbal, T.; Shaukat, A.; Akram, M.U.; Mustansar, Z.; Khan, A. Automatic Diagnosis of Pneumothorax from Chest Radiographs: A Systematic Literature Review. IEEE Access 2021, 9, 145817–145839. [Google Scholar] [CrossRef]
  21. Katzman, B.D.; Alabousi, M.; Islam, N.; Zha, N.; Patlas, M.N. Deep Learning for Pneumothorax Detection on Chest Radiograph: A Diagnostic Test Accuracy Systematic Review and Meta Analysis. Can. Assoc. Radiol. J. 2024, 75, 525–533. [Google Scholar] [CrossRef]
  22. Ahn, J.S.; Ebrahimian, S.; McDermott, S.; Lee, S.; Naccarato, L.; Di Capua, J.F.; Wu, M.Y.; Zhang, E.W.; Muse, V.; Miller, B.; et al. Association of Artificial Intelligence-Aided Chest Radiograph Interpretation with Reader Performance and Efficiency. JAMA Netw. Open 2022, 5, e2229289. [Google Scholar] [CrossRef]
  23. van Beek, E.J.R.; Ahn, J.S.; Kim, M.J.; Murchison, J.T. Validation study of machine-learning chest radiograph software in primary and emergency medicine. Clin. Radiol. 2023, 78, 1–7. [Google Scholar] [CrossRef]
  24. Brealey, S.; Scally, A.; Hahn, S.; Thomas, N.; Godfrey, C.; Coomarasamy, A. Accuracy of radiographer plain radiograph reporting in clinical practice: A meta-analysis. Clin. Radiol. 2005, 60, 232–241. [Google Scholar] [CrossRef]
  25. Cheung, T.; Harianto, H.; Spanger, M.; Young, A.; Wadhwa, V. Low accuracy and confidence in chest radiograph interpretation amongst junior doctors and medical students. Intern. Med. J. 2018, 48, 864–868. [Google Scholar] [CrossRef] [PubMed]
  26. Eisen, L.A.; Berger, J.S.; Hegde, A.; Schneider, R.F. Competency in chest radiography. A comparison of medical students, residents, and fellows. J. Gen. Intern. Med. 2006, 21, 460–465. [Google Scholar] [CrossRef] [PubMed]
  27. Woznitza, N.; Ghimire, B.; Devaraj, A.; Janes, S.M.; Piper, K.; Rowe, S.; Bhowmik, A.; Hayes, N.; Togher, D.; Arumalla, N.; et al. Impact of radiographer immediate reporting of X-rays of the chest from general practice on the lung cancer pathway (radioX): A randomised controlled trial. Thorax 2023, 78, 890–894. [Google Scholar] [CrossRef] [PubMed]
  28. Woznitza, N.; Piper, K.; Rowe, S.; Bhowmik, A. Immediate reporting of chest X-rays referred from general practice by reporting radiographers: A single centre feasibility study. Clin. Radiol. 2018, 73, 507.e1–507.e8. [Google Scholar] [CrossRef]
  29. Khalaf, A.; Alshammari, M.; Zayed, H.; Emnawer, M.; Esfahani, A. Exploring Radiographers’ Readiness for Artificial Intelligence in Kuwait: Insights and Applications. Health Sci. Rep. 2025, 8, e70465. [Google Scholar] [CrossRef]
Figure 1. Study flowchart.
Figure 1. Study flowchart.
Diagnostics 15 02639 g001
Figure 2. ROC curve of AI detecting pneumothorax.
Figure 2. ROC curve of AI detecting pneumothorax.
Diagnostics 15 02639 g002
Figure 3. ROC curve of each group, with and without AI-assistance (a) Group 1, (b) Group 2, (c) Group 3, (d) Group 4, (e) Group 5, (f) Group 6. Group 1, board-certified radiologists; 2, senior radiology residents; 3, junior radiology residents; 4, postgraduate year residents; 5, senior radiographers; 6, junior radiographers.
Figure 3. ROC curve of each group, with and without AI-assistance (a) Group 1, (b) Group 2, (c) Group 3, (d) Group 4, (e) Group 5, (f) Group 6. Group 1, board-certified radiologists; 2, senior radiology residents; 3, junior radiology residents; 4, postgraduate year residents; 5, senior radiographers; 6, junior radiographers.
Diagnostics 15 02639 g003
Figure 4. (a) The comparison of ROC curve of AI and each group without AI assistance; (b) The comparison of ROC curve of each group with AI assistance.
Figure 4. (a) The comparison of ROC curve of AI and each group without AI assistance; (b) The comparison of ROC curve of each group with AI assistance.
Diagnostics 15 02639 g004
Figure 5. Two examples to illustrate the impact of AI assistance on chest radiograph reading. (a) Supine chest radiograph of a 73-year-old male with massive pneumothorax (deep sulcus sign); (b) Supine chest radiograph of a 77-year-old man with skin fold artifact. Left panel, the native radiograph; right panel, with AI-heat maps and labels; middle panel: the score sheet without and with AI of each group. Group 1, board-certified radiologists; 2, senior radiology residents; 3, junior radiology residents; 4, postgraduate year residents; 5, senior radiographers; 6, junior radiographers. med: median; min–max: minimal to maximal.
Figure 5. Two examples to illustrate the impact of AI assistance on chest radiograph reading. (a) Supine chest radiograph of a 73-year-old male with massive pneumothorax (deep sulcus sign); (b) Supine chest radiograph of a 77-year-old man with skin fold artifact. Left panel, the native radiograph; right panel, with AI-heat maps and labels; middle panel: the score sheet without and with AI of each group. Group 1, board-certified radiologists; 2, senior radiology residents; 3, junior radiology residents; 4, postgraduate year residents; 5, senior radiographers; 6, junior radiographers. med: median; min–max: minimal to maximal.
Diagnostics 15 02639 g005aDiagnostics 15 02639 g005b
Figure 6. Panels represent the presence (“Yes”) or absence (“No”) of AI assistance. The x-axis denotes the presence of skinfold artifacts (0 = absent, 1 = present), and lines depict diagnostic accuracy for junior (red) and senior (blue) readers. Vertical bars represent 95% confidence intervals of the predicted values. The model controlled for pneumothorax severity (ptx012) and projection type.
Figure 6. Panels represent the presence (“Yes”) or absence (“No”) of AI assistance. The x-axis denotes the presence of skinfold artifacts (0 = absent, 1 = present), and lines depict diagnostic accuracy for junior (red) and senior (blue) readers. Vertical bars represent 95% confidence intervals of the predicted values. The model controlled for pneumothorax severity (ptx012) and projection type.
Diagnostics 15 02639 g006
Table 1. Patient demographics.
Table 1. Patient demographics.
Inpatient CXRs
(n = 125, 125 CXRs from 125 Patients)
Mean age ± standard deviation66.36 ± 19.2
Gender
Male88 (70.4%)
Female37 (29.6%)
Pneumothorax
Absent73 (58.4%)
Present52 (41.6%)
Small (<2 cm)30 (57.7%)
Large (≥2 cm)22 (42.3%)
Projection type
Upright PA48 (38.4%)
Upright AP16 (12.8%)
Supine AP61 (48.8%)
Skinfold artifacts
Present56 (44.8%)
No pneumothorax38 (67.9%)
Small pneumothorax12 (21.4%)
Large pneumothorax6 (10.7%)
Absent69 (55.2%)
Table 2. The comparison of AUC of each group and p value of difference (with Holm’s correction) between AI and each group, (a) without and (b) with AI assistance.
Table 2. The comparison of AUC of each group and p value of difference (with Holm’s correction) between AI and each group, (a) without and (b) with AI assistance.
(a)
AIGroup 1Group 2Group 3Group 4Group 5Group 6
AUC0.9650.9070.850.7760.7130.840.606
p value
Group 10.14
Group 2<0.010.14
Group 3<0.01<0.010.11
Group 4<0.01<0.01<0.010.15
Group 5<0.010.090.710.15<0.01
Group 6<0.01<0.01<0.01<0.010.04<0.01
(b)
Group 1Group 2Group 3Group 4Group 5Group 6
AUC 0.9330.8890.8640.8320.8790.837
p value
Group 1
Group 2 0.49
Group 3 0.07>0.99
Group 4 <0.010.40>0.99
Group 5 0.22>0.99>0.990.76
Group 6 <0.010.49>0.99>0.990.94
Table 3. GEE analysis of confounding factors.
Table 3. GEE analysis of confounding factors.
EstimateSEWaldSig.
(Intercept)0.650.1230.0***
AI assistance (Yes)0.760.1431.2***
Senior group0.940.1349.5***
No pneumothorax1.560.13147.2***
Small pneumothorax−1.160.11116.1***
Skinfold artifacts−0.160.190.7
Projection type: AP0.110.160.4
Projection type: portable−0.330.109.9**
AI assistance × Senior group−0.430.24.7*
AI assistance × SF0.020.270.0
Senior group × SF0.290.271.2
AI assistance × Senior group × Skinfold artifacts−0.270.400.4
Note. ‘:’ indicates the interaction between variables; * p < 0.05, ** p < 0.01, *** p < 0.001.
Table 4. Descriptive analysis of diagnostic outcome distributions stratified by skinfold artifacts, reader’s seniority and AI assistance.
Table 4. Descriptive analysis of diagnostic outcome distributions stratified by skinfold artifacts, reader’s seniority and AI assistance.
SeniorityAI AidDiagnosis OutcomeSF (−)SF (+)
SeniorYesTrue negative27%30%
False negative7%4%
True positive20%10%
False positive1%1%
NoTrue negative27%30%
False negative9%5%
True positive18%9%
False positive1%1%
JuniorYesTrue negative26%28%
False negative10%5%
True positive18%10%
False positive2%2%
NoTrue negative25%26%
False negative15%7%
True positive13%8%
False positive3%5%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ho, C.-W.; Wu, Y.-L.; Chen, Y.-C.; Ju, Y.-J.; Wu, M.-T. Impact of AI Assistance in Pneumothorax Detection on Chest Radiographs Among Readers of Varying Experience. Diagnostics 2025, 15, 2639. https://doi.org/10.3390/diagnostics15202639

AMA Style

Ho C-W, Wu Y-L, Chen Y-C, Ju Y-J, Wu M-T. Impact of AI Assistance in Pneumothorax Detection on Chest Radiographs Among Readers of Varying Experience. Diagnostics. 2025; 15(20):2639. https://doi.org/10.3390/diagnostics15202639

Chicago/Turabian Style

Ho, Chen-Wei, Yu-Lun Wu, Yi-Chun Chen, Yu-Jeng Ju, and Ming-Ting Wu. 2025. "Impact of AI Assistance in Pneumothorax Detection on Chest Radiographs Among Readers of Varying Experience" Diagnostics 15, no. 20: 2639. https://doi.org/10.3390/diagnostics15202639

APA Style

Ho, C.-W., Wu, Y.-L., Chen, Y.-C., Ju, Y.-J., & Wu, M.-T. (2025). Impact of AI Assistance in Pneumothorax Detection on Chest Radiographs Among Readers of Varying Experience. Diagnostics, 15(20), 2639. https://doi.org/10.3390/diagnostics15202639

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop