You are currently viewing a new version of our website. To view the old version click .
Cancers
  • Article
  • Open Access

4 December 2025

Performance of an Artificial Intelligence Support System on Screening Mammography Cases Proceeding to Stereotactic Biopsy

,
,
,
,
,
and
Department of Diagnostic, Molecular and Interventional Radiology, Icahn School of Medicine at Mount Sinai (ISMMS), 1470 Madison Ave., New York, NY 10029, USA
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Radiology for the Detection, Diagnosis, and Management of Breast Cancer

Simple Summary

Digital breast tomosynthesis (DBT) for screening mammography is widely adopted in the United States and deep learning-based artificial intelligence (AI) algorithms have been developed to increase accuracy and efficiency in breast cancer detection. Few studies have evaluated the performance of AI decision support in real-world settings, especially on DBT exams. Transpara AI decision support was recently integrated into the clinical workflow for screening mammography at our institution and our aim was to assess its standalone performance in detecting breast cancer on DBT cases proceeding to stereotactic biopsy. This article presents performance metrics for Transpara AI along with representative cases of AI misses and false positives to provide insights into how radiologist–AI synergy can improve patient care.

Abstract

Background/Objective: The objective was to evaluate the standalone performance of an AI system, Transpara 1.7.1 (ScreenPoint Medical), in screening mammography cases proceeding to stereotactic biopsy using histopathological results as ground truth. Methods: This retrospective study included 202 asymptomatic female patients (mean age: 57.8 years) who underwent stereotactic biopsy at a multicenter academic institution between October 2022 and September 2023 with a preceding screening mammogram within 14 months. Transpara AI risk scores were compared to pathology results (benign versus malignant). Performance metrics for AI including sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and area under the receiver operating characteristic curve (AUC) were calculated. Results: Transpara AI classified 20 of 39 malignant findings (51%) as elevated risk compared with 50 of 211 total findings (24%). AI score was positively correlated with malignancy (r = 0.29, p < 0.001). Sensitivity for detecting malignancy (classifying as intermediate or elevated risk) was 94.9% (95% CI: 81.4–94.1), specificity was 24.4% (95% CI: 18.3–31.7), PPV was 22.2% (95% CI: 16.3–29.4), and NPV was 95.5% (95% CI: 83.3–99.2). Transpara had fair performance in detecting breast cancer with AUC 0.73 (95% CI: 0.63–0.82). Conclusions: Transpara AI is a useful screening mammography triage tool. Given its high sensitivity and high negative predictive value, AI may be used to guide radiologists in making biopsy or follow up recommendations. However, the high false-positive rate and presence of two false negatives underscore the need for radiologists to use caution and clinical expertise when interpreting AI results.

1. Introduction

Breast cancer is the most frequently diagnosed cancer in women with an incidence of 13.1% and leads to approximately 42,780 deaths annually in the United States [1]. The combination of early detection through mammographic screening and treatment advances has led to an estimated 44% mortality decrease since 1989 [1]. Digital breast tomosynthesis (DBT), a quasi-three-dimensional imaging technique approved by the FDA in 2011, is widely implemented in the United States with over 90% of Mammography Quality Standards Act (MQSA)-certified facilities reporting units in 2024. DBT functions by taking multiple low dose X-rays from different angles along an arc and reconstructing the images to create a 3D scrollable stack, resulting in superior visualization of overlapping breast tissues. DBT has been shown to improve sensitivity and specificity while reducing recall rate compared to 2D digital mammography, as demonstrated by numerous studies and a meta-analysis [2,3,4]. However, DBT’s principal drawback is increased radiologist workload as interpretation times are about twice as long as 2D digital mammography [5].
Artificial intelligence (AI), a rapidly evolving field using computer-based algorithms to perform tasks, is transforming breast imaging with applications in diagnosis, risk-stratification, breast density assessment, image quality control, and prediction of treatment response [6,7,8,9]. Significant progress has been made in leveraging deep learning-based AI algorithms to improve the accuracy and efficiency of screening mammography. Deep learning convolutional neural networks (CNNs) are used to segment and classify images based on trained models from large datasets [10]. Several retrospective studies and the Mammography Screening with Artificial Intelligence (MASAI) randomized control trial have demonstrated that radiologists had higher accuracy for detecting cancer on 2D digital mammography with AI support than with unaided reading [11,12,13,14]. Additionally, large retrospective studies have shown that AI strategies may cut reading time in half and dramatically reduce the workload of screening mammography [15,16,17].
Prior to the rise of AI, traditional machine learning computer-aided detection (CAD) systems were implemented since 1998, which marked an unlimited number of suspicious lesions on mammograms based on pre-programmed features. However, the established benefit of CAD has been challenged as several studies have shown that CAD does not improve accuracy and increases biopsy rates [18,19]. Moreover, CAD was shown to increase the time needed to interpret each study by approximately 20% due to a high rate of false positives [20]. Deep learning-based AI algorithms have been shown to outperform CAD with significantly better sensitivity and specificity when applied to a single mammography dataset [21]. Additional advantages of AI algorithms which classify data for a specific task include the potential to identify imaging features beyond the limits of human detection and to continuously improve [7,22].
With the overall goal of improving patient care, our institution replaced CAD in 2022 with a commercially available AI system for DBT, Transpara 1.7.1. A retrospective study in an enriched cohort with 25% cancer cases comparing two DBT AI systems (Transpara and ProFound AI) against radiologists found high AI performance in breast cancer detection but lower performance compared with radiologists [23]. However, a meta-analysis by Yoon et al. including four studies suggested that standalone AI had better sensitivity and specificity compared to radiologists in interpreting DBT exams, although there were insufficient studies to assess performance [24]. There is a paucity of studies evaluating the utilization of AI for DBT interpretation in a real-world setting versus in cancer-enriched datasets [25]. A retrospective real-world study by Letter et al. found nonsignificant improvement in cancer detection rate (CDR) with AI support, while a recent retrospective real-world study of 6407 DBT exams by Nepute et al. demonstrated a significant improvement in CDR [26,27]. It is crucial to understand the strengths and limitations of AI decision support for DBT in real clinical practice to guide radiologists in making biopsy or follow-up recommendations.
The objective of this study is to evaluate how Transpara AI decision support performed in detecting breast cancer on screening DBT mammography cases proceeding to stereotactic biopsy with histopathological results as the ground truth.

2. Methods

2.1. Study Population

This single-institution retrospective study was compliant with the Health Insurance Portability and Accountability Act. The institutional review board approved the study and waived the requirement of informed consent.
This study was performed at four urban centers within a single academic health system. We searched our Mammography Information System (MIS) (Ikonopedia, Richardson, TX, USA) for all stereotactic breast biopsies for adult patients from 1 October 2022 to 30 September 2023, resulting in 1051 biopsies.
Patients were then excluded for the following reasons: no preceding screening mammogram at the same institution within 14 months of biopsy (535 biopsies) or AI was not used for preceding mammogram (304 biopsies). All patients who underwent screening mammograms were asymptomatic and biopsies performed based on recommendations from diagnostic mammograms or ultrasound were excluded as the objective was to evaluate the performance of artificial intelligence on screening mammograms. Figure 1 shows the flow of patient selection, which resulted in 212 stereotactic biopsies with preceding screening mammogram for AI analysis.
Figure 1. Study flow chart.
The following information was extracted from our MIS: age, self-reported race and ethnicity, breast density on screening mammogram, finding type on mammography exam (classified as calcifications, mass, asymmetry, or architectural distortion), personal history of breast cancer, and family history of breast cancer in a first-degree relative. Race and ethnicity were evaluated because of their potential association with breast cancer screening. Race and ethnicity were categorized into the following broad categories: White, Black/African American, Asian, Hispanic, and unknown. Patients with Hispanic ethnicity were categorized as Hispanic regardless of race.

2.2. Image Acquisition

Patients underwent screening mammography using digital breast tomography (DBT) (Hologic, Marlborough, MA, USA). All breast imaging examinations were interpreted by dedicated breast radiologists based on the BI-RADS Atlas (5th edition).
This study used Transpara 1.7.1 (ScreenPoint Medical, Nijmegen, The Netherlands), a commercially available AI designed for automated breast cancer detection in screening mammography and breast tomosynthesis processed images. Transpara 1.7.1 does not evaluate prior examinations in risk score assessment. Transpara generates an interactive decision support score between 1 and 100 (score of 100 indicates the highest suspicion of malignancy) for each lesion and an overall case score which represents the highest score of all lesions. Transpara marked a maximum of two lesions per breast (most suspicious) for this dataset and provided a score category defined by the vendor as follows: low risk (≤42), intermediate risk (43–74), and elevated risk (≥75). The exact numerical value was not available for lesions classified as low risk (≤42) and thus the low-risk lesions were considered to have a score of zero. The Transpara system was trained and validated using a proprietary multicenter mammogram database and the mammograms in this study were not used to train the algorithms. All false-negative cases were reported to ScreenPoint Medical as part of quality assurance; no other support was received from ScreenPoint for this study.

2.3. Ground Truth

All patients underwent stereotactic biopsy and histopathology results served as the ground truth. The histologic types of breast cancer included ductal carcinoma in situ (DCIS), invasive ductal carcinoma (IDC), and invasive lobular carcinoma (ILC). There was one case of malignant lymphoma which was excluded from the analysis. Atypical lesions with high-risk features were noted and included atypical ductal hyperplasia (ADH), atypical lobular hyperplasia (ALH), lobular carcinoma in situ (LCIS), radial scar/complex sclerosing lesion, and papilloma with ADH. Atypical lesions and all other histopathology results were considered benign. Transpara lesion-specific scores were compared to pathology results on the biopsy sample (benign versus malignant). Two study authors reviewed all stereotactic biopsy mammogram images in conjunction with the preceding screening mammogram to obtain lesion-specific scores and ensure the location of the AI markings corresponded to the biopsied lesion. If multiple lesions from a single patient underwent stereotactic biopsy, each lesion was considered independent and lesion-specific score was used for evaluating AI performance.

2.4. Statistical Analysis

Data was summarized descriptively, and all analyses were performed using R (R Foundation for Statistical Computing, Vienna, Austria, version 4.4.1). Statistical significance was set at p < 0.05.
A 2 × 2 contingency table was created by calculating true-positive, false-positive, true-negative, and false-negative findings from the reported data. For the null hypothesis that there is no association between AI risk classification (low vs. intermediate/elevated risk) and biopsy outcome, p was less than 0.001 suggesting significant association. We calculated sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV).
The diagnostic performance of AI for detecting malignant lesions was assessed using receiver operating characteristic (ROC) analysis. AI performance was also stratified by lesion type (mass, calcification, asymmetry, and architectural distortion). Correlation between AI risk score and malignancy was estimated using the Pearson correlation coefficient.

3. Results

3.1. Patient Characteristics

The final study sample comprised 202 patients. Of these patients, eight patients underwent biopsy of two lesions (same breast or both breasts) and one patient underwent biopsy of three lesions (both breasts).
Table 1 summarizes patient characteristics. All patients were female. The mean age for the patients was 57.7 years (range, 35–89 years). A total of 24 out of 202 (12%) patients had a prior breast cancer diagnosis. A total of 50 out of 202 (25%) patients had a family history of breast cancer. Breast density was as follows: 5% fatty, 43% scattered, 46% heterogenous, and 6% extremely dense. Patient’s self-identified race and ethnicity classifications were as follows: 44% White, 17% Black/African American, 13% Asian, 20% Hispanic, and 6% unknown.
Table 1. Patient characteristics. Age is expressed as mean while all other data are expressed as count with percentage in parentheses.

3.2. Lesion Characteristics

The distribution of stereotactic biopsy lesions was as follows: 10 (5%) masses, 162 (76%) calcifications, 15 (7%) asymmetries, and 25 (12%) architectural distortions. Calcifications are typically occult on ultrasound (unless there is an associated mass) whereas masses, asymmetries, and architectural distortion are more likely to have an ultrasound correlate.
All lesions in this study were interpreted as high risk by the radiologist on screening mammogram (BIRADS 0, 3, 4, 5) as further recommendation was to undergo stereotactic biopsy. There were no low-risk lesions (BIRADS 1, 2) as these patients would have returned to screening rather than proceeding to stereotactic biopsy. The eight patients who received a BIRADS score of 3, 4, or 5 all presented for screening and the radiologist converted the screening mammogram to a diagnostic mammogram due to an abnormality.
Stereotactic biopsy results showed 136 benign, 36 high-risk lesions, and 40 malignant lesions. Table 2 summarizes the lesion characteristics including sub-types of the malignant findings.
Table 2. Characteristics of stereotactic biopsy lesions. Data are expressed as count with percentage in parentheses.

3.3. AI Performance

Of the 212 total lesions, AI classified 45 (21%) lesions as low risk, 117 (55%) lesions as intermediate risk, and 50 (24%) lesions as elevated risk. Of the 39 malignant lesions, AI classified 2 (5%) as low risk, 17 (44%) as intermediate risk, and 20 (51%) as elevated risk. Figure 2 shows the distribution of AI score for all lesions and for malignant lesions.
Figure 2. Distribution of artificial intelligence (AI) risk categories for all findings (n = 212) and malignant findings (n = 39).
Correlation with histopathological ground truth revealed 37 AI true positives, 130 AI false positives, 2 AI false negatives, and 42 AI true negatives. AI sensitivity for detecting malignancy (classifying as intermediate or elevated risk) was 94.9% (95% CI: 81.4–94.1) and specificity was 24.4% (95% CI: 18.3–31.7). PPV was 22.2% (95% CI: 16.3–29.4) for intermediate or elevated risk findings and low risk had a NPV of 95.5% (95% CI: 83.3–99.2) (Table 3). We did not find a statistically significant difference for sensitivity or specificity when stratifying patients by age, race and ethnicity, or breast density (p > 0.05).
Table 3. Performance of AI in relation to pathology results with contingency table showing true positives (TP), false negatives (FN), false positives (FP), and true negatives (TN).
Representative images of AI true-positive, false-positive, false-negative, and true-negative cases are shown in Figure 3, Figure 4, Figure 5 and Figure 6.
Figure 3. A 43-year-old asymptomatic female patient presents for screening. (A) Right craniocaudal and (B) mediolateral oblique views demonstrate an architectural distortion (red boxes) that was flagged by AI with a lesion-specific score of 77 corresponding to elevated risk. The case was recalled by the radiologist and biopsy revealed invasive ductal carcinoma, representing an AI true positive.
Figure 4. A 77-year-old asymptomatic female patient presents for screening. (A) Right craniocaudal and (B) magnification view demonstrate calcifications (red box) that were flagged by AI and assigned a lesion-specific score of 71 corresponding to intermediate risk. The case was recalled by the radiologist and biopsy revealed benign breast calcifications, representing an AI false positive.
Figure 5. Images from a 56-year-old female patient. (A) Mass with calcifications on diagnostic mammogram recalled by the radiologist and found to represent ductal carcinoma in situ on biopsy. (B) Right mediolateral view with no finding flagged by Transpara 1.7.1 and classified as low risk, representing an AI false negative. Case was sent to Screenpoint Medical as part of quality assurance. (C) Right mediolateral view with lesion (red box) flagged by newer version Transpara 2.1.0 and classified as intermediate risk with a lesion-specific score of 48.
Figure 6. Images from a 42-year-old female patient. (A) Right craniocaudal and (B) mediolateral views demonstrate a mass (red circles) which was recalled by the radiologist although no finding was flagged by AI which classified the case as low risk. (C) Mass persists on spot compression view and (D) ultrasound demonstrates an oval hypoechoic mass. Biopsy revealed benign breast tissue with a microcyst, representing an AI true negative.
AI score was positively correlated with rate of malignancy (r = 0.29, p < 0.001). Higher AI score categories corresponded to increased PPV (Table 4). Elevated risk (≥75) had a PPV of 40%.
Table 4. Positive predictive value (PPV) of AI by score category.
The receiver operating characteristic (ROC) curve for Transpara AI performance is shown in Figure 7. Transpara AI had fair performance in detecting breast cancer with area under the curve (AUC) 0.73 (95% CI: 0.63–0.82). AI performance stratified by lesion type demonstrated good performance for mass (AUC 0.84, 95% CI: 0.57–1), fair performance for calcifications (AUC 0.70, 95% CI: 0.58–0.82), low performance for asymmetry (AUC 0.69, 95% CI: 0.17–1), and good performance for architectural distortion (AUC 0.81, 95% CI: 0.64–0.98).
Figure 7. ROC curve for Transpara AI performance in detecting breast cancer with AUC 0.73 (95% CI: 0.63–0.82).

4. Discussion

This real-world retrospective study evaluated the standalone performance of Transpara AI in detecting breast cancer on screening digital breast tomosynthesis (DBT) exams which proceeded to stereotactic biopsy. Within this high-risk subset, Transpara AI had high sensitivity and high negative predictive value of approximately 95%. The false-positive rate was high with specificity around 25%. Higher AI scores corresponded to increased positive predictive value (PPV). Overall, AI had fair performance in detecting breast cancer with AUC 0.73.
Digital breast tomosynthesis is rapidly emerging as the standard of care in the United States and evaluating AI performance from a standpoint of diagnostic accuracy and workflow efficiency is of utmost importance. Our study provides insights on how to interpret AI risk score in clinical practice and triage screening mammograms. We found low AI risk score (<43) had a malignancy rate of 4.5%, whereas elevated risk (>75) had a malignancy rate of up to 40%. Therefore, we recommend prioritizing elevated-risk exams over low-risk exams to prevent delays in breast cancer diagnosis and treatment. Given the high negative predictive value, AI could potentially be used to decrease the number of unnecessary biopsies. For instance, radiologists could classify a group of punctate calcifications in the setting of low AI risk score as BIRADS-3 (probably benign) instead of BI-RADS 4 (suspicious) with high confidence to avoid biopsy. However, radiologists must exercise caution and use clinical judgment when interpreting AI results to avoid missing breast cancers as our study showed two cases of AI false negatives.
A recent retrospective study of AI-missed cancers on mammograms showed the most common reasons for misses were overlapping dense breast tissue, nonmammary zone locations, architectural distortions, and amorphous microcalcifications [28]. It is imperative for radiologists to be aware of AI shortcomings and pay special attention to certain areas and features. Additionally, AI is prone to generate false positives as evident in our study. Two retrospective studies of DBT screening exams have shown AI false positives are most commonly benign calcifications easily dismissed by radiologists [29,30]. Our dataset was skewed towards calcifications, representing 162 of 212 (76%) stereotactic-biopsy lesions, which may explain the high false-positive rate. Our institution favors ultrasound-guided over stereotactic biopsy due to patient comfort, so masses, asymmetries, and architectural distortions with an ultrasound correlate were less likely to be included. We found AI had fair performance for calcifications with AUC 0.70 (95% CI: 0.58–0.82), indicating scope for improvement with future algorithm development. The expertise of breast radiologists remains valuable and increased knowledge of the reasons for AI misses and false positives may improve patient care.
Our findings showing that AI classified 51% of malignant lesions as elevated risk are in line with a prior retrospective study of 1016 screen-detected cancers in which Transpara 1.7.0 classified 38% of screening mammograms preceding breast cancer diagnosis as high risk [31]. A recent population-based DBT screening cohort study with 4824 exams by Chen et al. using Transpara 1.7.1 reported a sensitivity of 89.2%, specificity of 68.5%, and NPV of 99.8% when defining positive results as intermediate/elevated risk [32]. While our study also showed that AI has a high NPV, we found significantly lower specificity. This may reflect our inherently high-risk subset of patients recommended for stereotactic biopsy as opposed to a general screening population. We also defined a positive AI result as intermediate/elevated risk so specificity could be improved if positive results were defined as elevated risk only.
There has been tremendous growth in the development of deep learning-based AI DBT systems over recent years. As of July 2024, a review article by Lamb et al. states there are at least six FDA-approved applications for screening DBT: Genius AI detection from Hologic Inc. (Marlborough, MA, USA), Lunit INSIGHT DBT from Lunit Inc. (New York, NY, USA), MammoScreen 2D/3D from Therapixel (Nice, France), ProFound AI Software from iCAD Inc. (Nashua, NH, USA), Saige-Dx from DeepHealth Inc. (Somerville, MA, USA), and Transpara from ScreenPoint Medical (Nijmegen, The Netherlands) [33]. While there is limited data comparing the performances of different AI-based systems for DBT, one study showed Profound AI 3.0 outperformed Transpara 1.7.0 (AUC 0.93 and 0.86, respectively; p = 0.004) [23]. The rapid deployment of AI systems and constant evolution with subsequent versions poses significant challenges for rigorous validation and reproducibility across institutions and time. We reported AI errors to ScreenPoint Medical for quality assurance and to support further performance improvement in subsequent versions. One false-negative case was identified by the newer version Transpara 2.1.0 which has now become commercially available. A study exploring AI errors found that recent versions of Transpara (1.6.0 and 1.7.0) had lower false-negative rates compared with an earlier version (1.4.0) [34]. There is significant heterogeneity between different AI-based systems for DBT as well as subsequent versions which may impact patient care and outcomes.
There is a gap between AI performance in controlled testing environments and real-life due to complex factors including patient demographics and individual radiologist experience [35]. There are few published studies evaluating AI-based systems for DBT integrated into clinical practice and our study adds to the literature. Important considerations for successful AI implementation include accessibility and equitable performance across diverse patient populations. A recent review article demonstrated that there are limited open access mammography datasets used for development of AI algorithms which report race and ethnicity [36]. If data is missing on certain underserved populations, this may introduce bias and exacerbate health disparities [37,38]. In our small sample of 202 patients, we did not find a statistically significant difference in sensitivity or specificity for AI when stratifying patients by age, race and ethnicity, or breast density. However, further large prospective studies are necessary to ensure AI performs equally well, independent of patient characteristics.
This study had limitations. First, this was an urban multicenter single-institution study focused on a high-risk subset of mammograms which proceeded to stereotactic biopsy, which may limit generalizability. Results may be different for other institutions, including those in other geographic areas with a different patient population. Second, this was a retrospective study limited by a small number of false-negative cancers. Finally, by study design, we focused on lesions prompting stereotactic biopsy, which were predominantly calcifications. The outcomes of artificial intelligence findings not recalled by the radiologist or prompting biopsy through a different modality such as ultrasound or MRI were not evaluated, and this should be explored in future studies.

5. Conclusions

DBT is widely adopted in the United States and our study establishes the current value of Transpara AI in detecting breast cancer on screening mammography cases preceding stereotactic biopsy. AI can improve clinical workflow efficiency as a triage tool and can guide radiologists in making follow-up or biopsy recommendations given the high sensitivity and negative predictive value. However, the expertise of radiologists remains valuable as AI is prone to generate false positives and may miss cancers detected by radiologists. Large prospective studies are necessary to validate the performance of AI in the real world and to ensure equitable care across diverse patient populations.

Author Contributions

Conceptualization, A.M., C.M. and L.R.M.; methodology, A.M., C.M. and L.R.M.; formal analysis, A.M.; data curation, A.M., C.M., A.S., N.T. and S.S.; writing—original draft preparation, A.M.; writing—review and editing, A.M. and L.R.M.; supervision, D.S.M. and L.R.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with Mount Sinai’s Federal Wide Assurances (FWA#00005656, FWA#00005651) and approved by the Department of Health and Human Services and Institutional Review Board of the Icahn School of Medicine at Mount Sinai on 10 January 2024.

Data Availability Statement

The original contributions presented in the study are included in the article. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

A.M., C.M., A.S., N.T., S.S. and D.S.M. have no conflicts of interest to declare. L.R.M. provides consulting services for Screenpoint Medical, Sectra, Siemens Healthineers, and GE Healthcare.

Abbreviations

AI = artificial intelligence, AUC = area under the ROC curve, CAD = computer-aided detection, DBT = digital breast tomosynthesis, NPV = negative predictive value, PPV = positive predictive value, ROC = receiver operating curve

References

  1. Breast Cancer Facts & Figures 2024–2025. Available online: https://www.cancer.org/content/dam/cancer-org/research/cancer-facts-and-statistics/breast-cancer-facts-and-figures/2024/breast-cancer-facts-and-figures-2024.pdf (accessed on 1 November 2025).
  2. Rafferty, E.A.; Park, J.M.; Philpotts, L.E.; Poplack, S.P.; Sumkin, J.H.; Halpern, E.F.; Niklason, L.T. Assessing Radiologist Performance Using Combined Digital Mammography and Breast Tomosynthesis Compared with Digital Mammography Alone: Results of a Multicenter, Multireader Trial. Radiology 2013, 266, 104–113. [Google Scholar] [CrossRef]
  3. Marinovich, M.L.; Hunter, K.E.; Macaskill, P.; Houssami, N. Breast Cancer Screening Using Tomosynthesis or Mammography: A Meta-analysis of Cancer Detection and Recall. J. Natl. Cancer Inst. 2018, 110, 942–949. [Google Scholar] [CrossRef]
  4. Durand, M.A.; Friedewald, S.M.; Plecha, D.M.; Copit, D.S.; Barke, L.D.; Rose, S.L.; Hayes, M.K.; Greer, L.N.; Dabbous, F.M.; Conant, E.F. False-Negative Rates of Breast Cancer Screening with and Without Digital Breast Tomosynthesis. Radiology 2021, 298, 296–305. [Google Scholar] [CrossRef]
  5. Dang, P.A.; Freer, P.E.; Humphrey, K.L.; Halpern, E.F.; Rafferty, E.A. Addition of Tomosynthesis to Conventional Digital Mammography: Effect on Image Interpretation Time of Screening Examinations. Radiology 2014, 270, 49–56. [Google Scholar] [CrossRef]
  6. Shamir, S.B.; Sasson, A.L.; Margolies, L.R.; Mendelson, D.S. New Frontiers in Breast Cancer Imaging: The Rise of AI. Bioengineering 2024, 11, 451. [Google Scholar] [CrossRef] [PubMed]
  7. Bahl, M. Artificial Intelligence: A Primer for Breast Imaging Radiologists. J. Breast Imaging 2020, 2, 304–314. [Google Scholar] [CrossRef]
  8. Ali, A.; Alghamdi, M.; Marzuki, S.S.; Tengku Din, T.; Yamin, M.S.; Alrashidi, M.; Alkhazi, I.S.; Ahmed, N. Exploring AI Approaches for Breast Cancer Detection and Diagnosis: A Review Article. Breast Cancer 2025, 17, 927–947. [Google Scholar] [CrossRef]
  9. Sechopoulos, I.; Teuwen, J.; Mann, R. Artificial intelligence for breast cancer detection in mammography and digital breast tomosynthesis: State of the art. Semin. Cancer Biol. 2021, 72, 214–225. [Google Scholar] [CrossRef] [PubMed]
  10. Diaz, O.; Rodriguez-Ruiz, A.; Sechopoulos, I. Artificial Intelligence for breast cancer detection: Technology, challenges, and prospects. Eur. J. Radiol. 2024, 175, 111457. [Google Scholar] [CrossRef] [PubMed]
  11. Rodriguez-Ruiz, A.; Lang, K.; Gubern-Merida, A.; Broeders, M.; Gennaro, G.; Clauser, P.; Helbich, T.H.; Chevalier, M.; Tan, T.; Mertelmeier, T.; et al. Stand-Alone Artificial Intelligence for Breast Cancer Detection in Mammography: Comparison with 101 Radiologists. J. Natl. Cancer Inst. 2019, 111, 916–922. [Google Scholar] [CrossRef]
  12. Lang, K.; Josefsson, V.; Larsson, A.M.; Larsson, S.; Hogberg, C.; Sartor, H.; Hofvind, S.; Andersson, I.; Rosso, A. Artificial intelligence-supported screen reading versus standard double reading in the Mammography Screening with Artificial Intelligence trial (MASAI): A clinical safety analysis of a randomised, controlled, non-inferiority, single-blinded, screening accuracy study. Lancet Oncol. 2023, 24, 936–944. [Google Scholar] [CrossRef]
  13. Pacilè, S.; Lopez, J.; Chone, P.; Bertinotti, T.; Grouin, J.M.; Fillard, P. Improving Breast Cancer Detection Accuracy of Mammography with the Concurrent Use of an Artificial Intelligence Tool. Radiol. Artif. Intell. 2020, 2, e190208. [Google Scholar] [CrossRef] [PubMed]
  14. Hernstrom, V.; Josefsson, V.; Sartor, H.; Schmidt, D.; Larsson, A.M.; Hofvind, S.; Andersson, I.; Rosso, A.; Hagberg, O.; Lang, K. Screening performance and characteristics of breast cancer detected in the Mammography Screening with Artificial Intelligence trial (MASAI): A randomised, controlled, parallel-group, non-inferiority, single-blinded, screening accuracy study. Lancet Digit. Health 2025, 7, e175–e183. [Google Scholar] [CrossRef]
  15. Conant, E.F.; Toledano, A.Y.; Periaswamy, S.; Fotin, S.V.; Go, J.; Boatsman, J.E.; Hoffmeister, J.W. Improving Accuracy and Efficiency with Concurrent Use of Artificial Intelligence for Digital Breast Tomosynthesis. Radiol. Artif. Intell. 2019, 1, e180096. [Google Scholar] [CrossRef]
  16. Raya-Povedano, J.L.; Romero-Martin, S.; Elias-Cabot, E.; Gubern-Merida, A.; Rodriguez-Ruiz, A.; Alvarez-Benito, M. AI-based Strategies to Reduce Workload in Breast Cancer Screening with Mammography and Tomosynthesis: A Retrospective Evaluation. Radiology 2021, 300, 57–65. [Google Scholar] [CrossRef]
  17. Lauritzen, A.D.; Rodriguez-Ruiz, A.; von Euler-Chelpin, M.C.; Lynge, E.; Vejborg, I.; Nielsen, M.; Karssemeijer, N.; Lillholm, M. An Artificial Intelligence-based Mammography Screening Protocol for Breast Cancer: Outcome and Radiologist Workload. Radiology 2022, 304, 41–49. [Google Scholar] [CrossRef] [PubMed]
  18. Lehman, C.D.; Wellman, R.D.; Buist, D.S.; Kerlikowske, K.; Tosteson, A.N.; Miglioretti, D.L.; Breast Cancer Surveillance Consortium. Diagnostic Accuracy of Digital Screening Mammography with and Without Computer-Aided Detection. JAMA Intern. Med. 2015, 175, 1828–1837. [Google Scholar] [CrossRef] [PubMed]
  19. Fenton, J.J.; Taplin, S.H.; Carney, P.A.; Abraham, L.; Sickles, E.A.; D’Orsi, C.; Berns, E.A.; Cutter, G.; Hendrick, R.E.; Barlow, W.E.; et al. Influence of computer-aided detection on performance of screening mammography. N. Engl. J. Med. 2007, 356, 1399–1409. [Google Scholar] [CrossRef]
  20. Tchou, P.M.; Haygood, T.M.; Atkinson, E.N.; Stephens, T.W.; Davis, P.L.; Arribas, E.M.; Geiser, W.R.; Whitman, G.J. Interpretation Time of Computer-aided Detection at Screening Mammography. Radiology 2010, 257, 40–46. [Google Scholar] [CrossRef]
  21. Bahl, M.; Kshirsagar, A.; Pohlman, S.; Lehman, C.D. Traditional versus modern approaches to screening mammography: A comparison of computer-assisted detection for synthetic 2D mammography versus an artificial intelligence algorithm for digital breast tomosynthesis. Breast Cancer Res. Treat. 2025, 210, 529–537. [Google Scholar] [CrossRef]
  22. Lamb, L.R.; Lehman, C.D.; Gastounioti, A.; Conant, E.F.; Bahl, M. Artificial Intelligence (AI) for Screening Mammography, from the AJR Special Series on AI Applications. Am. J. Roentgenol. 2022, 219, 369–380. [Google Scholar] [CrossRef]
  23. Resch, D.; Lo Gullo, R.; Teuwen, J.; Semturs, F.; Hummel, J.; Resch, A.; Pinker, K. AI-enhanced Mammography with Digital Breast Tomosynthesis for Breast Cancer Detection: Clinical Value and Comparison with Human Performance. Radiol. Imaging Cancer 2024, 6, e230149. [Google Scholar] [CrossRef]
  24. Yoon, J.H.; Strand, F.; Baltzer, P.A.T.; Conant, E.F.; Gilbert, F.J.; Lehman, C.D.; Morris, E.A.; Mullen, L.A.; Nishikawa, R.M.; Sharma, N.; et al. Standalone AI for Breast Cancer Detection at Screening Digital Mammography and Digital Breast Tomosynthesis: A Systematic Review and Meta-Analysis. Radiology 2023, 307, e222639. [Google Scholar] [CrossRef] [PubMed]
  25. Park, E.K.; Kwak, S.; Lee, W.; Choi, J.S.; Kooi, T.; Kim, E.K. Impact of AI for Digital Breast Tomosynthesis on Breast Cancer Detection and Interpretation Time. Radiol. Artif. Intell. 2024, 6, e230318. [Google Scholar] [CrossRef]
  26. Letter, H.; Peratikos, M.; Toledano, A.; Hoffmeister, J.; Nishikawa, R.; Conant, E.; Shisler, J.; Maimone, S.; Diaz de Villegas, H. Use of Artificial Intelligence for Digital Breast Tomosynthesis Screening: A Preliminary Real-world Experience. J. Breast Imaging 2023, 5, 258–266. [Google Scholar] [CrossRef]
  27. Nepute, J.A.; Peratikos, M.; Toledano, A.Y.; Salvas, J.P.; Delks, H.; Shisler, J.L.; Hoffmeister, J.W.; Madden, C.M. Improved Breast Cancer Detection with Artificial Intelligence in a Real-World Digital Breast Tomosynthesis Screening Program. Clin. Breast Cancer 2025, 25, 808–816.e5. [Google Scholar] [CrossRef]
  28. Woo, O.H.; Song, S.E.; Choe, S.J.; Kim, M.; Cho, K.R.; Seo, B.K. Invasive Breast Cancers Missed by AI Screening of Mammograms. Radiology 2025, 315, e242408. [Google Scholar] [CrossRef]
  29. Koch, H.W.; Bergan, M.B.; Gjesvik, J.; Larsen, M.; Bartsch, H.; Haldorsen, I.H.S.; Hofvind, S. Mammographic features in screening mammograms with high AI scores but a true-negative screening result. Acta Radiol. 2025, 66, 1225–1232. [Google Scholar] [CrossRef] [PubMed]
  30. Shahrvini, T.; Wood, E.J.; Joines, M.M.; Nguyen, H.; Hoyt, A.C.; Chalfant, J.S.; Capiro, N.M.; Fischer, C.P.; Sayre, J.; Hsu, W.; et al. Artificial Intelligence Versus Radiologist False Positives on Digital Breast Tomosynthesis Examinations in a Population-Based Screening Program. Am. J. Roentgenol. 2025, in press. [Google Scholar] [CrossRef] [PubMed]
  31. Larsen, M.; Olstad, C.F.; Koch, H.W.; Martiniussen, M.A.; Hoff, S.R.; Lund-Hanssen, H.; Solli, H.S.; Mikalsen, K.O.; Auensen, S.; Nygard, J.; et al. AI Risk Score on Screening Mammograms Preceding Breast Cancer Diagnosis. Radiology 2023, 309, e230989. [Google Scholar] [CrossRef]
  32. Chen, I.E.; Joines, M.; Capiro, N.; Dawar, R.; Sears, C.; Sayre, J.; Chalfant, J.; Fischer, C.; Hoyt, A.C.; Hsu, W.; et al. Commercial Artificial Intelligence Versus Radiologists: NPV and Recall Rate in Large Population-Based Digital Mammography and Tomosynthesis Screening Mammography Cohorts. Am. J. Roentgenol. 2025, in press. [Google Scholar] [CrossRef] [PubMed]
  33. Lamb, L.R.; Lehman, C.D.; Do, S.; Kim, K.; Langarica, S.; Bahl, M. Artificial Intelligence (AI)-Based Computer-Assisted Detection and Diagnosis for Mammography: An Evidence-Based Review of Food and Drug Administration (FDA)-Cleared Tools for Screening Digital Breast Tomosynthesis (DBT). AI Precis. Oncol. 2024, 1, 195–206. [Google Scholar] [CrossRef] [PubMed]
  34. Zeng, A.; Houssami, N.; Noguchi, N.; Nickel, B.; Marinovich, M.L. Frequency and characteristics of errors by artificial intelligence (AI) in reading screening mammography: A systematic review. Breast Cancer Res. Treat. 2024, 207, 1–13. [Google Scholar] [CrossRef] [PubMed]
  35. Sosna, J.; Joskowicz, L.; Saban, M. Navigating the AI Landscape in Medical Imaging: A Critical Analysis of Technologies, Implementation, and Implications. Radiology 2025, 315, e240982. [Google Scholar] [CrossRef]
  36. Laws, E.; Palmer, J.; Alderman, J.; Sharma, O.; Ngai, V.; Salisbury, T.; Hussain, G.; Ahmed, S.; Sachdeva, G.; Vadera, S.; et al. Diversity, inclusivity and traceability of mammography datasets used in development of Artificial Intelligence technologies: A systematic review. Clin. Imaging 2024, 118, 110369. [Google Scholar] [CrossRef]
  37. Ozcan, B.B.; Patel, B.K.; Banerjee, I.; Dogan, B.E. Artificial Intelligence in Breast Imaging: Challenges of Integration into Clinical Practice. J. Breast Imaging 2023, 5, 248–257. [Google Scholar] [CrossRef]
  38. Nguyen, D.L.; Ren, Y.; Jones, T.M.; Thomas, S.M.; Lo, J.Y.; Grimm, L.J. Patient Characteristics Impact Performance of AI Algorithm in Interpreting Negative Screening Digital Breast Tomosynthesis Studies. Radiology 2024, 311, e232286. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.