Next Article in Journal
BadDomains: Early Detection of Phishing Domains Registration
Previous Article in Journal
A Post-Quantum Secure RFID Authentication Protocol Based on NTRU Encryption Algorithm
Previous Article in Special Issue
CHARMS: A CNN-Transformer Hybrid with Attention Regularization for MRI Super-Resolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Image Quality Standardization in Radiomics: A Systematic Review of Artifacts, Variability, and Feature Stability

by
Francesco Felicetti
1,
Francesco Lamonaca
1,2,
Domenico Luca Carnì
1 and
Sandra Costanzo
1,3,4,*
1
Department of Computer Engineering, Modeling, Electronics and Systems Engineering, University of Calabria, Ponte P. Bucci, 87036 Rende, Italy
2
Institute of Nanotechnology (CNR-NANOTEC), National Research Council of Italy, 87036 Rende, Italy
3
Consorzio Nazionale Interuniversitario per le Telecomunicazioni, 43124 Parma, Italy
4
Inter-University National Research Center on Interactions Between Electromagnetic Fields and Biosystems (ICEmB), 16126 Genova, Italy
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(3), 1039; https://doi.org/10.3390/s26031039
Submission received: 27 November 2025 / Revised: 28 January 2026 / Accepted: 31 January 2026 / Published: 5 February 2026
(This article belongs to the Special Issue Sensing Technologies in Digital Radiology and Image Analysis)

Abstract

This paper explores the role of metrology in the assessment of image quality in the field of radiomics. Image Quality Assessment (IQA) is central to ensuring the reliability and reproducibility of radiomic analyses, as it directly affects the accuracy of feature extraction and segmentation, ultimately impacting diagnostic outcomes. From the analysis of approximately 20,000 papers sourced from three databases (PubMed, Scopus, IEEE Xplore), last searched in December 2025, the need for standardized imaging protocols and quality control measures emerges as a critical theme. Studies were included if they involved radiomic feature extraction and evaluated the impact of image quality variations on feature robustness and no formal risk-of-bias assessment was performed. A total of 105 studies were included, covering different medical imaging modalities. Across the included studies, noise, motion, acquisition and reconstruction parameters, and other artifacts consistently emerged as major sources of radiomic feature instability. Indeed, in most papers, IQA is neglected, while the effect of poor-quality images is reported. This research identifies and discusses the relevant issues reported in clinical practice, as well as the main metrics adopted for image quality evaluation. Through a comprehensive review of current literature and an analysis of emerging trends, this paper highlights the urgent need for innovative solutions in image quality metrics tailored to radiomics applications.

1. Introduction

Radiomics is revolutionizing personalized medicine and oncology through its potential to enhance diagnosis, predict prognosis, and improve therapeutic responses from non-invasive imaging [1,2,3,4,5]. The core of this discipline relies on high-quality imaging data, as the accuracy of radiomic analysis is strictly related to the quality of the input images [6]. IQA is thus essential in ensuring the reliability and reproducibility of radiomics applications. Image Quality Assessment (IQA) covers the evaluation of various aspects of the image, such as resolution, contrast, noise, and artifacts, which can significantly impact the reliability of feature extraction and subsequent analysis [7,8,9,10,11]. First, it acts as a quality control filter, accurately screening images and excluding those that fail to meet minimum quality benchmarks. Second, it ensures that the data entering the radiomics pipeline are of sufficient quality to yield meaningful insights [12]. This is a crucial step because the robustness of radiomic features is highly dependent on the consistency and quality of the imaging data [13,14]. Variations in imaging protocols, scanner types, and even patient positioning can introduce biases, thus affecting the reproducibility of the results [15,16,17]. Moreover, advanced IQA methods can play a major role in standardizing images across different institutions and scanner types, facilitating multicentric studies and the generalization of radiomics models [18]. The introduction of automated tools for image assessment is becoming increasingly urgent because an efficient IQA tool could rapidly evaluate vast datasets [19,20], identifying, measuring, and, when possible, correcting image quality issues. This would support manual operators by reducing their workload and eliminating elements related to the subjectivity of a visual inspection [21]. In summary, the importance of IQA in radiomics applications cannot be overstated. It is an essential prerequisite for the accurate extraction and analysis of radiomic features, directly influencing the predictive power and clinical utility of radiomics models. As the field advances, ongoing developments in IQA methodologies will be critical to harnessing the full potential of radiomics in personalized medicine and beyond. When addressing image quality challenges in radiomics, it is important to distinguish between two complementary approaches: image quality standardization and image quality correction. Standardization refers to protocol-level control measures implemented at the acquisition stage to ensure consistent imaging parameters across scanners, institutions, and time points [18,22,23]. This includes harmonized acquisition protocols, calibration procedures, and quality control measures designed to prevent quality degradation before it occurs. In contrast, correction refers to algorithmic, post-acquisition interventions applied to mitigate degradations already present in acquired images, such as denoising algorithms [24,25,26], artifact reduction techniques [26,27], and harmonization methods [28,29,30]. While both approaches are essential and complementary, standardization should be prioritized, as it reduces variability at the source, whereas post hoc corrections may introduce additional uncertainty or alter radiomic feature distributions unpredictably. This paper aims to critically examine the role of IQA in radiomics by analyzing its presence and consideration in the existing scientific literature, trying to identify the IQA impact on radiomic feature extraction and diagnostic reliability.
While previous systematic reviews have addressed radiomic feature reproducibility or harmonization strategies, this review systematically quantifies the prevalence of IQA consideration in radiomics literature from a metrological perspective. Unlike existing reviews [14,18,28,29,30], primarily focusing on clinical or algorithmic aspects, our work introduces the following new aspects: (1) a quantitative analysis of the IQA gap in radiomics research, (2) a first attempt at systematic categorization of artifacts linked to affected feature classes, (3) a critical comparison of correction approaches, (4) practical recommendations for integrating IQA as a quality-control step in radiomics pipelines.

2. Materials and Methods

This systematic review was conducted following the PRISMA 2020 statement [31], with the primary objective of mapping and analyzing how image quality affects radiomic feature extraction and the stability of downstream predictive models. The focus of this review is methodological rather than clinical, aiming to provide a comprehensive overview of experimental designs, quality metrics, and correction strategies employed in radiomics research to mitigate the impact of image degradation. The literature search was carried out in three major scientific databases (PubMed, Scopus, and IEEE Xplore) to comprehensively capture biomedical, engineering, and computational imaging studies relevant to radiomics and IQA. Searches were conducted in each database up to December 2025, with no language or publication-type restrictions. An initial broad exploratory query using the single keyword “Radiomics” retrieved approximately 20,000 records across sources. To focus on works explicitly addressing IQ-related degradations in the context of radiomics, we then applied a refined query that combined Radiomics with IQA terms (noise, blur, artifact, image perturbation, image quality metrics). The exact database-specific queries were:
  • Scopus:
    TITLE-ABS-KEY (radiomics) AND (TITLE-ABS-KEY (noise)
    OR TITLE-ABS-KEY (artifact)
    OR TITLE-ABS-KEY (blur) OR TITLE-ABS-KEY (“image perturbation”)
    OR TITLE-ABS-KEY (“image quality metrics”))
  • PubMed:
    (Radiomics) AND ((Noise) OR (Blur) OR (artifact)
    OR (“image perturbation”) OR (“image quality metrics”))
  • IEEE Xplore:
    ((“Document Title”:“radiomics”) OR (“Abstract”:“radiomics”))
    AND (“noise” OR “blur” OR “artifact” OR “image perturbation”
    OR “image quality metrics”)
All search results were exported and merged into a unified dataset. Duplicates were identified and removed using semi-automated filtering followed by manual verification. The combined corpus served as the foundation for the subsequent screening and eligibility assessment.
Inclusion and exclusion criteria were defined a priori, according to the PICO framework adapted for methodological systematic reviews:
  • Population: Imaging studies (CT, MRI, PET, or hybrid modalities) employing radiomic feature extraction.
  • Intervention/Exposure: Image quality variations or artifacts, whether simulated, measured, or induced experimentally (e.g., noise, blur, motion, beam hardening, or acquisition parameters).
  • Comparator: Reference or baseline imaging conditions, or different levels of degradation.
  • Outcome: Quantitative assessment of feature variability, robustness, or reproducibility in relation to IQA metrics.
Studies were included if they met all the above conditions. Works that were reviews, editorials, or methodological protocols without original data, as well as those focusing exclusively on image quality without radiomic analysis, were excluded. A total of 21,766 records were retrieved, and 1866 duplicates were removed, leaving 19,990 unique records for title and abstract screening. After preliminary filtering, 187 reports were sought for full-text assessment, of which 26 could not be retrieved. The remaining 161 reports were evaluated for eligibility, and 23 were excluded for specific reasons: reviews or protocols (n = 10), radiomics without quantitative IQA metrics (n = 7), IQA without radiomics (n = 2), insufficient data (n = 3), and duplicate datasets (n = 1). At this stage, other 33 studied were were excluded for the following reasons: Extended abstract (n = 28), Editorial (n = 5). Consequently, 105 studies were included in the final qualitative synthesis. The overall study selection process is summarized in Figure 1, according to the PRISMA 2020 flow diagram.

3. Methodological and Scope Limitations

Since this systematic review was conducted with specific methodological adaptations justified by its exploratory and metrological scope, it was not prospectively registered (e.g., in PROSPERO [32]), and no formal risk-of-bias assessment, reporting-bias assessment, or certainty-of-evidence evaluation (e.g., QUADAS-2, GRADE [33,34]) was performed. These procedures are designed for clinical intervention or diagnostic accuracy studies, so they are not applicable to a methodological review that does not focus on quantitative estimations or clinical outcomes. However, it uniquely addresses image quality considerations in radiomics research. No inferential statistical analyses or meta-analytic pooling were performed due to substantial heterogeneity in imaging modalities (CT, MRI, PET, ultrasound), outcome metrics (ICC, SNR, CNR, Dice coefficients), and experimental designs (phantom vs. clinical studies). Data were instead qualitatively analyzed by emphasizing trends, methodological gaps, and reproducibility issues identified across the retrieved literature.
It is important to further clarify that the specific choice of the search queries adopted in the present review may have introduced selection bias. As stated in the Introduction, the adopted strategy was intentionally designed to focus on studies explicitly addressing image quality degradation and artifacts in radiomics through specific terminology (noise, artifact, blur, image perturbation, image quality metrics). Consequently, works indirectly dealing with image quality, such as studies primarily focused on preprocessing, harmonization strategies, robustness analyses, or downstream pipeline optimization, may not have been captured if image quality issues were not explicitly mentioned in the title, abstract, or keywords.
Furthermore, the bibliometric analysis was based on three major scientific databases (PubMed, Scopus, IEEE Xplore), providing comprehensive coverage of biomedical and engineering literature. Although this choice ensures broad coverage of peer-reviewed biomedical and engineering research, it may have excluded some studies, potentially limiting the full completeness of the bibliometric analysis.

4. Radiomics and Image Quality

Searching for the word “radiomics” across various scientific databases yields a considerable number of results. Specifically, a search for “radiomics” in titles, abstracts, and keywords on Scopus, PubMed, and IEEE Xplore databases finds just under 20,000 unique results. The topic is of clear scientific interest, as evidenced by the increasing number of studies on the topic (Figure 2a). The analysis of the most frequent words in this research field reveals “radiomics” as the main topic, but other prominent words include “feature” or “features” and “image/images/imaging” (Figure 2b). The term “radiomics” was introduced by Lambin et al. in 2012 [35], marking its first appearance in the scientific literature to describe the extraction of quantitative features from radiological images for cancer studies using computed tomography (CT). This introduction discusses the use of the term “radiomics” in the context of medical imaging and analysis, focusing on the extraction of numerous quantitative features from medical images for various clinical applications, including but not limited to cancer research. Although the attribution of this term to the mentioned study is clear, it is also necessary to state that previous studies exist that discuss the extraction of information from medical images [7,8,9].
From the analysis of the most frequent words and the definition of radiomics, it becomes clear that images play a central role in the radiomic workflow. Images can indeed be considered the fundamental link connecting data acquisition to the extraction of information, which allows qualitative or quantitative evaluations and provides the basis for constructing prognostic models. In a more detailed examination of the same databases previously considered, by performing refined queries to include not just “Radiomics” but also one of the following terms: “noise”, “artifact”, “blur”, “image perturbation,” or “image quality metrics” (each highlighting the crucial role of image quality), the initial set of nearly 20,000 results reduced to less than 750 works. This reduction to 3.75% is clearly in contrast with expectations, especially considering the absence of an established and acknowledged effective metric for image evaluation. The question that soon arises is whether it is appropriate to introduce an analysis of image quality within the radiomic workflow. The location and selection activity was carried out with the help of the MySLR digital platform [36].

5. Image Quality Assessment Tools: A Need

Altered images in radiology can significantly impact patient care, leading to diagnostic inaccuracies that may result in misdiagnosis or delayed treatments. This not only increases patient’s anxiety and dissatisfaction but also increases the risk of legal consequences. According to a 2018 study [37] conducted on a sample of 1,210,858 radiology exams of various types (Figure 3), concerns about image quality were reported in 2.4% of cases during clinical evaluation. Although this percentage may seem almost negligible, it is crucial to consider several aspects potentially leading to a significantly underestimated evaluation, as compared to the potentially real one, namely:
  • In 2.4% of cases, operators who assessed the images asserted the importance of highlighting defects potentially leading to data misinterpretation. However, it is impossible to determine the number of examiners who did not identify potential defects.
  • Reporting of defects may have been omitted by experienced operators relying on their personal expertise.
  • Reporting of defects may have been omitted by less experienced operators who were unable to recognize the problem.
  • Although 2.4% may appear to be a small percentage, it is important to emphasize that this percentage, within a sample of over 1,200,000 cases, translates to approximately 30,000 potential misdiagnoses, with all the associated consequences.
  • If only Positron Emission Tomography (PET), CT, and Magnetic Resonance Imaging (MRI) are considered, the percentage of defects potentially leading to data misinterpretation highlighted by the operators increases up to 6%. This means that these last techniques, typically adopted for automatic cancer detection, are more prone to being affected by alterations during acquisition.
Anyway, radiologists typically attempt to analyze images and provide a diagnosis, even in cases of low-quality images, frequently documenting the details concerning poor quality directly in the report [37].
The prevalence of concerns about image quality (Figure 3c), as reported in clinical evaluations, underscores the critical need for robust assessment methods to ensure the reliability of radiomic data. In cases where artifacts in images are not identified, failure to address the above issues can lead to misinterpretation of radiomic features and subsequent misdiagnosis or delayed treatments [38]. Furthermore, the inherent subjectivity in detecting image artifacts highlights the importance of objective assessment tools to standardize quality evaluation across different radiology practices and expertise levels. In addition to technical aspects, there is also an economic need. Even if investing resources in high-quality imaging technologies for radiomics involves upfront costs related to purchasing equipment, training personnel, and potentially increased operational costs, these investments can lead to significant economic benefits in the long term. By improving diagnostic accuracy and efficiency, high-quality imaging can reduce unnecessary procedures, misdiagnoses, and treatment complications. It can also enable personalized treatment plans that are more cost-effective and have better outcomes, potentially leading to overall healthcare cost savings.

6. Main Concerns in Clinical Practice

An artifact in imaging refers to any distortion or anomaly present in medical images, leading to non-accurate representation of the underlying anatomy or pathology of the scanned tissue or organ [39]. These discrepancies can emerge from different sources, including the imaging process itself, and other patient-related factors. Artifacts may reveal unexpected shadows, streaks, blurs, or other forms of distortion, that potentially could lead to misleading diagnoses or treatment planning. Table 1 summarizes the most common types of imaging artifacts, categorizing them by imaging modality. Their presence is a significant concern in radiology and medical imaging analysis, with worrying rates in some settings, such as the reported 38% of cases in [40].

6.1. Motion

In medical imaging, particularly with PET/CT and MRI modalities, patient motion is the main source of artifacts that significantly degrade image quality (Figure 4) [41,42]. The long acquisition times required for these modalities amplify the issue, with MRI being notably sensitive due to its complex nature and the extended duration needed to capture high-quality images [43]. This problem is not confined to a specific patient demographic; however, it is pronounced in elderly patients with movement disorders [41] and highly kinetic participants such as those in developing and aging populations. As reported in [43], various techniques have been proposed to address motion artifacts, including mechanical restraints, synchronized sensors, and advanced computational methods, such as machine learning and image processing. Additionally, the advent of simultaneous PET-MR imaging offers new ways for motion correction, exploiting MR-based techniques for real-time monitoring and correction during PET data acquisition [44]. Despite these advancements, motion correction in medical imaging remains a challenging field, with no single method proving to be effective in all situations, underscoring the need for continuous research and development in this area [45,46].

6.2. Beam Hardening Streaks and Cupping Artifacts

Beam hardening artifacts, specifically in CT scans, are substantial due to the polychromatic nature of X-ray beams. As these beams penetrate the body, lower-energy photons are absorbed more readily than higher-energy photons, causing the beam to harden as it passes through the tissue [48]. This hardening alters the expected attenuation profile, producing streak and cupping artifacts, which can lead to significant image distortions and can affect clinical diagnoses and quantitative radiomic analyses [27]. Streaking artifacts appear as lines or streaks in the CT image, and they are especially noticeable around high-density objects within the scan, such as metal implants (Figure 5a [49]). The cupping effect is observed as an abnormal, cup-shaped distortion in the homogeneity of an image, particularly in the case of homogeneous substances such as soft tissues (Figure 5b [50]). Various approaches exist to mitigate the above artifacts. Traditionally, hardware-based solutions such as iterative reconstruction techniques and dual-energy CT have been employed [51,52]. These methods adjust the CT data acquisition process, or they use computational techniques to compensate for the artifacts; however, they often involve significant computational resources or additional hardware, which may not always be available in clinical settings [53]. Alternatively, recent advancements have introduced software-based solutions using artificial intelligence (AI) to address the problem of beam hardening. These AI models, such as the one proposed by Kalare et al. [27], take advantage of large datasets of CT images to learn the characteristics of beam-hardening artifacts and apply corrections during the image reconstruction process. The AI-driven approach is designed to improve the accuracy of the reconstructed images, thereby enhancing the quality of radiomic data for better clinical and research outcomes [54].

6.3. Noise

Noise in radiomics refers to random variations in the gray level of image data pixels, which are not related to the true signal (Figure 6). Noise in radiomic data can originate at different stages of the imaging pipeline, each characterized by distinct physical or algorithmic mechanisms, and by a specific impact on the stability of quantitative features. In particular, it is possible to distinguish between acquisition noise, reconstruction noise, and post-processing noise. Acquisition noise is intrinsic to the image formation process, and it is primarily determined by the physical limitations of the detection system. This type of noise shows a direct dependence on parameters such as radiation dose, acquisition time, or signal-to-noise ratio, and it tends to manifest as an increase in voxel-wise variability [55]. The noise introduced during the reconstruction phase, in addition to representing the propagation of acquisition noise, is conditioned by algorithmic choices. Iterative reconstruction methods, as well as the use of different reconstruction kernels or smoothing levels, can modify the spatial distribution of noise, introducing medium-range correlations or structured artifacts [56]. The reproducibility of radiomic features in noisy images is commonly assessed using techniques such as test-retest studies, phantom studies, and statistical stability analysis. These methods evaluate whether features extracted from the same region of interest remain consistent under different conditions or across multiple imaging sessions [12]. Studies have identified several radiomic features that are robust to noise, including shape-based features and some textural features, such as entropy and energy. These features tend to exhibit less variation under different noise levels compared to more complex textural features, such as local binary patterns or high-order statistical features [57]. Generative models, particularly those based on deep learning, show promise in improving the quality of low-dose CT images by reducing noise and enhancing relevant structures. These models, such as Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs), learn to generate or reconstruct images that are close to high-dose or high-quality images, thus preserving important radiomic features while reducing noise [24]. The aforementioned generative models demonstrate promising capabilities for image denoising and quality enhancement, but their application in radiomics workflows raises important methodological concerns that should be carefully considered. A critical risk associated with deep learning-based correction methods is the potential introduction of feature hallucination or distribution shift, i.e., alterations to texture, intensity patterns, or spatial correlations not perceptually obvious but significantly affecting extracted radiomic features [58,59]. In particular, GANs are trained to generate visually “plausible” images by learning statistical patterns from training data. However, this process can introduce synthetic texture components or may suppress tissue heterogeneity in such a way as to alter radiomic features such as Gray-Level Co-occurrence Matrices (GLCM), Gray-Level Run-Length Matrices (GLRLM), and wavelet-based descriptors [11,58]. As an example, GAN-based denoising may inadvertently smooth tumor heterogeneity patterns that are clinically relevant for prognosis prediction, or conversely, they can introduce artificial texture patterns not present in the original anatomy [59,60]. These alterations can propagate through the entire radiomics pipeline, potentially leading to overfitted models, loss of generalizability, or biased clinical predictions when deployed on uncorrected images. In contrast, traditional correction methods, such as wavelet-based denoising [61], or iterative reconstruction algorithms [62], operate on solid mathematical principles with predictable effects on image statistics. These approaches may be less effective compared to deep learning methods when noise is severe, but they typically preserve feature distributions more consistently, and they have a reduced risk of unpredictable feature alterations. However, traditional methods also have limitations: they may require manual tuning, exhibit reduced effectiveness, or introduce artifact patterns.

6.4. Blur

Image blur significantly impacts the accuracy of texture analysis in radiomic studies, as it can obscure or alter image details and texture, similar to other artifacts. Common sources of blur include patient movement, system resolution, imaging technique variations, and processing algorithms. Consequently, blur can lead to increased false positive or false negative rates in classifying lesions [65]. Without addressing blur, the extracted features may not accurately reflect the true pathological condition, potentially leading to incorrect disease staging, incorrect treatment planning, and poor patient outcomes. Effective methods for correcting blur in radiomic feature extraction include deconvolution techniques [66] and the use of sharpness enhancement algorithms [67]. From a radiomics perspective, image blur reduces high-frequency spatial information, introducing systematic bias and reduced stability, in particular for texture-based and higher-order radiomic features, and it may also affect shape features through inaccurate lesion boundary delineation [66].

6.5. Body Habitus/Patient Size

Overweight and obese patients increasingly present imaging challenges that can degrade image quality in diagnostic radiology (Figure 6b) [64,68], leading to increased image noise, reduced contrast, beam hardening artifacts, and streak artifacts, which can obscure anatomical structures and compromise diagnostic accuracy [69,70,71]. As patient thickness increases, photon attenuation also rises exponentially. Therefore, the goal in imaging obese patients should be to enhance image quality in accordance with the ALARA (As Low As Reasonably Achievable) principle [72]. These issues can complicate the interpretation of images in routine clinical practice, thus affecting the accuracy of feature extraction used to predict tumor behavior [73]. Additionally, there is a recognized link between diagnostic radiation from CT scans and a small, yet heightened, risk of cancer, particularly among younger populations [74]. Beyond the direct impact on image noise and artifact generation, patient size and body habitus introduce an additional level of complexity in radiomics analyses because of their interaction with scanner exposure control systems and reconstruction parameters. Scanners that employ Automatic Tube Current Modulation (ATCM) or Automatic Exposure Control (AEC) to dynamically adjust the radiation dose based on patient attenuation have demonstrated that AEC efficiency is strongly dependent on patient body size, with potential dose increases in certain patient populations [75].

7. The Impact of Image Quality on Segmentation

Image segmentation is a computer vision technique that aims to divide a digital image into distinct groups of pixels with the same characteristics, known as segments. This technique is crucial for object detection and similar tasks because it simplifies the image’s complex visual information into more manageable sections. Through image segmentation, the process of image analysis becomes quicker and more sophisticated, thus facilitating advanced image processing techniques. Factors such as noise, contrast, and artifacts have a direct impact on the performance of segmentation algorithms [76]. The segmentation accuracy also largely depends on the algorithm’s ability to identify and delineate objects, which is influenced by these image quality factors. Therefore, enhancing image quality through methods such as noise reduction and contrast enhancement can directly improve segmentation outcomes, underscoring image quality as a critical determinant of the success of image segmentation [77,78]. In diagnostic radiology, where images are analyzed to support decision-making in clinical practice, precision in segmenting anatomical structures from medical images is crucial for accurate delineation, disease diagnosis, treatment planning, and monitoring, ensuring effective patient care [79]. The accuracy of segmentation directly influences the reliability of subsequent radiomic feature extraction and, by extension, the utility of these features as biomarkers for disease characterization, prognosis, and prediction of therapy response [80]. It should be noted that different segmentation strategies such as manual delineation, semi-automatic methods, and fully automatic deep learning-based approaches exhibit sensitivity that varies with image quality degradations. Manual segmentation, while time-consuming, may benefit from expert visual compensation for artifacts; radiologists can exploit anatomical knowledge to interpret degraded regions, although image alterations may still introduce diagnostic uncertainty. Fully automatic deep learning-based segmentation models, while offering speed and consistency, can be highly sensitive to deviations in image quality characteristics that differ from their training distributions. This sensitivity can lead to segmentation failures or systematic biases when deployed on data acquired with different protocols or scanner settings. Consequently, the interaction between image quality and segmentation accuracy introduces an additional source of uncertainty in radiomic feature extraction, underscoring the critical importance of robust IQA protocols applied upstream of the segmentation step [11,76,81]. Image quality plays a pivotal role in guaranteeing the precision of image segmentation [82]. Moreover, image artifacts can lead to the extraction of unreliable radiomic features, as these features are highly sensitive to the quality and dimensions of the segmented region [83,84,85]. Features such as texture, shape, and intensity distribution are directly affected by segmentation quality, and any errors in segmentation can propagate through to the extracted features, driving misleading conclusions about the presence or characteristics of disease (Table 2).

8. Image Quality Improves Radiomic Biomarker Extraction

Image quality is critical for the fidelity of radiomics biomarkers, emphasizing the importance of optimizing imaging protocols to maximize information content while minimizing artifacts and distortions. High-quality imaging data enable the extraction of subtle and clinically relevant features, facilitating more accurate tumor characterization and therapy planning in oncology. Furthermore, standardization initiatives such as the Image Biomarker Standardization Initiative (IBSI) [18] emphasize the crucial role of image quality in ensuring consistency and reproducibility across different imaging platforms and clinical settings [22]. In addition, IBSI Phase 2 extends these standardization efforts to convolutional image filters commonly used in radiomics pipelines, such as wavelet and Laplacian-of-Gaussian filters, providing reference filtered images and reference feature values to improve the reproducibility of filtered radiomic features [22]. Cobo et al. [86] also emphasized that variations in image quality can significantly affect radiomic feature stability and reproducibility, highlighting the importance of standardized imaging protocols to maintain consistency in radiomic analyses [23,87,88]. The implementation of standardized reporting frameworks has become essential for ensuring methodological rigor in radiomics research. The Transparent Reporting of a multivariable prediction model for Individual prognosis Or Diagnosis (TRIPOD) statement provides comprehensive guidelines for reporting predictive models [89], while the Checklist for Artificial Intelligence in Medical Imaging (CLAIM) framework specifically addresses AI-based imaging studies [90]. Finally, in the context of radiomics research, the CheckList for EvaluAtion of Radiomics research (CLEAR) is designed to assess the repeatability and reproducibility of radiomic features, providing detailed guidance on the key stages of the radiomics pipeline, which include image acquisition, segmentation, preprocessing, feature extraction, and statistical analysis [23]. These frameworks complement IBSI standardization by ensuring transparent and reproducible research practices. By establishing standardized imaging protocols and quality control measures, it will be possible to mitigate variability introduced by imaging artifacts and scanner-dependent factors, thereby enhancing robustness and generalization of radiomics biomarkers. Moreover, Larue et al. [91] highlight the impact of image quality parameters, such as gray level discretization, on the stability and reliability of radiomics features. Similarly, Yadav et al. [92] conducted a phantom study to systematically evaluate how acquisition parameters such as slice thickness, dose, and reconstruction algorithm affect radiomic feature stability, identifying metrics robust to such variations. Their comprehensive phantom study underscores the importance of optimizing imaging parameters to minimize variability and maximize feature stability, thus improving the clinical utility of radiomics biomarkers [93]. Improved image quality enhances the delineation of anatomical structures and pathological regions, thus enabling more precise feature extraction. Furthermore, suitable image quality minimizes artifacts and distortions, ensuring the fidelity of radiomic features and reducing variability in quantitative measurements. While most conventional IQA metrics are designed to capture perceptual or global image degradations, several emerging approaches have been proposed to assess image quality directly in terms of radiomic feature robustness. In particular, perturbation-based stability analyses have been increasingly adopted as a practical surrogate for test-retest experiments, enabling the evaluation of feature sensitivity to controlled variations in noise, spatial resolution, voxel resampling, and segmentation uncertainty [18]. By computing stability metrics, such as the Intraclass Correlation Coefficient (ICC) across perturbed images, these methods allow the identification of features that remain robust under clinically relevant quality variations. Complementary to perturbation strategies, simulation and phantom-based frameworks have been proposed to systematically characterize radiomic feature sensitivity to acquisition and reconstruction parameters, under controlled conditions [94,95]. In this context, the Intraclass Correlation Coefficient (ICC) [96] is widely employed in radiomics studies as a practical strategy to assess feature robustness and repeatability, particularly in test-retest, multi-observer, and multi-scanner settings. Across the literature, ICC-based analyses are commonly used to exclude unstable or noise-sensitive features, with reported thresholds typically ranging between 0.7 and 0.9. However, several methodological works emphasize that ICC alone may be insufficient to fully characterize feature robustness, as it depends on the underlying data variance and acquisition conditions, and should therefore be interpreted in conjunction with image quality considerations and complementary standardization frameworks [11,12,13,14]. From an operational perspective, Image Quality Assessment should therefore be considered as a preliminary quality-control and gating step within the radiomics pipeline, to be applied before segmentation and feature extraction to identify unsuitable images for quantitative analysis, as well as to contextualize feature stability analyses, rather than as a direct modeling variable. In conclusion, the evidence from these studies collectively supports the notion that image quality improvements play a critical role in enhancing the extraction and utility of radiomics biomarkers, contributing to improved patient stratification, treatment response assessment, and personalized medicine in clinical practice [97]. A summary of the main image quality degradation sources and the corresponding mitigation strategies adopted in radiomics workflows is reported in Table 3.

9. Image Quality Metrics

Considering the significant impact of image quality on radiomic analyses, and the dependency of extracted information reliability on such quality, establishing metrics to accurately evaluate an image’s usability is fundamental [98]. Therefore, the development, selection, and implementation of appropriate metrics become critical to ensure data consistency, comparability, and ultimately, diagnostic reliability [99]. Given the diverse imaging conditions and acquisition protocols in clinical practice, robust metrics are required to assess whether an image is suitable for radiomic analysis. Traditional metrics such as Signal-to-Noise Ratio (SNR), Spatial Resolution, Contrast-to-Noise Ratio (CNR), and Uniformity have been widely adopted [100,101,102]. More recently, advanced metrics like the Learned Perceptual Image Patch Similarity (LPIPS) [103] and no-reference quality estimators like Blind/Reference less Image Spatial Quality Evaluator (BRISQUE) [104,105] have demonstrated good performance in detecting image degradations. Radiomic-specific IQA metrics should be designed to reflect not only perceptual image fidelity, but also their impact on downstream radiomic analyses. Conventional IQA measures, largely developed for visual assessment or general image processing tasks, may fail to capture degradations that minimally affect human interpretation yet substantially alter radiomic feature values, segmentation accuracy, or model robustness. For this reason, the clinical and methodological relevance of radiomic-specific IQA metrics should be evaluated through their association with quantitative endpoints intrinsic to the radiomics pipeline. Potential validation strategies include assessing the relationship between IQA scores and radiomic feature reproducibility (e.g., intraclass correlation coefficients), analyzing the sensitivity of segmentation performance metrics (e.g., Dice similarity coefficient) to controlled variations in image quality, and comparing feature stability before and after quality-improving interventions, such as denoising or harmonization. Graphical analyses, such as feature stability curves or correlation plots between IQA scores and downstream performance metrics, may provide an intuitive and reproducible framework for testing the effectiveness of radiomic-oriented IQA measures. Such approaches would allow future metrics to be benchmarked not only against visual quality, but against their ability to preserve the reliability and clinical interpretability of radiomic biomarkers.

Full-Reference vs. No-Reference Metrics

Image quality metrics can be classified into two main categories: full-reference (FR) and no-reference (NR) metrics. Full-reference metrics require a “pristine” reference image for comparing and quantifying the deviation of the target image from this ideal. These include metrics such as Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM) [106], and Learned Perceptual Image Patch Similarity (LPIPS) [103]. However, their application in radiomics is limited due to their full-reference nature. A critical question in IQA is what can be considered a true “reference” image when using FR metrics; the definition of such a reference in clinical practice remains ambiguous and context-dependent. In practical workflows, the reference is implicitly defined by expert clinical judgment, where a radiologist evaluates an image as being of “diagnostically acceptable” quality. In this context, the gold standard becomes human interpretation, which introduces subjectivity and inter-observer variability.
NR metrics, on the other hand, evaluate image quality directly from an image without any ground truth comparison. These are significantly more practical in medical imaging pipelines, particularly for large-scale clinical datasets, where reference images are rarely available. Traditional metrics such as SNR, CNR, Spatial Resolution, and Uniformity fall within this category and are frequently used in radiomics literature [100,101,102]. Among the NR metrics, BRISQUE has gained attention in image processing. BRISQUE assesses image quality by extracting natural scene statistics (NSS) from image patches and comparing them to statistical models of distortion-free images. It has been shown to effectively capture blur, noise, compression artifacts, and contrast loss without the need for a reference image. While BRISQUE offers the clear advantage of being a no-reference metric, it also presents certain limitations. Specifically, BRISQUE provides an overall image quality score without offering detailed information about the nature or source of the degradation. This lack of interpretability makes it less effective in scenarios where the identification of specific artifacts is needed. It is essential to clarify that the IQA metrics discussed in the present review (SNR, CNR, SSIM, BRISQUE, LPIPS) are designed to assess image-level quality degradations rather than to directly quantify the stability or reproducibility of radiomic features themselves. These metrics serve as upstream quality control filters to identify images with fundamental acquisition issues that would preclude reliable downstream analysis. However, traditional metrics present critical limitations when applied to radiomics workflows. SNR and CNR provide global intensity assessments, but they both fail to capture spatially localized texture disruptions that disproportionately affect second-order radiomic features such as GLCM and GLRLM [11]. The analysis of PET radiomics studies revealed that spatial resolution has the strongest effect on radiomic feature variability, underscoring that local spatial information is more critical for texture features than global intensity metrics suggest [58]. SSIM, while effective for perceptual quality tasks, exhibits decreased correlation when applied to blurred and noisy images and performs poorly for recognition threshold tasks [106]. BRISQUE, though suitable for no-reference assessment, requires retraining for medical images, since it was originally designed for natural scene statistics [105]. The fundamental limitation is that passing traditional IQA thresholds represents a necessary but not sufficient condition for robust radiomics. An image may achieve acceptable SNR or BRISQUE scores yet still yield unstable texture features due to inconsistent processing conditions such as quantization range, bin number, and reconstruction parameters, which directly affect probability distributions underlying second-order feature calculations [91].

10. Conclusions

This work provides a critical overview of measurement methods for assessing image quality in radiomics. While previous systematic reviews have addressed radiomic feature reproducibility or harmonization strategies, this review systematically quantifies the prevalence of IQA consideration in radiomics literature from a metrological standpoint. Analysis of the recent literature indicates that research efforts to date have largely focused on the development of automated tools for disease detection and evaluation. Very few studies address the correlation between poor-quality images and their effect on disease detection, both in human interpretation and automated analysis. The reviewed literature reports that approximately 2.4% of 1,210,000 radiology images were affected by alterations, corresponding to approximately 30,000 cases in which image quality could contribute to misinterpretation and potential misdiagnosis. It is worth noting that this percentage derives from a heterogeneous clinical cohort that spans multiple imaging modalities and clinical contexts. Direct extrapolation of these figures should be carefully interpreted, as the prevalence and impact of quality issues may vary depending on imaging protocols, scanner technologies, clinical applications, and the specific adopted feature extraction pipelines. Nevertheless, this underscores the substantial clinical and economic burden of undetected or unaddressed image quality degradation in quantitative imaging analyses. Moreover, it must be highlighted that this number may substantially increase in the case of PET, CT, and MRI, which are widely used in tumor detection. When the analysis is performed by human operators, the clinical results include notes about poor image quality, strongly affecting the reliability of the diagnosis. This overview has also reported the main alterations by classifying them according to the radiomic image type they affect. The most commonly used metrics for assessing image quality are reported as well. Despite the critical role of these metrics, there are notable gaps and challenges when applying them to radiomics [107]. Many existing metrics were developed for visual IQA, not for quantitative analysis of high-dimensional data extracted from images [91]. Another challenge is the variability in imaging conditions and protocols across different institutions and even within the same institution over time. This variability can lead to significant differences in image quality and, consequently, in the extracted radiomic features, thus complicating multicenter studies and the generalization of radiomic models [28,29,30]. Moreover, the interaction between image quality metrics and radiomic features is complex and not fully understood [108]. Some radiomic features may be more sensitive to changes in certain image quality metrics than others, and the impact of this sensitivity on clinical decision-making is still being investigated [109]. It should be acknowledged that radiomic feature reliability is strongly modality-dependent, introducing an additional layer of complexity in interpreting aggregated quality metrics across heterogeneous imaging studies. Cross-sectional imaging modalities such as CT and MRI, which rely on standardized acquisition protocols and well-controlled physical parameters, generally demonstrate higher feature stability and reproducibility compared to operator-dependent or real-time imaging techniques such as ultrasound. Future work should take into account radiomic feature reproducibility across modalities using standardized phantoms and harmonized protocols, as recommended by initiatives such as IBSI, to establish specific quality thresholds and feature selection criteria in different modalities. Future systematic reviews could complement this work by performing stratified quantitative analyses that link specific artifact types to feature-class stability (e.g., first-order, texture, higher-order features) across imaging modalities, using standardized phantom datasets and harmonized feature extraction protocols. Due to the metrological nature of the review, meta-analysis and formal risk of bias assessment are omitted, but they will be considered in a future applied paper. To address these gaps, there is a growing demand for the development of new metrics and standards that are specifically designed for radiomics [107,110,111]. Addressing this need will be crucial for unlocking the full potential of radiomics in clinical applications, ensuring that the quantitative features extracted from medical images are accurate, reliable, and meaningful for patient care.

Author Contributions

Conceptualization, S.C. and F.L.; methodology, S.C. and F.L.; formal analysis, F.F.; investigation, F.F.; validation, D.L.C. and S.C.; writing—original draft preparation, F.F.; writing—review and editing, D.L.C. and S.C.; supervision, S.C. and F.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by: POS, RADIOAMICA—Open Network per la RADIOmica/rAdiogenoMIca Cooperativa basata su intelligenza Artificiale. CUP H53C22000650006.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IQAImage Quality Assessment
CTComputed Tomography
MRIMagnetic Resonance Imaging
PETPositron Emission Tomography
AIArtificial Intelligence
GANGenerative Adversarial Network
VAEVariational Autoencoder
SNRSignal-to-Noise Ratio
CNRContrast-to-Noise Ratio
SSIMStructural Similarity Index
LPIPSLearned Perceptual Image Patch Similarity
BRISQUEBlind/Referenceless Image Spatial Quality Evaluator
IBSIImage Biomarker Standardization Initiative
TRIPODTransparent Reporting of a Multivariable Prediction Model for
Individual Prognosis or Diagnosis
CLAIMChecklist for Artificial Intelligence in Medical Imaging
ALARAAs Low As Reasonably Achievable
FRFull-Reference
NRNo-Reference
ICCIntraclass Correlation Coefficient
GLCMGray-Level Co-occurrence Matrix
GLRLMGray-Level Run-Length Matrix
PRISMAPreferred Reporting Items for Systematic Reviews and Meta-Analyses
QUADAS-2Quality Assessment of Diagnostic Accuracy Studies, 2nd edition
GRADEGrading of Recommendations Assessment, Development and Evaluation

References

  1. Yip, S.S.; Aerts, H.J. Applications and limitations of radiomics. Phys. Med. Biol. 2016, 61, R150–R166. [Google Scholar] [CrossRef]
  2. Mariotti, F.; Agostini, A.; Borgheresi, A.; Marchegiani, M.; Zannotti, A.; Giacomelli, G.; Pierpaoli, L.; Tola, E.; Galiffa, E.; Giovagnoni, A. Insights into radiomics: A comprehensive review for beginners. Clin. Transl. Oncol. 2025, 27, 4091–4102. [Google Scholar] [CrossRef] [PubMed]
  3. Avanzo, M.; Stancanello, J.; El Naqa, I. Beyond imaging: The promise of radiomics. Phys. Medica 2017, 38, 122–139. [Google Scholar] [CrossRef]
  4. McCague, C.; Ramlee, S.; Reinius, M.; Selby, I.; Hulse, D.; Piyatissa, P.; Bura, V.; Crispin-Ortuzar, M.; Sala, E.; Woitek, R. Introduction to radiomics for a clinical audience. Clin. Radiol. 2023, 78, 83–98. [Google Scholar] [CrossRef] [PubMed]
  5. Shur, J.D.; Doran, S.J.; Kumar, S.; ap Dafydd, D.; Downey, K.; O’Connor, J.P.B.; Papanikolaou, N.; Messiou, C.; Koh, D.-M.; Orton, M.R. Radiomics in Oncology: A Practical Guide. RadioGraphics 2021, 41, 1717–1732. [Google Scholar] [CrossRef]
  6. Park, J.E.; Kim, D.; Kim, H.S.; Park, S.Y.; Kim, J.Y.; Cho, S.J.; Shin, J.H.; Kim, J.H. Quality of science and reporting of radiomics in oncologic studies: Room for improvement according to radiomics quality score and TRIPOD statement. Eur. Radiol. 2020, 30, 523–536. [Google Scholar] [CrossRef]
  7. Fass, L. Imaging and cancer: A review. Mol. Oncol. 2008, 2, 115–152. [Google Scholar] [CrossRef]
  8. Li, H.; Giger, M.L.; Yuan, Y.; Chen, W.; Horsch, K.; Lan, L.; Jamieson, A.R.; Sennett, C.A.; Jansen, S.A. Evaluation of Computer-Aided Diagnosis on a Large Clinical Full-Field Digital Mammographic Dataset. Acad. Radiol. 2008, 15, 1437–1445. [Google Scholar] [CrossRef]
  9. El Naqa, I.; Grigsby, P.; Apte, A.; Kidd, E.; Donnelly, E.; Khullar, D.; Chaudhari, S.; Yang, D.; Schmitt, M.; Laforest, R.; et al. Exploring feature-based approaches in PET images for predicting cancer treatment outcomes. Pattern Recognit. 2009, 42, 1162–1171. [Google Scholar] [CrossRef] [PubMed]
  10. Kumar, V.; Gu, Y.; Basu, S.; Berglund, A.; Eschrich, S.A.; Schabath, M.B.; Forster, K.; Aerts, H.J.W.L.; Dekker, A.; Fenstermacher, D.; et al. Radiomics: The process and the challenges. Magn. Reson. Imaging 2012, 30, 1234–1248. [Google Scholar] [CrossRef]
  11. van Timmeren, J.; Cester, D.; Tanadini-Lang, S.; Alkadhi, H.; Baessler, B. Radiomics in medical imaging—“how-to” guide and critical reflection. Insights Imaging 2020, 11, 91. [Google Scholar] [CrossRef]
  12. Zhao, B.; Tan, Y.; Tsai, W.Y.; Qi, J.; Xie, C.; Lu, L.; Schwartz, L.H. Reproducibility of radiomics for deciphering tumor phenotype with imaging. Sci. Rep. 2016, 6, 23428. [Google Scholar] [CrossRef]
  13. Cui, Y.; Yin, F.F. Impact of image quality on radiomics applications. Phys. Med. Biol. 2022, 67, 15TR03. [Google Scholar] [CrossRef] [PubMed]
  14. Traverso, A.; Wee, L.; Dekker, A.; Gillies, R. Repeatability and Reproducibility of Radiomic Features: A Systematic Review. Int. J. Radiat. Oncol. 2018, 102, 1143–1158. [Google Scholar] [CrossRef]
  15. Gillies, R.J.; Kinahan, P.E.; Hricak, H. Radiomics: Images Are More than Pictures, They Are Data. Radiology 2016, 278, 563–577. [Google Scholar] [CrossRef]
  16. Chauvie, S.; Mazzoni, L.N.; O’Doherty, J. A review on the use of imaging biomarkers in oncology clinical trials: Quality assurance strategies for technical validation. Tomography 2023, 9, 1876–1902. [Google Scholar] [CrossRef]
  17. Jahanshahi, A.; Soleymani, Y.; Fazel Ghaziani, M.; Khezerloo, D. Radiomics reproducibility challenge in computed tomography imaging as a nuisance to clinical generalization: A mini-review. J. Radiol. Nucl. Med. 2023, 54, 83. [Google Scholar] [CrossRef]
  18. Zwanenburg, A.; Vallières, M.; Abdalah, M.A.; Aerts, H.J.W.L.; Andrearczyk, V.; Apte, A.; Ashrafinia, S.; Bakas, S.; Beukinga, R.J.; Boellaard, R.; et al. The Image Biomarker Standardization Initiative: Standardized Quantitative Radiomics for High-Throughput Image-based Phenotyping. Radiology 2020, 295, 328–338. [Google Scholar] [CrossRef] [PubMed]
  19. Shen, D.; Wu, G.; Suk, H.-I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef]
  20. Zhao, B. Understanding sources of variation to improve the reproducibility of radiomics. Front. Oncol. 2021, 11, 633176. [Google Scholar] [CrossRef] [PubMed]
  21. Aerts, H.J.W.L.; Velazquez, E.R.; Leijenaar, R.T.H.; Parmar, C.; Grossmann, P.; Carvalho, S.; Bussink, J.; Monshouwer, R.; Haibe-Kains, B.; Rietveld, D.; et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat. Commun. 2014, 5, 4006. [Google Scholar] [CrossRef] [PubMed]
  22. Depeursinge, A.; Andréarczyk, V.; Whybra, P.; van Griethuysen, J.J.; Müller, H.; Schaer, R.; Vallières, M.; Zwanenburg, A. Standardised convolutional filtering for radiomics. arXiv 2020, arXiv:2006.05470. [Google Scholar]
  23. Pfaehler, E.; Zhovannik, I.; Wei, L.; Boellaard, R.; Dekker, A.; Monshouwer, R.; El Naqa, I.; Bussink, J.; Gillies, R.; Wee, L.; et al. A systematic review and quality of reporting checklist for repeatability and reproducibility of radiomic features. Phys. Imaging Radiat. Oncol. 2021, 20, 69–75. [Google Scholar] [CrossRef]
  24. Yi, X.; Walia, E.; Babyn, P. Generative Adversarial Network in Medical Imaging: A Review. Med. Image Anal. 2019, 58, 101552. [Google Scholar] [CrossRef] [PubMed]
  25. Wang, Q.; Liu, H.; Xie, G.; Zhang, Y. Image Denoising Using an Improved Generative Adversarial Network with Wasserstein Distance. In Proceedings of the 2021 40th Chinese Control Conference (CCC), Shanghai, China, 26–28 July 2021; pp. 7027–7032. [Google Scholar]
  26. Wolterink, J.M.; Leiner, T.; Viergever, M.A.; Išgum, I. Generative Adversarial Networks for Noise Reduction in Low-Dose CT. IEEE Trans. Med. Imaging 2017, 36, 2536–2545. [Google Scholar] [CrossRef]
  27. Kalare, K.; Bajpai, M.; Sarkar, S.; Dubey, S.R. Deep neural network for beam hardening artifacts removal in image reconstruction. Appl. Intell. 2022, 52, 6037–6056. [Google Scholar] [CrossRef]
  28. Mali, S.A.; Ibrahim, A.; Woodruff, H.C.; Andrearczyk, V.; Müller, H.; Primakov, S.; Salahuddin, Z.; Chatterjee, A.; Lambin, P. Making Radiomics More Reproducible across Scanner and Imaging Protocol Variations: A Review of Harmonization Methods. J. Pers. Med. 2021, 11, 842. [Google Scholar] [CrossRef]
  29. Kállos-Balogh, Z.; Egri, Á.; Sera, A.; Horváth, I.K.; Trencsényi, E.; Garai, I. Multicentric study on the reproducibility and robustness of PET-based radiomics features with a realistic activity painting phantom. PLoS ONE 2024, 19, e0309540. [Google Scholar] [CrossRef]
  30. Horvat, N.; Papanikolaou, N.; Koh, D.-M. Radiomics Beyond the Hype: A Critical Evaluation Toward Oncologic Clinical Use. Radiol. Artif. Intell. 2024, 6, e230437. [Google Scholar] [CrossRef]
  31. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef] [PubMed]
  32. Schiavo, J.H. PROSPERO: An International Register of Systematic Review Protocols. Med. Ref. Serv. Q. 2019, 38, 171–180. [Google Scholar] [CrossRef] [PubMed]
  33. Whiting, P.F.; Rutjes, A.W.; Westwood, M.E.; Mallett, S.; Deeks, J.J.; Reitsma, J.B.; Leeflang, M.M.; Sterne, J.A.; Bossuyt, P.M. QUADAS-2: A revised tool for the quality assessment of diagnostic accuracy studies. Ann. Intern. Med. 2011, 155, 529–536. [Google Scholar] [CrossRef] [PubMed]
  34. Minozzi, S.; Cinquini, M.; Arienti, C.; Battain, P.C.; Brigadoi, G.; Del Vicario, M.; Di Domenico, G.; Farma, T.; Federico, S.; Innocenti, T.; et al. Grading of Recommendations Assessment, Development, and Evaluation and Confidence in Network Meta-Analysis showed moderate to substantial concordance in the evaluation of certainty of the evidence. J. Clin. Epidemiol. 2025, 184, 111811. [Google Scholar] [CrossRef] [PubMed]
  35. Lambin, P.; Rios-Velazquez, E.; Leijenaar, R.T.H.; Carvalho, S.; van Stiphout, R.G.P.M.; Granton, P.; Zegers, C.M.L.; Gillies, R.J.; Boellard, R.; Dekker, A.; et al. Radiomics: Extracting more information from medical images using advanced feature analysis. Eur. J. Cancer 2012, 48, 441–446. [Google Scholar] [CrossRef]
  36. Ammirato, S.; Felicetti, A.M.; Rogano, D.; Linzalone, R.; Corvello, V. Digitalising the Systematic Literature Review process: The MySLR platform. Knowl. Manag. Res. Pract. 2022, 21, 777–794. [Google Scholar] [CrossRef]
  37. Mabotuwana, T.; Bhandarkar, V.S.; Hall, C.S.; Gunn, M.L. Detecting Technical Image Quality in Radiology Reports. AMIA Annu. Symp. Proc. 2018, 2018, 780–788. [Google Scholar]
  38. Krupa, K.; Bekiesińska-Figatowska, M. Artifacts in magnetic resonance imaging. Pol. J. Radiol. 2015, 80, 93–106. [Google Scholar] [CrossRef] [PubMed]
  39. Smith, T.B.; Nayak, K.S. MRI artifacts and correction strategies. Imaging Med. 2010, 2, 445–457. [Google Scholar] [CrossRef]
  40. Budrys, T.; Veikutis, V.; Lukosevicius, S.; Gleizniene, R.; Monastyreckiene, E.; Kulakiene, I. Artifacts in magnetic resonance imaging: How it can really affect diagnostic image quality and confuse clinical diagnosis? J. Vibroengineering 2018, 20, 1202–1213. [Google Scholar] [CrossRef]
  41. Ye, H.; Wong, K.P.; Wardak, M.; Dahlbom, M.; Kepe, V.; Barrio, J.R.; Nelson, L.D.; Small, G.W.; Huang, S.C. Automated movement correction for dynamic PET/CT images: Evaluation with phantom and patient data. PLoS ONE 2014, 9, e103745. [Google Scholar] [CrossRef] [PubMed]
  42. Sommer, K.; Saalbach, A.; Brosch, T.; Hall, C.; Cross, N.M.; Andre, J.B. Correction of Motion Artifacts Using a Multiscale Fully Convolutional Neural Network. AJNR. Am. J. Neuroradiol. 2020, 41, 416–423. [Google Scholar] [CrossRef]
  43. Mohebbian, M.R.; Walia, E.; Wahid, K.A. Which K-Space Sampling Schemes is good for Motion Artifact Detection in Magnetic Resonance Imaging? arXiv 2021, arXiv:2103.08516. [Google Scholar] [CrossRef]
  44. Huang, C.; Ackerman, J.L.; Petibon, Y.; Brady, T.J.; El Fakhri, G.; Ouyang, J. MR-based motion correction for PET imaging using wired active MR microcoils in simultaneous PET-MR: Phantom study. Med. Phys. 2014, 41, 041910. [Google Scholar] [CrossRef]
  45. Zaitsev, M.; Maclaren, J.; Herbst, M. Motion artifacts in MRI: A complex problem with many partial solutions. J. Magn. Reson. Imaging 2015, 42, 887–901. [Google Scholar] [CrossRef]
  46. Küstner, T.; Armanious, K.; Yang, J.; Yang, B.; Schick, F.; Gatidis, S. Retrospective correction of motion-affected MR images using deep learning frameworks. Magn. Reson. Med. 2019, 82, 1527–1540. [Google Scholar] [CrossRef] [PubMed]
  47. Kadah, Y.M. Motion artifact suppression in MRI using registration of segmented acquisitions. J. Eng. Appl. Sci. 2007, 54, 673–688. [Google Scholar]
  48. Pessis, E.; Campagna, R.; Sverzut, J.-M.; Bach, F.; Rodallec, M.; Guerini, H.; Feydy, A.; Drapé, J.-L. Virtual monochromatic spectral imaging with fast kilovoltage switching: Reduction of metal artifacts at CT. Radiographics 2013, 33, 573–583. [Google Scholar] [CrossRef]
  49. Revision Radiology. Revision Radiology—Radiology Education Resources. Available online: http://www.revisionrads.com/ (accessed on 30 January 2026).
  50. Xie, S.; Li, C.; Li, H.; Ge, Q. A level set method for cupping artifact correction in cone-beam CT. Med. Phys. 2015, 42, 4888–4895. [Google Scholar] [CrossRef]
  51. De Man, B.; Nuyts, J.; Dupont, P.; Marchal, G.; Suetens, P. Reduction of metal streak artifacts in X-ray computed tomography using a transmission maximum a posteriori algorithm. IEEE Trans. Nucl. Sci. 2000, 47, 977–981. [Google Scholar] [CrossRef]
  52. Yu, L.; Leng, S.; McCollough, C.H. Dual-energy CT-based monochromatic imaging. AJR Am. J. Roentgenol. 2012, 199, S9–S15. [Google Scholar] [CrossRef]
  53. Stowe, J. Suppression of the CT beam hardening streak artifact using predictive correction on detector data. Univers. J. Med. Sci. 2016, 4, 69–80. [Google Scholar] [CrossRef]
  54. Zhang, Y.; Yu, H. Convolutional neural network based metal artifact reduction in X-ray computed tomography. IEEE Trans. Med. Imaging 2018, 37, 1370–1381. [Google Scholar] [CrossRef]
  55. Mackin, D.; Fave, X.; Zhang, L.; Fried, D.; Yang, J.; Taylor, B.; Rodriguez-Rivera, E.; Dodge, C.; Jones, A.K.; Court, L. Measuring Computed Tomography Scanner Variability of Radiomics Features. J. Comput. Assist. Tomogr. 2015, 39, 903–911. [Google Scholar] [CrossRef]
  56. Meyer, M.; Ronald, J.; Vernuccio, F.; Nelson, R.C.; Ramirez-Giraldo, J.C.; Solomon, J.; Patel, B.N.; Samei, E.; Marin, D. Reproducibility of CT Radiomic Features within the Same Patient: Influence of Radiation Dose and CT Reconstruction Settings. Radiology 2020, 294, 583–591. [Google Scholar] [CrossRef]
  57. O’Connor, J.P.B.; Aboagye, E.; Adams, J.E.; Aerts, H.J.W.L.; Barrington, S.F.; Beer, A.J.; Boellaard, R.; Bohndiek, S.E.; Brady, M.; Brown, G.; et al. Imaging biomarker roadmap for cancer studies. Nat. Rev. Clin. Oncol. 2017, 14, 169–186. [Google Scholar] [CrossRef]
  58. Shiri, I.; Rahmim, A.; Ghaffarian, P.; Geramifar, P.; Abdollahi, H.; Bitarafan-Rajabi, A. The impact of image reconstruction settings on 18F-FDG PET radiomic features: Multi-scanner phantom and patient studies. Eur. Radiol. 2020, 30, 4498–4509. [Google Scholar] [CrossRef]
  59. Umehara, K.; Ota, J.; Ishida, T. Application of super-resolution convolutional neural network for enhancing image resolution in chest CT. J. Digit. Imaging 2018, 31, 441–450. [Google Scholar] [CrossRef] [PubMed]
  60. Jiang, J.; Hu, Y.-C.; Liu, C.; Halpenny, D.; Hellmann, M.D.; Deasy, J.O.; Mageras, G.; Veeraraghavan, H. Multiple resolution residually connected feature streams for automatic lung tumor segmentation from CT images. IEEE Trans. Med. Imaging 2021, 40, 134–144. [Google Scholar] [CrossRef] [PubMed]
  61. Juntu, J.; Sijbers, J.; Van Dyck, D.; Gielen, J. Machine learning study of several classifiers trained with texture analysis features to differentiate benign from malignant soft-tissue tumors in T1-MRI images. J. Magn. Reson. Imaging 2010, 31, 680–689. [Google Scholar] [CrossRef]
  62. Geyer, L.L.; Schoepf, U.J.; Meinel, F.G.; Nance, J.W., Jr.; Bastarrika, G.; Leipsic, J.A.; Paul, N.S.; Rengo, M.; Laghi, A.; De Cecco, C.N. State of the art: Iterative CT reconstruction techniques. Radiology 2015, 276, 339–357. [Google Scholar] [CrossRef] [PubMed]
  63. Misko, O.; Veseliak, V. Singular Value Decomposition in Image Noise Filtering; Ukrainian Catholic University: Lviv, Ukraine, 2019. [Google Scholar]
  64. Uppot, R.N. Technical Challenges of Imaging and Image-Guided Interventions in Obese Patients. Br. J. Radiol. 2018, 91, 20170931. [Google Scholar] [CrossRef] [PubMed]
  65. Tumakov, D.; Kayumov, Z.; Zhumaniezov, A.; Chikrin, D.; Galimyanov, D. Elimination of Defects in Mammograms Caused by a Malfunction of the Device Matrix. J. Imaging 2022, 8, 128. [Google Scholar] [CrossRef] [PubMed]
  66. Gang, G.J.; Deshpande, R.; Stayman, J.W. Standardization of Histogram- and Gray-Level Co-Occurrence Matrices-Based Radiomics in the Presence of Blur and Noise. Phys. Med. Biol. 2021, 66, 074004. [Google Scholar] [CrossRef] [PubMed]
  67. Lee, M.H.; Yun, C.S.; Kim, K.; Lee, Y.; Alzheimer Disease Neuroimaging Initiative. Image Restoration Algorithm Incorporating Methods to Remove Noise and Blurring from Positron Emission Tomography Imaging for Alzheimer’s Disease Diagnosis. Phys. Medica 2022, 103, 181–189. [Google Scholar] [CrossRef]
  68. Feldmane, I.; Gampp, C.; Hausmann, D.; Mavridis, S.; Euler, A.; Hefermehl, L.J.; Knoth, F.; Kubik-Huch, R.A.; Nocito, A.; Niemann, T. Evaluation of image quality of overweight and obese patients in CT using high data rate detectors. In Vivo 2023, 37, 1186–1191. [Google Scholar] [CrossRef]
  69. Schindera, S.T.; Nelson, R.C.; Toth, T.L.; Nguyen, G.T.; Toncheva, G.I.; DeLong, D.M.; Yoshizumi, T.T. Effect of patient size on radiation dose for abdominal MDCT with automatic tube current modulation: Phantom study. AJR Am. J. Roentgenol. 2008, 190, W100–W105. [Google Scholar] [CrossRef]
  70. Wang, A.J.; Goldsmith, Z.G.; Wang, C.; Nguyen, G.; Astroza, G.M.; Neisius, A.; Iqbal, M.W.; Neville, A.M.; Lowry, C.; Toncheva, G.; et al. Obesity triples the radiation dose of stone protocol computerized tomography. J. Urol. 2013, 189, 2142–2146. [Google Scholar] [CrossRef]
  71. Ginde, A.A.; Foianini, A.; Renner, D.M.; Valley, M.; Camargo, C.A., Jr. The challenge of CT and MRI imaging of obese individuals who present to the emergency department: A national survey. Obesity 2008, 16, 2549–2551. [Google Scholar] [CrossRef]
  72. Seo, G.; Robinson, J.; Punch, A.; Jimenez, Y.; Lewis, S. Understanding radiographic decision-making when imaging obese patients: A Think-Aloud study. J. Med. Radiat. Sci. 2022, 69, 13–23. [Google Scholar] [CrossRef]
  73. Uppot, R.N.; Sahani, D.V.; Hahn, P.F.; Kalra, M.K.; Saini, S.S.; Mueller, P.R. Effect of obesity on image quality: Fifteen-year longitudinal study for evaluation of dictated radiology reports. Radiology 2006, 240, 435–439. [Google Scholar] [CrossRef]
  74. Mathews, J.D.; Forsythe, A.V.; Brady, Z.; Butler, M.W.; Goergen, S.K.; Byrnes, G.B.; Giles, G.G.; Wallace, A.B.; Anderson, P.R.; Guiver, T.A.; et al. Cancer risk in 680,000 people exposed to computed tomography scans in childhood or adolescence: Data linkage study of 11 million Australians. BMJ 2013, 346, f2360. [Google Scholar] [CrossRef] [PubMed]
  75. Israel, G.M.; Cicchiello, L.; Brink, J.; Huda, W. Patient Size and Radiation Exposure in Thoracic, Pelvic, and Abdominal CT Examinations Performed with Automatic Exposure Control. Am. J. Roentgenol. 2010, 195, 1340–1346. [Google Scholar] [CrossRef]
  76. Pednekar, G.V.; Udupa, J.K.; McLaughlin, D.J.; Wu, X.; Tong, Y.; Simone, C.B.; Camaratta, J.; Torigian, D.A. Image Quality and Segmentation. Proc. SPIE—Int. Soc. Opt. Eng. 2018, 10576, 105762N. [Google Scholar] [PubMed]
  77. Soewondo, W.; Haji, S.O.; Eftekharian, M.; Marhoon, H.A.; Dorofeev, A.E.; Jalil, A.T.; Jawad, M.A.; Jabbar, A.H. Noise reduction and mammography image segmentation optimization with novel QIMFT-SSA method. Comput. Opt. 2022, 46, 298–307. [Google Scholar] [CrossRef]
  78. Koçak, B.; Yüzkan, S.; Mutlu, S.; Karagülle, M.; Kala, A.; Kadıoğlu, M. Influence of image preprocessing on the segmentation-based reproducibility of radiomic features: In vivo experiments on discretization and resampling parameters. Diagn. Interv. Radiol. 2024, 30, 152–162. [Google Scholar] [CrossRef]
  79. Pluim, J.P.W.; Maintz, J.B.A.; Viergever, M.A. Mutual-information-based registration of medical images: A survey. IEEE Trans. Med. Imaging 2003, 22, 986–1004. [Google Scholar] [CrossRef]
  80. deSouza, N.M.; van der Lugt, A.; Deroose, C.M.; Alberich-Bayarri, A.; Bidaut, L.; Fournier, L.; Costaridou, L.; Oprea-Lager, D.E.; Kotter, E.; Smits, M.; et al. Standardised lesion segmentation for imaging biomarker quantitation: A consensus recommendation from ESR and EORTC. Insights Imaging 2022, 13, 159. [Google Scholar] [CrossRef]
  81. Chupetlovska, K.; Akinci D’Antonoli, T.; Bodalal, Z.; Abdelatty, M.A.; Erenstein, H.; Santinha, J.; Huisman, M.; Visser, J.J.; Trebeschi, S.; Lipman, G. ESR Essentials: A step-by-step guide of segmentation for radiologists—practice recommendations by the European Society of Medical Imaging Informatics. Eur. Radiol. 2025, 35, 6894–6904. [Google Scholar] [CrossRef]
  82. Ullah, F.; Nadeem, M.; Abrar, M.; Amin, F.; Salam, A.; Khan, S. Enhancing Brain Tumor Segmentation Accuracy through Scalable Federated Learning with Advanced Data Privacy and Security Measures. Mathematics 2023, 11, 4189. [Google Scholar] [CrossRef]
  83. Roy, S.; Whitehead, T.D.; Quirk, J.D.; Salter, A.; Ademuyiwa, F.O.; Li, S.; An, H.; Shoghi, K.I. Optimal co-clinical radiomics: Sensitivity of radiomic features to tumour volume, image noise and resolution in co-clinical T1-weighted and T2-weighted magnetic resonance imaging. eBioMedicine 2020, 59, 102963. [Google Scholar] [CrossRef]
  84. Ahmed, I.; Balestrieri, E.; Daponte, P.; Imperatore, R.; Lamonaca, F.; Paolucci, M.; Picariello, F. Morphometric measurement of fish blood cell: An image processing and ellipse fitting technique. IEEE Trans. Instrum. Meas. 2024, 73, 5011712. [Google Scholar] [CrossRef]
  85. Ceccarelli, M.; Speranza, A.; Grimaldi, D.; Lamonaca, F. Automatic detection and surface measurements of micronucleus by a computer vision approach. IEEE Trans. Instrum. Meas. 2010, 59, 2383–2390. [Google Scholar] [CrossRef]
  86. Cobo, M.; Menéndez Fernández-Miranda, P.; Bastarrika, G.; Iglesias, L.L. Enhancing radiomics and Deep Learning systems through the standardization of medical imaging workflows. Sci. Data 2023, 10, 732. [Google Scholar] [CrossRef] [PubMed]
  87. Soliman, M.A.S.; Kelahan, L.C.; Magnetta, M.; Savas, H.; Agrawal, R.; Avery, R.J.; Aouad, P.; Liu, B.; Xue, Y.; Chae, Y.K.; et al. A Framework for Harmonization of Radiomics Data for Multicenter Studies and Clinical Trials. JCO Clin. Cancer Inform. 2022, 6, e2200023. [Google Scholar] [CrossRef]
  88. Trojani, V.; Bassi, M.C.; Verzellesi, L.; Bertolini, M. Impact of Preprocessing Parameters in Medical Imaging-Based Radiomic Studies: A Systematic Review. Cancers 2024, 16, 2668. [Google Scholar] [CrossRef]
  89. Collins, G.S.; Reitsma, J.B.; Altman, D.G.; Moons, K.G. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): The TRIPOD statement. BMJ 2015, 350, g7594. [Google Scholar] [CrossRef]
  90. Mongan, J.; Moy, L.; Kahn, C.E., Jr. Checklist for Artificial Intelligence in Medical Imaging (CLAIM): A guide for authors and reviewers. Radiol. Artif. Intell. 2020, 2, e200029. [Google Scholar] [CrossRef]
  91. Larue, R.T.H.M.; van Timmeren, J.E.; de Jong, E.E.C.; Feliciani, G.; Leijenaar, R.T.H.; Schreurs, W.M.J.; Sosef, M.N.; Raat, F.H.P.J.; van der Zande, F.H.R.; Das, M.; et al. Influence of gray level discretization on radiomic feature stability for different CT scanners, tube currents and slice thicknesses: A comprehensive phantom study. Acta Oncol. 2017, 56, 1544–1553. [Google Scholar] [CrossRef]
  92. Yadav, N.; Turangan, A.; Han, H.; Lei, X.; Hwang, D.H.; Cen, S.; Levy, J.; Jensen, K.; Goodenough, D.J.; Varghese, B.A. Systematic Approach to Identifying Sources of Variation in CT Radiomics: A Phantom Study. In Proceedings of the 20th International Symposium on Medical Information Processing and Analysis (SIPAIM), Antigua, Guatemala, 13–15 November 2024; pp. 1–4. [Google Scholar]
  93. Teng, X.; Song, J.; Zhang, X.; Wang, Y.; Shao, Z. Enhancing the Clinical Utility of Radiomics: Addressing Key Challenges in Repeatability and Reproducibility. Diagnostics 2024, 14, 1835. [Google Scholar] [CrossRef] [PubMed]
  94. Hertel, A.; Tharmaseelan, H.; Rotkopf, L.T.; Nörenberg, D.; Riffel, P.; Nikolaou, K.; Weiss, J.; Bamberg, F.; Schoenberg, S.O.; Froelich, M.F. Phantom-based radiomics feature test–retest stability analysis on photon-counting detector CT. Eur. Radiol. 2023, 33, 4905–4914. [Google Scholar] [CrossRef] [PubMed]
  95. Flouris, K.; Jimenez-del-Toro, O.; Aberle, C.; Bach, M.; Schaer, R.; Obmann, M.; Stieltjes, B.; Müller, H.; Depeursinge, A.; Konukoglu, E. Assessing radiomics feature stability with simulated CT acquisitions. Sci. Rep. 2022, 12, 6942. [Google Scholar] [CrossRef]
  96. Shrout, P.E.; Fleiss, J.L. Intraclass correlation coefficient. In Encyclopedia of Statistical Sciences; Kotz, S., Johnson, N.L., Eds.; John Wiley & Sons: New York, NY, USA, 1982; Volume 4, pp. 213–217. [Google Scholar]
  97. Limkin, E.J.; Sun, R.; Dercle, L.; Zacharaki, E.I.; Robert, C.; Reuzé, S.; Schernberg, A.; Paragios, N.; Deutsch, E.; Ferté, C. Promises and challenges for the implementation of computational medical imaging (radiomics) in oncology. Ann. Oncol. 2017, 28, 1191–1206. [Google Scholar] [CrossRef] [PubMed]
  98. Felicetti, F.; Carnì, D.L.; Lamonaca, F. Fish blood cell as biological dosimeter: In between measurements, radiomics, preprocessing, and artificial intelligence. In Innovations in Computational Intelligence and Computer Vision; Lecture Notes in Networks and Systems; Springer: Singapore, 2024; Volume 1117. [Google Scholar]
  99. Perniciano, A.; Loddo, A.; Di Ruberto, C.; Pes, B. Insights into radiomics: Impact of feature selection and classification. Multimed. Tools Appl. 2025, 84, 31695–31721. [Google Scholar] [CrossRef]
  100. Nadipally, M. Optimization of Methods for Image-Texture Segmentation Using Ant Colony Optimization. In Intelligent Data Analysis for Biomedical Applications; Academic Press: Cambridge, MA, USA, 2019; pp. 21–47. [Google Scholar]
  101. Welvaert, M.; Rosseel, Y. On the definition of signal-to-noise ratio and contrast-to-noise ratio for FMRI data. PLoS ONE 2013, 8, e77089. [Google Scholar] [CrossRef] [PubMed]
  102. Raunig, D.L.; McShane, L.M.; Pennello, G.; Gatsonis, C.; Carson, P.L.; Voyvodic, J.T.; Wahl, R.L.; Kurland, B.F.; Schwarz, A.J.; Gönen, M.; et al. Quantitative imaging biomarkers: A review of statistical methods for technical performance assessment. Stat. Methods Med. Res. 2015, 24, 27–67. [Google Scholar] [CrossRef] [PubMed]
  103. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 586–595. [Google Scholar]
  104. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  105. Chow, L.S.; Rajagopal, H. Modified-BRISQUE as no reference image quality assessment for structural MR images. Magn. Reson. Imaging 2017, 43, 74–87. [Google Scholar] [CrossRef]
  106. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  107. D’Antonoli, T.A.; Cavallo, A.U.; Vernuccio, F.; Stanzione, A.; Klontzas, M.E.; Cannella, R.; Ugga, L.; Baran, A.; Fanni, S.C.; Petrash, E.; et al. Reproducibility of radiomics quality score: An intra- and inter-rater reliability study. Eur. Radiol. 2024, 34, 2024. [Google Scholar] [CrossRef]
  108. Mayerhoefer, M.A.; Papanikolaou, N.; Aerts, H.J.W.L. Introduction to Radiomics. J. Nucl. Med. 2020, 61, 488–495. [Google Scholar] [CrossRef]
  109. Fernández Patón, M.; Cerdá Alberich, L.; Sangüesa Nebot, C.; Martínez de Las Heras, B.; Veiga Canuto, D.; Cañete Nieto, A.; Martí-Bonmatí, L. MR Denoising Increases Radiomic Biomarker Precision and Reproducibility in Oncologic Imaging. J. Digit. Imaging 2021, 34, 1134–1145. [Google Scholar] [CrossRef] [PubMed]
  110. Koçak, B.; Baessler, B.; Bakas, S.; Cuocolo, R.; Fedorov, A.; Maier-Hein, L.; Mercaldo, N.; Müller, H.; Orlhac, F.; dos Santos, D.P.; et al. CheckList for EvaluAtion of Radiomics research (CLEAR): A step-by-step reporting guideline for authors and reviewers endorsed by ESR and EuSoMII. Insights Imaging 2023, 14, 75. [Google Scholar] [CrossRef] [PubMed]
  111. Kocak, B.; D’Antonoli, T.A.; Mercaldo, N.; Alberich-Bayarri, A.; Baessler, B.; Ambrosini, I.; Andreychenko, A.E.; Bakas, S.; Beets-Tan, R.G.H.; Bressem, K.; et al. METhodological RadiomICs Score (METRICS): A quality scoring tool for radiomics research endorsed by EuSoMII. Insights Imaging 2024, 15, 8. [Google Scholar] [CrossRef] [PubMed]
Figure 1. PRISMA 2020 flow diagram illustrating the study identification, screening, and inclusion process. A total of 21,766 records were retrieved from databases (PubMed, Scopus, IEEE Xplore), and 1866 duplicates were removed prior to screening. After title and abstract screening (19,990 records), 187 reports were sought for full-text retrieval, of which 26 could not be accessed. Among the 161 reports assessed for eligibility, 59 were excluded for the following reasons: editorial (n = 5), extended abstract (n = 28), and out of topic (n = 23). Finally, 105 studies were included in the qualitative synthesis [31].
Figure 1. PRISMA 2020 flow diagram illustrating the study identification, screening, and inclusion process. A total of 21,766 records were retrieved from databases (PubMed, Scopus, IEEE Xplore), and 1866 duplicates were removed prior to screening. After title and abstract screening (19,990 records), 187 reports were sought for full-text retrieval, of which 26 could not be accessed. Among the 161 reports assessed for eligibility, 59 were excluded for the following reasons: editorial (n = 5), extended abstract (n = 28), and out of topic (n = 23). Finally, 105 studies were included in the qualitative synthesis [31].
Sensors 26 01039 g001
Figure 2. (a) Distribution of scientific articles containing the word “radiomics” divided by year of publication [36]. (b) Histogram of the most frequent words in radiomics literature. The asterisk groups words with the same root (e.g., Imag* includes image, images, imaging).
Figure 2. (a) Distribution of scientific articles containing the word “radiomics” divided by year of publication [36]. (b) Histogram of the most frequent words in radiomics literature. The asterisk groups words with the same root (e.g., Imag* includes image, images, imaging).
Sensors 26 01039 g002
Figure 3. Infographics related to medical imaging over a three-year period from 1 January 2012, to 31 December 2014, according to [37]. (a) Pie chart representing the distribution of 1,219,876 total exams across different scanning modalities. (b) Pie chart representing a subset of the data from panel (a), showing the distribution of exams for which an imaging defect was identified. (c) Bar chart detailing the number of exams with quality concerns, by type of concern. The concerns are quantified as percentages of the total number of quality concerns observed.
Figure 3. Infographics related to medical imaging over a three-year period from 1 January 2012, to 31 December 2014, according to [37]. (a) Pie chart representing the distribution of 1,219,876 total exams across different scanning modalities. (b) Pie chart representing a subset of the data from panel (a), showing the distribution of exams for which an imaging defect was identified. (c) Bar chart detailing the number of exams with quality concerns, by type of concern. The concerns are quantified as percentages of the total number of quality concerns observed.
Sensors 26 01039 g003
Figure 4. Motion artifacts: (a) original image; (b) motion artifact due to translational motion; and (c) motion artifact due to rotational motion [47].
Figure 4. Motion artifacts: (a) original image; (b) motion artifact due to translational motion; and (c) motion artifact due to rotational motion [47].
Sensors 26 01039 g004
Figure 5. Imaging artifacts: (a) streak artifact due to embolization coils [49]; and (b) axial view of a skull phantom affected by cupping artifacts [50].
Figure 5. Imaging artifacts: (a) streak artifact due to embolization coils [49]; and (b) axial view of a skull phantom affected by cupping artifacts [50].
Sensors 26 01039 g005
Figure 6. (a) MRI of the human brain corrupted by noise [63]; (b) streak artifacts generated by large body habitus [64].
Figure 6. (a) MRI of the human brain corrupted by noise [63]; (b) streak artifacts generated by large body habitus [64].
Sensors 26 01039 g006
Table 1. Typical concerns arising during image acquisition and formation, categorized by imaging modality. A check mark indicates that the modality is typically affected by the artifact.
Table 1. Typical concerns arising during image acquisition and formation, categorized by imaging modality. A check mark indicates that the modality is typically affected by the artifact.
ConcernCTPETMRIX-RaysUltrasoundTypical Source
MotionPatient movement, long acquisition times
Beam hardening Polychromatic X-ray beams, high-density materials
NoiseLow dose, acquisition settings, patient size
BlurMotion, system resolution limits, reconstruction effects
Patient size Photon attenuation, scattering, reduced signal-to-noise ratio
Table 2. Radiomic feature classes affected by image artifacts.
Table 2. Radiomic feature classes affected by image artifacts.
Artifact TypeImaging ModalityAffected FeatureReported Impact
MotionMRI, PET, CTTexture, shapeReduced reproducibility, segmentation errors
Beam hardeningCTFirst-order, textureFeature bias, intensity distortion
NoiseCT, MRI, PETTexture (GLCM, GLSZM), first-orderIncreased variance, instability
BlurCT, MRI, PETTexture, shapeLoss of edge definition, misclassification
Body habitusCT, PETFirst-order, textureContrast loss, increased noise
Table 3. Mitigation strategies for image artifacts in radiomics.
Table 3. Mitigation strategies for image artifacts in radiomics.
Artifact TypeMitigation ApproachLevel of ApplicationLimitations
MotionMotion correction, AI-based methodsImage-levelComputational cost, modality dependence
Beam hardeningDual-energy CT, iterative reconstruction, deep learningAcquisition/reconstructionHardware availability, training data
NoiseFiltering, GAN-based denoisingImage-levelPossible alteration of texture features
BlurDeconvolution, sharpness enhancementImage-levelRequires PSF estimation
Scanner/protocol variabilityStatistical harmonizationFeature-levelRequires batch definition, possible signal masking
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Felicetti, F.; Lamonaca, F.; Carnì, D.L.; Costanzo, S. Image Quality Standardization in Radiomics: A Systematic Review of Artifacts, Variability, and Feature Stability. Sensors 2026, 26, 1039. https://doi.org/10.3390/s26031039

AMA Style

Felicetti F, Lamonaca F, Carnì DL, Costanzo S. Image Quality Standardization in Radiomics: A Systematic Review of Artifacts, Variability, and Feature Stability. Sensors. 2026; 26(3):1039. https://doi.org/10.3390/s26031039

Chicago/Turabian Style

Felicetti, Francesco, Francesco Lamonaca, Domenico Luca Carnì, and Sandra Costanzo. 2026. "Image Quality Standardization in Radiomics: A Systematic Review of Artifacts, Variability, and Feature Stability" Sensors 26, no. 3: 1039. https://doi.org/10.3390/s26031039

APA Style

Felicetti, F., Lamonaca, F., Carnì, D. L., & Costanzo, S. (2026). Image Quality Standardization in Radiomics: A Systematic Review of Artifacts, Variability, and Feature Stability. Sensors, 26(3), 1039. https://doi.org/10.3390/s26031039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop