Next Article in Journal
A High-Performance Branch Control Mechanism for GPGPU Based on RISC-V Architecture
Next Article in Special Issue
Machine Learning-Based Blood Pressure Prediction Using Cardiovascular Disease Data: A Comprehensive Comparative Study
Previous Article in Journal
Enhancing IoT Common Service Functions with Blockchain: From Analysis to Standards-Based Prototype Implementation
Previous Article in Special Issue
A SimAM-Enhanced Multi-Resolution CNN with BiGRU for EEG Emotion Recognition: 4D-MRSimNet
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DnCNN-Based Denoising Model for Low-Dose Myocardial CT Perfusion Imaging

1
Computer Science Department, Western University, London, ON N6A 3K7, Canada
2
Imaging Program, Lawson Research Institute, London, ON N6A 4V2, Canada
3
Department of Medical Biophysics, Western University, London, ON N6A 3K7, Canada
*
Author to whom correspondence should be addressed.
Electronics 2026, 15(1), 124; https://doi.org/10.3390/electronics15010124
Submission received: 25 November 2025 / Revised: 19 December 2025 / Accepted: 23 December 2025 / Published: 26 December 2025

Abstract

Unlike high-dose scans, low-dose cardiac CT perfusion imaging reduces patient radiation exposure and thereby the risk of potential health effects. However, it introduces significant image noise, degrading diagnostic quality and limiting clinical assessment. Denoising is thus a critical preprocessing step to enhance image quality without compromising anatomical or perfusion details. Traditionally used reconstruction-domain methods, such as Iterative Reconstruction and Compressed Sensing, are often limited by algorithmic complexity, dependence on raw sinogram data, and restricted adaptability. Conversely, image-domain methods offer more adaptable denoising options. Recently, learning-based approaches have further expanded this flexibility and demonstrated state-of-the-art performance across various denoising tasks. In this work, we present a deep learning-based denoising method specifically tuned for low-dose cardiac CT perfusion imaging. Our model is trained to reduce noise while preserving structural integrity and temporal contrast dynamics, which are critical for downstream analysis. Unlike many existing methods, our approach is optimized for perfusion data, where temporal consistency is essential. Residual cardiac motion remains a separate challenge, which we aim to address in our future work. Experimental results show significant improvements in quantitative image quality, using both reference-based and no-reference metrics, such as MSE/PSNR/SSIM and NIQE/FID/KID, as well as improved accuracy of perfusion measurements.

1. Introduction

Myocardial computed tomography perfusion (CTP) imaging plays a critical role in the evaluation of coronary artery disease by providing quantitative information on myocardial blood flow [1,2,3]. However, acquiring high-quality CTP images often requires repeated scanning across multiple cardiac cycles, resulting in substantial patient radiation exposure [4]. In clinical practice, reducing the current or voltage of the X-ray tube is a common strategy to reduce the radiation dose. However, this increases the noise level in the reconstructed images [5]. This increased noise not only degrades visual quality but also negatively affects quantitative perfusion measurements, potentially leading to diagnostic inaccuracies [6].
Traditional denoising approaches for low-dose CT, including linear filtering [7], nonlinear diffusion [8,9], and iterative reconstruction algorithms [10], have been widely used to suppress noise in low-dose acquisitions. Although these methods can effectively reduce noise, they often blur fine anatomical and structural details and introduce reconstruction artifacts. Such issues are critical in myocardial CTP imaging, where preserving perfusion details is essential. An alternative approach, compressed sensing [11], has achieved popularity by utilizing sparsity constraints in transform domains to reconstruct high-quality images from severely undersampled or low-dose datasets. Although compressed sensing can improve noise suppression while preserving edges, it is computationally expensive, sensitive to parameter tuning, and may introduce incoherent artifacts if the sparsity assumptions are incorrect. More advanced statistical model-based iterative reconstruction methods, such as Adaptive Statistical Iterative Reconstruction (ASIR) and Advanced Modelled Iterative Reconstruction (ADMIRE), can achieve improved noise-resolution trade-offs [12,13]. However, they require proprietary scanner software, involve high computational costs, and may still produce texture inconsistencies compared to standard-dose images [14].
In recent years, deep learning-based image denoising has been popular as a powerful alternative, utilizing convolutional neural networks (CNNs) to learn a direct mapping from noisy to high-quality images [15,16]. Supervised training with paired low-dose and standard-dose data has demonstrated substantial improvements in noise reduction while preserving anatomical details in CT, MRI, and PET modalities [17,18]. Several architectures, such as DnCNN [15], UNet [19,20], and GAN-based networks [21], have been adapted for specific anatomical applications and have reported superior performance compared to traditional techniques. However, most studies in this domain focus on non-medical natural images or pay little attention to general CT or thoracic imaging, with limited work addressing the unique challenges of low-dose myocardial CTP, including temporal consistency requirements, dynamic contrast enhancement, and preservation of small anatomical structures [22].
To address these challenges, we present a DnCNN-inspired denoising framework specifically tuned for myocardial CTP imaging. By learning from domain-specific data, the network is optimized to suppress noise while preserving delicate myocardial structures and safeguarding perfusion-related intensity patterns. The proposed model incorporates a variance-stabilizing transformation to unify the mixed Poisson–Gaussian noise distribution, followed by its corresponding inverse transform after reconstruction to preserve intensity fidelity. To enhance denoising efficiency, the input volumes were spatially downsampled, and multi-channel feature maps were generated to match the architecture requirements of FFDNet-style processing. The model is trained on synthetically generated low-dose images derived from high-dose porcine datasets, enabling controlled noise modelling and the availability of ground truth. To our knowledge, this is the first controlled low-dose simulation and denoising pipeline designed explicitly for dynamic myocardial CTP, evaluated using both reference-based and no-reference IQA metrics. Quantitative evaluation using both reference-based metrics (MSE, PSNR, SSIM) and no-reference image quality measures (NIQE, FID, KID) demonstrates that the proposed approach achieves superior noise suppression with minimal loss of diagnostic detail, providing a potential pathway toward safer low-dose myocardial CTP imaging protocol without compromising clinical utility.

2. Related Work

Iterative reconstruction (IR) has served as a fundamental technique in CT reconstruction, providing superior noise reduction compared to conventional filtered back projection (FBP). Unlike FBP, which applies a single-pass analytical reconstruction, IR repeatedly refines the image estimate using projection data to minimize reconstruction error. This iterative process reduces noise, improves contrast, and enhances overall image quality [23]. Over the years, IR has evolved into several well-known variants, including statistical IR, adaptive statistical IR, model-based IR and hybrid approaches [24,25,26]. These methods aim to balance image quality and computational efficiency, producing diagnostic-quality images at lower radiation doses. However, IR is constrained by two key factors: high computational cost and reliance on unprocessed projection data or sinogram.
Despite its effectiveness, iterative reconstruction faces practical challenges in ultra-low-dose CT settings. At extremely low photon counts, noise characteristics change significantly, reducing the ability of IR algorithms to maintain structural detail [27]. The statistical models underlying IR can fail to converge optimally when photon statistics are poor, leading to image artifacts and degraded diagnostic performance. Furthermore, IR methods depend on sinogram data, which is limited in availability and computationally expensive. These constraints have encouraged researchers to explore alternative strategies that can simultaneously suppress noise and preserve structure more efficiently and adaptively. Following IR, compressed sensing (CS) emerged as a significant advancement in CT reconstruction. CS exploits sparsity priors to enable accurate image recovery from undersampled projection data, reducing radiation dose and acquisition requirements. However, CS reconstructions are limited by high computational cost and dependence on careful parameter tuning, which have constrained their widespread clinical adoption [28].
The shortcomings of iterative reconstruction have influenced the development of learning-based denoising techniques. Advances in machine learning have opened a new direction for CT denoising. Learning-based methods leverage large datasets to learn a mapping from noisy to clean images, enabling networks to distinguish structured anatomical signals from unwanted noise [29]. Unlike fixed analytical filters, deep learning models adapt their processing to the specific characteristics of the input data. Approaches span both the reconstruction domain, where models process raw projection data [30], and the image domain, where denoising is applied to reconstructed images [21]. This adaptability has allowed deep learning-based methods to outperform traditional approaches in many benchmark studies.
Reconstruction domain denoising holds promise for tackling noise at its source, directly in the sinogram or projection space. By operating before image formation, these methods can suppress noise without introducing certain artifacts that arise post-reconstruction. Examples include deep networks that predict and subtract noise from sinograms or reconstruct cleaner images through learned forward and inverse transforms [31]. However, applying deep learning in this domain faces significant barriers. Access to raw projection data is typically restricted to CT scanner manufacturers due to proprietary formats and data privacy concerns. This limits the availability of large-scale datasets required for practical training and has slowed broader adoption compared to image-domain techniques, where reconstructed images are more readily accessible.
Several notable deep learning approaches have targeted the reconstruction domain. Yin et al. [31] employed a ResNet-based architecture across three stages: sinogram, FBP, and image domains, progressively removing noise at each stage. Ernest et al. [32] used a UNet model to directly denoise sinogram data, reporting measurable improvements in PSNR and SSIM. Ramon et al. [33] adopted a 3D convolutional autoencoder to predict high-dose myocardial CTP images from low-dose inputs. Yuan et al. [30] proposed a UNet approach that interpolates sparse sinograms before reconstruction to improve fidelity, while these methods demonstrate strong technical results, their reliance on projection data has limited scalability and clinical translation, especially in multi-center or multi-vendor settings.
Image-domain denoising has therefore become the most widely studied and implemented approach for low-dose CT (LDCT) imaging, primarily because reconstructed CT images are readily available for both training and evaluation. By operating directly on pixel or voxel intensities, these methods integrate easily with existing clinical workflows. They can be applied retrospectively to archived datasets or in real time during routine practice. Early convolutional neural network (CNN) methods, such as the patch-based approach proposed by Chen et al. [16], demonstrated measurable improvements in PSNR and SSIM compared to classical algorithms like BM3D and K-SVD, but often failed to preserve delicate anatomical structures. This motivated enhancements such as residual and deconvolution layers [18], which offered better structural recovery but still faced limitations when relying solely on pixel-wise loss functions.
To address these shortcomings, subsequent research focused on structural and edge preservation. For example, You et al. [34] proposed a 3D Structurally Sensitive Loss (SSL) combining SSIM, adversarial, and L 1 / L 2 terms, while Gholizadeh et al. [35] leveraged dilated convolutions to capture broader contextual information while preserving edge details. ResNet-based models with fused attention mechanisms [36] integrate spatial and channel attention to enhance detail preservation while maintaining computational efficiency. UNet variants leverage hierarchical feature extraction and skip connections for stable reconstruction performance. GAN-based approaches [21,37] achieve superior perceptual quality and texture reproduction, though often at the cost of modest trade-offs in quantitative metrics such as PSNR. These innovations improved perceptual quality but sometimes showed variable performance across datasets, highlighting the domain-dependence of learned models. Other approaches, like Kadimesetty et al. [38], applied CNN-based denoising directly to perfusion map reconstruction, bypassing the need for separate preprocessing, while this streamlined method eliminates intermediate denoising steps, it does not guarantee alignment with high-dose reference standards when manual validation is necessary.
In recent years, advanced architectures have extended the capabilities of image-domain LDCT denoising. The DnCNN model [15], adapted for CT imaging, has achieved competitive performance by modelling residual noise rather than directly estimating the clean signal. However, in the context of myocardial CTP imaging, DnCNN presents two limitations. First, to the best of our knowledge, it has not yet been explored for myocardial CTP, with prior applications focusing primarily on low-dose abdominal and lung CT [39,40]. Second, its original formulation is tailored to Gaussian noise only, whereas LDCT data typically contain a blend of Poisson and Gaussian noise, with Poisson noise, arising from the stochastic nature of X-ray photon emission [41,42], being dominant and signal-dependent, and Gaussian noise primarily from detector electronics, being signal-independent. This mismatch makes the default DnCNN suboptimal for realistic LDCT scenarios. FFDNet [43], an extension of DnCNN by the same group of authors, introduces noise level maps to handle more general noise, providing a potential pathway for signal-dependent Poisson noise. Yet, its efficacy on Poisson-dominated LDCT data remains untested. To bridge this gap while leveraging the power of the learning models, we propose a modified DnCNN designed to learn and remove blended Gaussian–Poisson noise patterns in low-dose cardiac CTP.

3. Proposed Method

3.1. Model Summary

We propose a deep learning framework inspired by DnCNN-FFDNet for denoising low-dose CT images corrupted by blended Poisson–Gaussian noise. Our proposed model works in several stages as described below.
Stage 1: Simulating Poisson–Gaussian Noise. In the first stage, we generate the simulated low-dose CT image from high-dose CT images. We also generate a noise map based on the simulated noise to feed into the FFDNet model. Equation (1) illustrates how simulated noise is generated. Equation (2) defines the process for generating the noise map.
I LD ( x , y ) = Poisson I HD ( x , y ) k + N 0 , σ Gaussian , k = Dose reduction factor
Stage 2: Variance-Stabilizing Transform (VST) Preprocessing. The second stage employs a Generalized Anscombe Transform (GAT) [44] to convert signal-dependent Poisson noise into approximately signal-independent Gaussian noise, stabilizing the noise variance. This transformation is essential because it prepares the data for CNN-based Gaussian denoising, as models like FFDNet/DnCNN perform optimally with Gaussian noise. By preprocessing the input data with GAT, we create a noise-level map compatible with FFDNet, thereby enhancing denoising performance and ensuring more effective training [45]. The generation of the noise level map is defined in Equation (2).
Stage 3: Noise-Adaptive CNN Denoising. The third stage employs a DnCNN-inspired model, with added improvements from FFDNet, enhanced with dilated convolutions for multi-scale feature extraction and residual learning (DnCNN-style). The network takes the simulated low-dose image and the noise level map as input, enabling adaptive denoising. The residual learning strategy of DnCNN predicts the noise component, which is subtracted from the input to produce the denoised output in the VST domain. After denoising, the Inverse Generalized Anscombe Transform (iGAT) reconstructs the final denoised CT image in the original intensity domain, with optional bias correction for improved accuracy. Figure 1 illustrates the schematic diagram of the proposed denoising architecture.
σ map ( x , y ) = λ ( x , y ) + σ Gaussian 2 where λ ( x , y ) = I HD ( x , y ) k

3.2. Data Description

Data Acquisition: Our study utilized porcine cardiac CT perfusion (CTP) datasets acquired at three tube-current settings: high-dose (80 mA), mid-dose (40 mA), and low-dose (20 mA). Dose level classification depends on subject size, exposure duration, and tube current. For example, a 20 mA scan with a 0.35 s gantry rotation, 80 kilovolts (kV) and 20 perfusion acquisitions corresponds to a practical exposure of 20 × 0.35 s × 20. While 40 mA is typically considered a low dose in humans, the same current may be mid- or high-dose in animal studies.
Data Splitting: For training, we have 4 myocardial CTP studies with 2240 images each, a total of 8960 images to train the model. For testing, we have 8 myocardial CTP studies available, which were scanned with different tube current settings: 80 mA for high-dose radiation and 20 mA for low-dose radiation. Table 1 shows the study names and our given aliases to those studies, along with their scan dose.

3.3. Training Procedure

Noise Simulation: To create realistic low-dose training pairs, we formed three dose combinations: (80 mA, 20 mA), (80 mA, 40 mA), and (40 mA, 20 mA). In each pair, the higher-dose image was designated as the reference, while the lower-dose image defined the target noise characteristics. Controlled Poisson–Gaussian noise was simulated by modelling quantum noise as a Poisson process (signal-dependent) and electronic noise as additive Gaussian noise (signal-independent). The Poisson component was scaled according to the photon flux ratio between dose levels, and the Gaussian variance was estimated from detector noise measurements. Noise was applied to the high-dose image such that the resulting synthetic image matched the noise statistics of the corresponding low-dose acquisition. This process ensured that training pairs reflected realistic mixed-noise conditions.
Variance Stabilizing Transform (VST) Preprocessing: Because the Poisson component is signal-dependent, a generalized Anscombe transform (GAT) was applied to all simulated noisy images to stabilize variance and approximate the noise distribution as additive white Gaussian noise (AWGN). This transformation facilitated the use of Gaussian-denoising networks, as their performance is tested only with Gaussian noise. Along with the transformed noisy image, a noise-level map was generated to capture spatially varying noise levels. This map was provided as an auxiliary input to the FFDNet network, enabling noise-adaptive processing. This improvement over the DnCNN network allows it to use a fixed noise level during training.
Network Training: We employed a DnCNN-inspired architecture adapted from FFDNet to leverage both residual learning and noise-level conditioning. The network consisted of sequential convolutional layers with batch normalization and ReLU activation, dilated convolutions for multi-scale feature capture, and skip connections for improved gradient flow. Residual learning was implemented such that the network predicted the noise component, which was then subtracted from the input to obtain the denoised image in the VST domain. The inverse GAT was subsequently applied to return the image to its original intensity scale. The primary loss function was mean squared error (MSE) between the predicted and ground-truth high-dose images in the VST domain. Training was conducted on NVIDIA GPU hardware. Realistic Poisson–Gaussian blended noise simulation, variance stabilization, and noise-adaptive residual learning ensured that the network was explicitly trained to address the spatially varying, mixed-noise characteristics of low-dose CT while preserving diagnostically relevant myocardial structures.
The dataset was derived from four CTP studies, each contributing 2240 images, resulting in 8960 image pairs. Of these, 7960 pairs were randomly selected for training and 1000 for validation to avoid bias toward any particular study. Each noisy image was accompanied by a noise-level map, serving as an auxiliary input to the FFDNet architecture and enabling adaptive denoising. The network operated by downsampling noisy inputs, processing them through the multi-scale architecture, and reconstructing denoised sub-images, which were then combined into a clean output, which was finally transformed back to the original intensity domain via the inverse Anscombe transform. Training was performed with a batch size of 16, corresponding to 498 iterations per epoch. A maximum of 50 epochs was allowed, but early stopping with a tolerance of 8 epochs was applied; training converged after 12 epochs when the validation loss plateaued. Optimization used Stochastic Gradient Descent (SGD) with an initial learning rate of 10 3 and momentum of 0.8. We used FFDNet’s default downsampling. The average inference latency (conducted in Core i5-4570 CPU @ 3.20 GHz (4CPUs)) was 1.2 s per image, corresponding to a throughput of approximately 0.83 frames per second.

3.4. Novelty and Contributions

The objective of this work is not to introduce a new network topology, but to adapt and specialize a proven denoising architecture for the unique challenges of dynamic low-dose myocardial CTP imaging. Although convolutional denoising networks such as DnCNN have been widely studied for static natural and medical images, their direct application to dynamic myocardial CTP is challenging due to mixed signal-dependent and independent noise characteristics, strong temporal dependencies, and the need to preserve perfusion dynamics. The novelty of this work lies in the following contributions:
  • First adaptation of DnCNN/FFDNet to dynamic myocardial CTP: To the best of our knowledge, this work presents the first systematic adaptation and evaluation of a DnCNN/FFDNet-based denoising framework for dynamic myocardial CT perfusion sequences rather than static CT images.
  • Noise-adaptive denoising pipeline tailored to myocardial CTP: We introduce a noise-aware preprocessing strategy based on the Generalized Anscombe Transform (GAT) to stabilize the combined Poisson–Gaussian noise inherent to low-dose CTP, enabling effective use of Gaussian-noise–oriented denoising models while respecting CT noise physics.
  • Controlled high-dose porcine simulation framework: A controlled porcine CTP dataset is used to generate paired high-dose and synthetically degraded low-dose sequences, enabling rigorous reference-based evaluation under realistic motion and perfusion conditions.
  • Comprehensive quantitative and perceptual evaluation: We present a systematic assessment combining reference-based metrics (MSE, PSNR, SSIM), no-reference perceptual metrics (FID, KID, NIQE) retrained for the myocardial CTP domain, temporal consistency analysis, and expert radiologist assessment. To the best of our knowledge, such retraining and application of no-reference IQAs has not previously been reported for dynamic myocardial CTP denoising.

4. Results

4.1. Reference-Based Image Quality Assessment (IQA)

Reference-based image quality metrics are a conventional means of assessing reconstructed image quality. These metrics are beneficial when a reference image is available, enabling the reconstructed image to be quantitatively compared against the reference to obtain an objective quality score. In our study, high-dose (80 mA) images were available and subsequently simulated to represent low-dose (20 mA) acquisitions. After reconstructing these low-dose images to approximate 80 mA quality, we compared them against the original high-dose (80 mA) reference images to evaluate reconstruction performance.

4.1.1. Means Squared Error (MSE)

MSE measures the average squared difference between corresponding pixels. Equation (3) defines Mean Squared Error, where X is the reference image and Y is the reconstructed image.
MSE = 1 N i = 1 N ( X i Y i ) 2
Table 2 presents the Mean Squared Error (MSE) values for all simulated test cases, comparing the simulated noisy low-dose images (20 mA) with their corresponding reconstructed outputs. The MSE values presented here are the averages across all MSEs calculated for each image in each test case. Across all test cases, the reconstructed images consistently achieve significantly lower MSE values, reducing the error by more than four times compared to the original noisy inputs. For example, in test case STC1, the MSE drops from 37.50 ± 0.87 to 7.62 ± 3.57, and in STC5, it is reduced from 35.41 ± 0.72 to 6.08 ± 2.17. This substantial and consistent reduction in MSE clearly demonstrates the effectiveness of the reconstruction process in recovering image fidelity and minimizing deviation from the high-quality reference (80 mA), thereby confirming the enhanced accuracy of the reconstructed images.

4.1.2. Peak Signal-to-Noise Ratio (PSNR)

PSNR is derived from MSE and quantifies the ratio of the maximum possible pixel value to the noise power. Equation (4) defines PSNR, where M A X I denotes the maximum possible intensity value of the image.
PSNR = 10 · log 10 M A X I 2 MSE
Table 3 presents the Peak Signal-to-Noise Ratio (PSNR) values for the simulated test cases, comparing the original noisy images acquired at a low dose (20 mA) with their corresponding reconstructed outputs. In all cases, the reconstructed images demonstrate a substantial improvement in PSNR, with increases of over 7–8 dB relative to the noisy inputs. For example, STC2 improves from 32.45 ± 0.09 to 40.59 ± 2.11 dB, and STC5 from 32.64 ± 0.09 to 40.51 ± 1.30 dB. Since PSNR is a widely used metric for quantifying image quality, where higher values indicate greater similarity to the reference image, these consistent gains confirm that the reconstruction process effectively suppresses noise and preserves structural detail, resulting in images that are significantly closer in quality to high-dose reference scans.

4.1.3. Structural Similarity Index (SSIM)

SSIM measures perceptual similarity by comparing luminance, contrast, and structure. Equation (5) defines SSIM, where μ X and μ Y are the means, σ X 2 and σ Y 2 the variances, and σ X Y the covariance between images.
SSIM ( X , Y ) = ( 2 μ X μ Y + C 1 ) ( 2 σ X Y + C 2 ) ( μ X 2 + μ Y 2 + C 1 ) ( σ X 2 + σ Y 2 + C 2 )
Table 4 presents the Structural Similarity Index (SSIM) values for the simulated test cases, comparing the noisy input images acquired at low dose (20 mA) with the corresponding reconstructed outputs. In every case, the reconstructed images exhibit a significant increase in SSIM, improving from values in the range of 0.63–0.68 (noisy) to 0.87–0.90. For instance, STC2 shows an increase from 0.63 ± 0.006 to 0.90 ± 0.002, and STC7 from 0.67 ± 0.008 to 0.89 ± 0.003. These consistently high SSIM values for the reconstructed images indicate strong structural resemblance to the reference 80 mA images. Given that SSIM is specifically designed to evaluate perceived image quality by comparing local structures, contrast, and luminance, the results clearly demonstrate that the reconstruction method effectively restores structural integrity and preserves critical anatomical details degraded in low-dose acquisitions.

4.2. No Reference Image Quality Assessments (IQA)

To further assess the proposed method’s capacity to generate high-quality HDCT from LDCT acquisitions, we employed no-reference image quality assessment (NR-IQA) metrics, also known as blind image quality metrics. This choice reflects the practical reality that, during clinical acquisition of low-dose scans, a corresponding high-dose reference is typically unavailable for direct comparison. Consequently, NR-IQA provides a viable framework to objectively evaluate model performance in scenarios where reference images are absent, thereby offering insights into the method’s applicability to real-world clinical workflows. It is important to note that blind image quality metrics rely on a global reference model, whereas reference-based metrics use a signal-dependent reference for evaluation.
Although FID/KID and NIQE were initially developed using statistical properties of natural scene images, their sensitivity to perceptual attributes such as smoothness, sharpness, and structural consistency makes them suitable for relative, no-reference image quality assessment in medical imaging. For FID and KID, a reference distribution is still required; therefore, we computed domain-specific feature statistics using a set of high-dose CT images, rather than relying on generic ImageNet-based statistics. Similarly, NIQE was re-estimated on the same high-dose dataset to create a medical-domain NIQE model, which was then applied to evaluate the reconstructed low-dose CT images.
It is worth noting that no-reference image quality metrics such as FID, KID, and NIQE should be interpreted as complementary statistical descriptors of feature-space similarity, rather than as substitutes for established reference-based image quality metrics. To our knowledge, these no-reference metrics have not previously been applied or reported in the myocardial CT perfusion literature. In this work, we therefore also aimed to demonstrate that these models can be retrained in this domain and used reliably to assess reconstructed CT perfusion images in the absence of a reference, while their behaviour correlates with reference-based IQAs in our experiments, a comprehensive study of their reliability and task-level diagnostic relevance is beyond the scope of this work.

4.2.1. FID

Fréchet Inception Distance (FID) measures the distance between the feature distributions of generated images and those of high-quality real images. Lower FID indicates better perceptual similarity. Equation (6) defines Fréchet Inception Distance (FID),
F I D = μ r μ g 2 + Tr Σ r + Σ g 2 Σ r Σ g 1 / 2
where:
-
μ r , Σ r are the mean and covariance of reference image features,
-
μ g , Σ g are the mean and covariance of generated image features,
-
Tr ( · ) denotes the trace of a matrix, i.e., the sum of the diagonal elements:
Tr ( A ) = i = 1 n A i i , where A i i denotes the diagonal elements of matrix A.
Table 5 presents the Fréchet Inception Distance (FID) scores for each test case, comparing noisy low-dose CT images with their corresponding reconstructed versions. In all cases, the reconstructed images exhibit noticeably lower FID values, indicating a closer match to the distribution of the high-dose reference images. Since lower FID values indicate greater similarity in the feature distributions of the test and reference images, these results quantitatively confirm that the reconstruction method improves image realism and perceptual fidelity.

4.2.2. KID

Kernel Inception Distance (KID) estimates the squared Maximum Mean Discrepancy (MMD) between feature representations of generated and real images using a polynomial kernel. Unlike FID, KID has an unbiased estimator and is more robust on smaller sample sizes. Lower KID indicates better similarity. Equation (7) defines Kernel Inception Distance (KID).
K I D = 1 n i = 1 n ϕ ( x i ) 1 m j = 1 m ϕ ( y j ) 2
where:
-
x i are features of real images,
-
y j are features of generated images,
-
ϕ ( · ) denotes the feature mapping induced by the polynomial kernel,
-
n and m are the number of real and generated samples, respectively.
Table 6 presents the Kernel Inception Distance (KID) mean and standard deviation values computed for each test case, comparing the noisy low-dose acquisitions with the corresponding reconstructed outputs. In every case, the reconstructed images achieve significantly lower KID scores, both in mean and variability, indicating a closer match to the reference distribution and greater perceptual similarity. These consistent improvements across all test cases confirm the effectiveness of the reconstruction process in producing images with statistical properties that better resemble those of high-dose reference scans. As KID is specifically designed to measure distribution-level similarity in the deep feature space, the results provide strong evidence of enhanced image quality and perceptual fidelity.

4.2.3. NIQE

Naturalness Image Quality Evaluator (NIQE) quantifies the statistical deviation of the image from natural scene statistics. Lower NIQE scores indicate higher perceptual quality. NIQE is defined by Equation (8).
N I Q E = ( μ n μ d ) T Σ n + Σ d 2 1 ( μ n μ d )
where:
-
μ n , Σ n are the mean and covariance of features from natural pristine images,
-
μ d , Σ d are the mean and covariance of features from the distorted image,
-
T indicates transpose, and −1 denotes matrix inverse.
Table 7 presents the NIQE (Natural Image Quality Evaluator) scores computed for each test case under evaluation. Like other quantitative parameters used in this chapter, this score was calculated per image for one data set, and the average was then taken; while NIQE was initially developed based on statistical regularities observed in natural images, its ability to assess perceptual quality such as image smoothness, sharpness, and structural regularity makes it a useful no-reference metric, even in medical imaging contexts, when used for relative comparisons. In this study, we retrained NIQE on domain-specific data and then applied it consistently across both noisy and reconstructed CT images. As presented in Table 7, the reconstructed images consistently yield lower NIQE scores compared to their noisy counterparts, indicating improved perceptual quality as perceived by the metric. When interpreted alongside other objective metrics such as PSNR, SSIM, MSE, FID, and KID; and subjective metrics such as AUC and expert ratings, NIQE further supports the conclusion that our proposed model significantly enhances image quality, particularly in terms of noise suppression and structural coherence.

4.2.4. SSIM

The Structural Similarity Index (SSIM) is traditionally defined for evaluating spatial similarity between two images as defined in Equation (5). To extend this concept to dynamic imaging, the temporal SSIM (tSSIM) measures structural consistency across the time dimension of a reconstructed CT perfusion sequence. Mathematically, for a temporal sequence { I t } t = 1 T consisting of T frames, temporal SSIM is computed as the average SSIM between consecutive time frames:
tSSIM = 1 T 1 t = 1 T 1 SSIM ( I t , I t + 1 ) ,
where SSIM ( I t , I t + 1 ) denotes the standard structural similarity index between frame I t and the subsequent frame I t + 1 .
In this work, temporal SSIM was computed by calculating SSIM between consecutive reconstructed frames along the time dimension of the CT perfusion sequence and averaging the resulting values. A higher temporal SSIM indicates improved preservation of temporal consistency, reduced flickering, and more physiologically coherent myocardial perfusion dynamics, reflecting the reconstruction method’s ability to maintain structural stability over time. The resulting tSSIM values summarized in Table 8 demonstrate consistently high temporal fidelity across all test cases, indicating that the proposed reconstruction approach effectively preserves frame-to-frame structural coherence in myocardial CT perfusion imaging.

4.3. Subjective Inspection

In addition to objective quality metrics, we provide a subjective visual example in Figure 2 illustrating the model’s effectiveness in correcting artifacts and restoring image quality degraded by simulated conditions. Visual inspection against the reference standard reveals that our method has enhanced capability to reproduce fine textures and anatomical structures.
The myocardial and pulmonary regions from Figure 2 were magnified and are presented in Figure 3. A comparative inspection of the three images clearly demonstrates that the noise present in the 20 mA acquisition (middle) is substantially reduced in the reconstructed 80 mA image (right). The myocardial structures appear well preserved, while the pulmonary vessels are more distinctly delineated, indicating effective noise suppression and structural recovery achieved by the proposed reconstruction method. This is comparable with the original 80 mA scan (right).

4.4. Comparative Performance

Although a comparative study is informative, in our case it is not an exact like-for-like evaluation because the reference studies available for comparison and listed in Table 9 were conducted on a different dataset—the 2016 NIH-AAPM-Mayo Clinic Low Dose CT Grand Challenge dataset, where quarter-dose refers to ¼th of the high-dose scan, which is specific to the Mayo acquisition protocol and cannot be directly translated to our experimental setting. Furthermore, the myocardial CTP imaging data used in our study are inherently susceptible to both noise and residual cardiac motion, unlike the only noise present in low-dose CT data used in works listed in Table 9. In our porcine myocardial CTP scans, the full dose is 80 mA, and the low dose is 20 mA, which likewise represents a quarter of the reference dose. Thus, the comparison is provided only to offer a relative performance context rather than a strict one-to-one benchmark.
Despite this limitation, the quantitative performance of our proposed method compares favourably with several recent deep learning-based denoising approaches. Yang et al. [21] proposed a GAN with Wasserstein distance and perceptual loss for LDCT denoising; Ma et al. [37] introduced a hybrid-loss GAN designed to capture low-dose noise statistics better; Zhang et al. [46] developed a highly efficient transformer architecture (Hformer) tailored for LDCT restoration; and Zhang et al. [47] presented an adaptive dual-branch CNN using multi-scale residual attention. These prior GAN, CNN, and transformer-based methods typically reported PSNR values in the range of 24 33 dB and SSIM values of 0.81 0.92 on the NIH-AAPM-Mayo dataset [21,37,46,47]. In contrast, our model achieves a mean PSNR of 39.96 dB and SSIM of 0.8906 on dynamically acquired myocardial CTP data under realistic Poisson–Gaussian noise conditions. These values indicate performance that is competitive with, or, in the majority of cases, exceeds, that of several state-of-the-art approaches, despite the added complexity of perfusion imaging, where temporal contrast changes and cardiac motion produce substantially more challenging noise characteristics than in static abdominal or thoracic CT. Please note that the results for the comparative methods were taken from their respective publications, as reported. We did not re-implement or reproduce them.

4.5. Effect of GAT

A key contribution of the proposed method is the use of the Generalized Anscombe Transform (GAT) to stabilize the signal-dependent Poisson noise present in low-dose CT perfusion data, approximating it as signal-independent Gaussian noise. This enables a more effective application of denoising models such as DnCNN/FFDNet, which were initially designed under Gaussian noise assumptions. To assess the impact of variance stabilization, we compare the baseline DnCNN model applied with and without GAT.
As shown in Table 10, applying the baseline model without GAT improves image quality relative to the noisy inputs; however, the gains remain limited due to the mismatch between the noise statistics and the model assumptions. In contrast, incorporating GAT consistently yields substantially lower MSE, higher PSNR, and improved SSIM across all test cases. These results demonstrate that variance stabilization is essential for effectively denoising low-dose myocardial CT perfusion images with Gaussian noise-oriented deep learning models.

4.6. Expert’s Scoring

A qualitative assessment was conducted by an expert radiologist to evaluate the denoising performance relative to the corresponding high-dose (80 mA) reference images. Two clinically relevant criteria were considered: (i) whether the reconstructed images provide sufficient anatomical detail to support confident diagnosis, and (ii) whether the reconstructed images can be considered diagnostically equivalent to high-dose acquisitions. Each criterion was scored on a five-point Likert scale, where 1 indicates minimal confidence and 5 indicates full confidence.
As summarized in Table 11, the reconstructed images consistently received high scores across all test cases. For diagnostic sufficiency, a uniform score of 4.0 was assigned for all STCs, indicating consistent confidence in the preservation of diagnostically relevant detail. For diagnostic equivalence to high-dose imaging, the mean score was 4.06, with most cases rated at 4.0 or higher. These results suggest that the proposed denoising approach yields reconstructions that are not only diagnostically adequate but also closely comparable to high-dose CT perfusion images, supporting its potential clinical utility.

5. Limitations

Despite the encouraging results, several limitations of this study should be acknowledged. First, the proposed method focuses on domain adaptation and noise modelling rather than architectural novelty; the backbone network builds upon the established DnCNN/FFDNet method. A full architectural ablation isolating the individual contributions of dilated convolutions and noise-level maps was therefore not conducted, as the effectiveness of these components has been well documented in prior literature. Instead, this work emphasizes their integration and validation within the specific context of dynamic myocardial CT perfusion imaging.
Second, although a controlled porcine myocardial perfusion dataset enables reliable paired low-dose and reference evaluations, its size remains limited and does not include human clinical cases. While porcine CTP closely mimics human myocardial perfusion dynamics and is commonly used in translational imaging research, future validation in larger human datasets will be necessary to establish its generalizability and clinical applicability fully.
Third, quantitative evaluation primarily relies on established reference-based metrics (PSNR, SSIM) and domain-specific-trained no-reference perceptual metrics (NIQE, FID, KID). Although confidence intervals are reported to reflect variability, formal paired statistical hypothesis testing was not performed, as the consistently non-overlapping confidence ranges across dose levels already indicate statistically meaningful improvements. Additional task-based statistical analyses may further strengthen clinical interpretation.
Finally, while temporal stability was assessed using temporal SSIM to capture frame-to-frame consistency, more advanced temporal or task-driven metrics, as well as fully spatiotemporal (2.5D or 3D) network architectures, were beyond the scope of this feasibility study and represent important directions for future work.

6. Conclusions

This study presented a deep learning-based denoising approach for low-dose myocardial CTP imaging, addressing the challenge of preserving diagnostic quality while minimizing radiation exposure. The model was trained on synthetically generated low-dose data, created by applying a blended Poisson–Gaussian noise model to high-dose (80 mA) reference scans, enabling controlled training with ground truth for evaluation.
On synthetically degraded images, the proposed network achieved substantial improvements in quantitative metrics, including MSE, PSNR, and SSIM, demonstrating its ability to suppress noise and restore fine anatomical details. Evaluation on real low-dose (20 mA) animal scans, using no-reference IQA metrics such as NIQE, FID, KID, and subjective inspection, confirmed the method’s robustness and generalizability.
These findings highlight the potential of the proposed denoising framework to approximate high-dose image quality from low-dose acquisitions, offering a pathway to safer myocardial CTP studies without compromising clinical utility. This work was not intended as a benchmarking study against all state-of-the-art denoisers, but rather as a feasibility investigation focused on domain-specific noise modelling, radiation dose reduction potential, and generalization to real data.

7. Future Direction

Future work will extend the current validation in several directions. Expert reader studies will be conducted to assess clinical acceptability, and additional testing on real low-dose data will further evaluate robustness beyond simulation-based training. Methodologically, we plan to investigate alternative architectures and loss functions to balance reconstruction fidelity and computational cost, and to adapt the framework to other CT applications. In addition, future extensions will explore spatiotemporal network designs, including 2.5D and 3D temporal architectures. Such architectures would enable modelling the unique need of spatiotemporal consistency in the context of myocardial CTP. Finally, while we explored the availability of pretrained U-Net- and transformer-based models (e.g., HFormer), no publicly available models were found that are directly applicable to dynamic low-dose myocardial CTP. Therefore, we reported only the numbers available in the literature, which limits the comparison between existing models and the proposed method. Adapting and retraining existing implementations under consistent data and noise assumptions is therefore deferred to future work.

Author Contributions

Conceptualization, M.H., A.S. and M.R.E.-S.; methodology and software/computer program, M.H.; validation, M.R.E.-S., A.S. and M.H.; investigation, M.H.; resources, M.H., M.R.E.-S. and A.S.; data acquisition, A.S.; writing—original draft preparation, M.H.; writing—review and editing, M.R.E.-S. and A.S.; supervision, M.R.E.-S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Project identification code (Protocol Approval Number): #2009-092; Date of approval: 2009; Name of the ethics committee/institutional review board: The University of Western Ontario, Animal Use Subcommittee/University Council on Animal Care.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data is not public domain data and not available for sharing.

Acknowledgments

The authors would like to thank Tamanna Zahan, Department of Radiology and Imaging, Holy Family Red Crescent Medical College & Hospital, Bangladesh, for her valuable contribution as a subject matter expert in the qualitative evaluation and scoring of the reconstructed CT perfusion images.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AWGNAdditive White Gaussian Noise
BM3DBlock Matching and 3D Filtering
CNNConvolutional Neural Network
CTPComputed Tomography Perfusion
FBPFiltered Back Projection
FIDFréchet Inception Distance
GATGeneralized Anscombe Transform
GANGenerative Adversarial Network
HDCTHigh-Dose Computed Tomography
IQAImage Quality Assessments
IRIterative Reconstruction
KIDKernel Inception Distance
K-SVDK-Singular Value Decomposition
LDCTLow-Dose Computed Tomography
MSEMean Square Error
NIQENaturalness Image Quality Evaluator
PSNRPeak Signal to Noise Ratio
SSIMStructural Similarity Index
SSLStructural Sensitive Loss
VSTVariance-Stabilizing Transform

References

  1. Nous, F.M.A.; Geisler, T.; Kruk, M.B.; Alkadhi, H.; Kitagawa, K.; Vliegenthart, R.; Hell, M.M.; Hausleiter, J.; Nguyen, P.K.; Budde, R.P.J.; et al. Dynamic Myocardial Perfusion CT for the Detection of Hemodynamically Significant Coronary Artery Disease. JACC Cardiovasc. Imaging 2022, 15, 75–87. [Google Scholar] [CrossRef]
  2. Sliwicka, O.; Sechopoulos, I.; Baggiano, A.; Pontone, G.; Nijveldt, R.; Habets, J. Dynamic Myocardial CT Perfusion Imaging—State of the Art. Eur. Radiol. 2023, 33, 5509–5525. [Google Scholar] [CrossRef]
  3. Fogante, M.; Paolini, E.; Cela, F.; Esposto Pirani, P.; Balardi, L.; Perna, G.P.; Schicchi, N. Feasibility, Added Value, and Radiation Dose of Combined Coronary CT Angiography and Stress Dynamic CT Myocardial Perfusion Imaging in Moderate Coronary Artery Disease: A Real-World Study. J. Cardiovasc. Dev. Dis. 2025, 12, 241. [Google Scholar] [CrossRef] [PubMed]
  4. Narula, J.; Chandrashekhar, Y.; Ahmadi, A.; Abbara, S.; Berman, D.S.; Blankstein, R.; Leipsic, J.; Newby, D.; Nicol, E.D.; Nieman, K.; et al. SCCT 2021 Expert Consensus Document on Coronary CT Angiography: A Report of the Society of Cardiovascular Computed Tomography. J. Cardiovasc. Comput. Tomogr. 2021, 15, 192–217. [Google Scholar] [CrossRef]
  5. Koetzier, L.R.; Mastrodicasa, D.; Szczykutowicz, T.P.; Willemink, M.J.; Noël, P.B. Deep Learning Image Reconstruction for CT: Technical Principles and Clinical Prospects. Radiology 2023, 306, e221257. [Google Scholar] [CrossRef]
  6. Zdanowicz, A.; Guziński, M.; Puła, M.; Witkowska, A.; Reczuch, K. Dynamic CT Myocardial Perfusion: The Role of Functional Evaluation in the Diagnosis of Coronary Artery Disease. J. Clin. Med. 2023, 12, 7062. [Google Scholar] [CrossRef] [PubMed]
  7. Nagayama, Y.; Sakabe, D.; Goto, M.; Emoto, T.; Oda, S.; Nakaura, T.; Kidoh, M.; Uetani, H.; Funama, Y.; Hirai, T. Deep Learning-based Reconstruction for Lower-Dose Pediatric CT: Technical Principles, Image Characteristics, and Clinical Implementations. RadioGraphics 2021, 41, 1936–1953. [Google Scholar] [CrossRef] [PubMed]
  8. Benz, D.C.; Ersözlü, S.; Mojon, F.L.; Messerli, M.; Mitulla, A.K.; Ciancone, D.; Kenkel, D.; Schaab, J.A.; Gebhard, C.; Pazhenkottil, A.P.; et al. Radiation Dose Reduction with Deep-Learning Image Reconstruction for Coronary Computed Tomography Angiography. Eur. Radiol. 2022, 32, 2620–2628. [Google Scholar] [CrossRef]
  9. Kamilov, U.S.; Bouman, C.A.; Buzzard, G.T.; Wohlberg, B. Plug-and-Play Methods for Integrating Physical and Learned Models in Computational Imaging: Theory, algorithms, and applications. IEEE Signal Process. Mag. 2023, 40, 85–97. [Google Scholar] [CrossRef]
  10. Greffier, J.; Hamard, A.; Pereira, F.; Barrau, C.; Pasquier, H.; Beregi, J.-P.; Frandon, J. Effect of a New Deep Learning Image Reconstruction Algorithm (AiCE) for Abdominal CT on Image Quality and Dose Reduction: A Phantom Study. Quant. Imaging Med. Surg. 2022, 12, 229–243. [Google Scholar] [CrossRef]
  11. Li, Z.; Sun, Z.; Lv, L.; Liu, Y.; Wang, X.; Xu, J.; Xing, J.; Babyn, P.; Sun, F. Ultra-sparse view lung CT image reconstruction using generative adversarial networks and compressed sensing. J. X-Ray Sci. Technol. 2025, 33, 803–816. [Google Scholar] [CrossRef]
  12. Jensen, C.T.; Gupta, S.; Saleh, M.M.; Liu, X.; Wong, V.K.; Salem, U.; Qiao, W.; Samei, E.; Wagner-Bartak, N.-A. Reduced-Dose Deep Learning Reconstruction for Abdominal CT of Liver Metastases. Radiology 2022, 303, 620–631. [Google Scholar] [CrossRef]
  13. Greffier, J.; Hamard, A.; Pereira, F.; Barrau, C.; Pasquier, H.; Beregi, J.P.; Frandon, J. Image Quality and Dose Reduction Opportunity of Deep Learning Image Reconstruction Algorithm for CT: A Phantom Study. Eur. Radiol. 2020, 30, 3951–3959. [Google Scholar] [CrossRef]
  14. Greffier, J.; Frandon, J.; Larbi, A.; Beregi, J.-P.; Hamard, A.; Pereira, F.; Pasquier, H. CT Iterative Reconstruction Algorithms: A Task-Based Image Quality Assessment. Eur. Radiol. 2020, 30, 487–500. [Google Scholar] [CrossRef] [PubMed]
  15. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef]
  16. Chen, H.; Zhang, Y.; Zhang, W.; Liao, P.; Li, K.; Zhou, J.; Wang, G. Low-Dose CT via Convolutional Neural Network. Biomed. Opt. Express 2017, 8, 679–694. [Google Scholar] [CrossRef] [PubMed]
  17. Kang, E.; Min, J.; Ye, J.C. A Deep Convolutional Neural Network Using Directional Wavelets for Low-Dose X-Ray CT Reconstruction. Med. Phys. 2017, 44, e360–e375. [Google Scholar] [CrossRef]
  18. Chen, H.; Zhang, Y.; Kalra, M.K.; Lin, F.; Chen, Y.; Liao, P.; Wang, G. Low-Dose CT with a Residual Encoder-Decoder Convolutional Neural Network. IEEE Trans. Med. Imaging 2017, 36, 2524–2535. [Google Scholar] [CrossRef] [PubMed]
  19. Wang, J.; Raihana, N.I.; Fu, P. A Comprehensive Review of U-Net and Its Variants: Advances and Applications in Medical Image Segmentation. IET Image Process. 2025, 19, 1751–9659. [Google Scholar] [CrossRef]
  20. Isensee, F.; Jaeger, P.F.; Kohl, S.A.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A Self-Configuring Method for Deep Learning–Based Biomedical Image Segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef]
  21. Yang, Q.; Yan, P.; Zhang, Y.; Yu, H.; Shi, Y.; Mou, X.; Kalra, M.K.; Zhang, L.; Sun, L.; Wang, G. Low-Dose CT Image Denoising Using a Generative Adversarial Network with Wasserstein Distance and Perceptual Loss. IEEE Trans. Med. Imaging 2018, 37, 1348–1357. [Google Scholar] [CrossRef] [PubMed]
  22. Patel, A.R.; Bamberg, F.; Branch, K.R.; Carrascosa, P.; De Cecco, C.N.; Einstein, A.J.; Fuld, M.K.; George, R.T.; Grizzard, J.D.; Han, B.K.; et al. Society of cardiovascular computed tomography expert consensus document on myocardial computed tomography perfusion imaging. J. Cardiovasc. Comput. Tomogr. 2020, 14, 87–100. [Google Scholar] [CrossRef]
  23. Zhu, S.; Zhang, B.; Tian, Q.; Li, A.; Liu, Z.; Hou, W.; Zhao, W.; Huang, X.; Xiao, Y.; Wang, Y.; et al. Reduced-Dose Deep Learning Iterative Reconstruction for Abdominal CT with Low Tube Voltage and Tube Current: A Prospective Study. BMC Med. Inform. Decis. Mak. 2024, 24, 389. [Google Scholar] [CrossRef]
  24. Kang, H.J.; Lee, J.M.; Park, S.J.; Lee, S.M.; Joo, I.; Yoon, J.H. Image quality improvement of low-dose abdominal CT using deep learning image reconstruction compared with the Second generation iterative reconstruction. Curr. Med. Imaging 2024, 20, e250523217310. [Google Scholar] [CrossRef]
  25. De Santis, D.; Polidori, T.; Tremamunno, G.; Rucci, C.; Piccinni, G.; Zerunian, M.; Pugliese, M.; Del Gaudio, A.; Guido, G.; Barbato, L.; et al. Deep Learning Image Reconstruction Algorithm: Impact on Image Quality in Coronary CT Angiography. La Radiol. Medica 2023, 128, 434–444. [Google Scholar] [CrossRef] [PubMed]
  26. Laurent, G.; Villani, N.; Hossu, G.; Rauch, A.; Noël, A.; Blum, A.; Gondim Teixeira, P.A. Full Model-Based Iterative Reconstruction (MBIR) in Abdominal CT Increases Objective Image Quality, but Decreases Subjective Acceptance. Eur. Radiol. 2019, 29, 4016–4025. [Google Scholar] [CrossRef]
  27. Dudhe, S.S.; Mishra, G.; Parihar, P.; Nimodia, D.; Kumari, A.; Mishra, G.V. Radiation dose optimization in radiology: A comprehensive review of safeguarding patients and preserving image fidelity. Cureus 2024, 16, e60846. [Google Scholar] [CrossRef]
  28. Szczykutowicz, T.P.; Toia, G.V.; Dhanantwari, A.; Nett, B. A Review of Deep Learning CT Reconstruction: Concepts, Limitations, and Promise in Clinical Practice. Curr. Radiol. Rep. 2022, 10, 101–115. [Google Scholar] [CrossRef]
  29. Willemink, M.J.; Noël, M.P. The Evolution of Image Reconstruction for CT—From Filtered Back Projection to Artificial Intelligence. Eur. Radiol. 2019, 29, 2185–2195. [Google Scholar] [CrossRef]
  30. Yuan, H.; Jia, J.; Zhu, Z. SIPID: A deep learning framework for sinogram interpolation and image denoising in low-dose CT reconstruction. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 1521–1524. [Google Scholar] [CrossRef]
  31. Yin, X.; Yu, L.; Zhou, Z.; Xie, T.; Wang, J.; Chen, W.; Zhou, Z. Domain Progressive 3D Residual Convolution Network to Improve Low-Dose CT Imaging. IEEE Trans. Med. Imaging 2019, 38, 2903–2913. [Google Scholar] [CrossRef] [PubMed]
  32. Ernst, P.; Chatterjee, S.; Rose, G.; Speck, O.; Nürnberger, A. Sinogram upsampling using Primal–Dual UNet for undersampled CT and radial MRI reconstruction. Neural Netw. 2023, 166, 704–721. [Google Scholar] [CrossRef]
  33. Ramon, A.J.; Yang, Y.; Pretorius, P.H.; Johnson, K.L.; King, M.A.; Wernick, M.N. Improving diagnostic accuracy in low-dose SPECT myocardial perfusion imaging with convolutional denoising networks. IEEE Trans. Med. Imaging 2020, 39, 2893–2903. [Google Scholar] [CrossRef] [PubMed]
  34. You, C.; Yang, Q.; Gao, Y.; Mou, X.; Wang, G. Structurally-Sensitive Multi-Scale Deep Neural Network for Low-Dose CT Denoising. IEEE Access 2018, 6, 41839–41855. [Google Scholar] [CrossRef] [PubMed]
  35. Gholizadeh-Ansari, M.; Alirezaie, J.; Babyn, P. Deep learning for low-dose CT denoising using perceptual loss and edge detection layer. J. Digit. Imaging 2020, 33, 504–515. [Google Scholar] [CrossRef] [PubMed]
  36. Marcos, L.; Alirezaie, J.; Babyn, P. Low-Dose CT Denoising by ResNet with Fused Attention Modules and Integrated Loss Functions. Front. Signal Process. 2022, 1, 812193. [Google Scholar] [CrossRef]
  37. Ma, Y.; Wei, B.; Feng, P.; He, P.; Guo, X.; Wang, G. Low-dose CT image denoising using a generative adversarial network with a hybrid loss function for noise learning. IEEE Access 2020, 8, 67519–67529. [Google Scholar] [CrossRef]
  38. Kadimesetty, V.; Aradhya, A.K.R.; Prasad, S.S.; Raghavendra, M.S. Convolutional Neural Network-Based Robust Denoising of Low-Dose Computed Tomography Perfusion Maps. IEEE Trans. Radiat. Plasma Med. Sci. 2018, 3, 137–152. [Google Scholar] [CrossRef]
  39. Zhao, T.; McNitt-Gray, M.F.; Ruan, D. A Convolutional Neural Network for Ultra-Low-Dose CT Denoising and Emphysema Screening. Med. Phys. 2019, 46, 3941–3950. [Google Scholar] [CrossRef]
  40. Usui, K.; Ogawa, K.; Goto, M.; Sakano, Y.; Kyougoku, S.; Daida, H. Quantitative Evaluation of Deep Convolutional Neural Network-Based Image Denoising for Low-Dose Computed Tomography. Vis. Comput. Ind. Biomed. Art 2021, 4, 21. [Google Scholar] [CrossRef]
  41. Diwakar, M.; Kumar, M. A Review on CT Image Noise and Its Denoising. Biomed. Signal Process. Control 2018, 42, 73–88. [Google Scholar] [CrossRef]
  42. Park, H.S.; Baek, J.; You, S.K.; Choi, J.K.; Seo, J.K. Unpaired Image Denoising Using a Generative Adversarial Network in X-Ray CT. IEEE Access 2019, 7, 110414–110425. [Google Scholar] [CrossRef]
  43. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a Fast and Flexible Solution for CNN-Based Image Denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef]
  44. Mäkitalo, M.; Foi, A. Optimal Inversion of the Generalized Anscombe Transformation for Poisson–Gaussian Noise. IEEE Trans. Image Process. 2012, 22, 91–103. [Google Scholar] [CrossRef]
  45. Jing, J.; Xia, W.; Hou, M.; Chen, H.; Liu, Y.; Zhou, J.; Zhang, Y. Training low dose CT denoising network without high quality reference data. Phys. Med. Biol. 2022, 67, 084002. [Google Scholar] [CrossRef]
  46. Zhang, S.Y.; Wang, Z.X.; Yang, H.B.; Chen, Y.L.; Li, Y.; Pan, Q.; Wang, H.K.; Zhao, C.X. Hformer: Highly efficient vision transformer for low-dose CT denoising. Nucl. Sci. Tech. 2023, 34, 61. [Google Scholar] [CrossRef]
  47. Zhang, J.; Ye, L.; Gong, W.; Chen, M.; Liu, G.; Cheng, Y. A novel network for low-dose ct denoising based on dual-branch structure and multi-scale residual attention. J. Imaging Inform. Med. 2025, 38, 1245–1264. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Schematic diagram of the proposed model, inspired by the FFDNet/DnCNN architecture.
Figure 1. Schematic diagram of the proposed model, inspired by the FFDNet/DnCNN architecture.
Electronics 15 00124 g001
Figure 2. Comparison of image quality across different dose levels and reconstruction. The 20 mA image (left) shows degradation from the low-dose simulation; the reconstructed 80 mA image (middle) demonstrates the recovery achieved by the proposed noise correction model; and the 80 mA image (right) serves as the high-quality reference. The window level and window width used in this example are −875 and 4298, respectively, representing a full dynamic view in this case.
Figure 2. Comparison of image quality across different dose levels and reconstruction. The 20 mA image (left) shows degradation from the low-dose simulation; the reconstructed 80 mA image (middle) demonstrates the recovery achieved by the proposed noise correction model; and the 80 mA image (right) serves as the high-quality reference. The window level and window width used in this example are −875 and 4298, respectively, representing a full dynamic view in this case.
Electronics 15 00124 g002
Figure 3. Comparison of image quality across different dose levels and reconstruction when zoomed. The 20 mA image (left) shows degradation from the low-dose simulation; the reconstructed 80 mA image (middle) demonstrates the recovery achieved by the proposed noise correction model; and the 80 mA image (right) serves as the high-quality reference. The window level and window width used in this example are −875 and 4298, respectively, representing a full dynamic view in this case.
Figure 3. Comparison of image quality across different dose levels and reconstruction when zoomed. The 20 mA image (left) shows degradation from the low-dose simulation; the reconstructed 80 mA image (middle) demonstrates the recovery achieved by the proposed noise correction model; and the 80 mA image (right) serves as the high-quality reference. The window level and window width used in this example are −875 and 4298, respectively, representing a full dynamic view in this case.
Electronics 15 00124 g003
Table 1. Summary of test data sets, their radiation dose levels, and number of images.
Table 1. Summary of test data sets, their radiation dose levels, and number of images.
Test CaseDose# Images
STC180 mA2376
STC280 mA304
STC380 mA392
STC480 mA176
STC580 mA1120
STC640 mA1120
STC780 mA1120
STC840 mA1120
Table 2. Mean Squared Errors (MSE) from simulated reconstructed test cases with 95% Confidence Intervals.
Table 2. Mean Squared Errors (MSE) from simulated reconstructed test cases with 95% Confidence Intervals.
Test CaseMSE NoisyMSE Recon. (80 mA)
MeanMinMaxStd. Dev.MeanMinMaxStd. Dev.95% CI (±)
STC137.5035.5539.780.877.622.7929.523.570.15
STC237.0334.9238.340.726.452.6427.343.670.41
STC338.0635.9540.361.037.763.1040.274.540.45
STC438.0135.8239.540.917.233.1018.453.550.53
STC535.4133.1037.520.726.083.8822.232.170.13
STC635.1932.3537.640.767.774.9128.362.790.16
STC732.2328.8235.671.366.253.8826.082.630.15
STC832.1428.7046.281.468.255.1632.133.210.19
Table 3. Peak Signal-to-Noise Ratio (PSNR) from simulated reconstructed test cases with 95% Confidence Intervals.
Table 3. Peak Signal-to-Noise Ratio (PSNR) from simulated reconstructed test cases with 95% Confidence Intervals.
Test CasePSNR NoisyPSNR Recon. (80 mA)
MeanMinMaxStd. Dev.MeanMinMaxStd. Dev.95% CI (±)
STC132.3932.1332.620.1039.7433.4343.681.930.08
STC232.4532.2932.700.0940.5933.7643.912.110.24
STC332.3332.0732.570.1239.7932.0843.212.110.21
STC432.3332.1632.590.1040.0135.4743.212.010.30
STC532.6432.3932.930.0940.5134.6642.241.300.08
STC632.6732.3733.030.0939.4433.6041.221.300.08
STC733.0532.6133.530.1840.4533.9742.251.450.09
STC833.0631.4833.550.1939.2133.0641.011.350.08
Table 4. Structural Similarity Index (SSIM) from simulated reconstructed test cases with 95% Confidence Intervals.
Table 4. Structural Similarity Index (SSIM) from simulated reconstructed test cases with 95% Confidence Intervals.
Test CaseSSIM NoisySSIM Recon. (80 mA)
MeanMinMaxStd. Dev.MeanMinMaxStd. Dev.95% CI (±)
STC10.6370.6280.6480.0040.9000.8930.9060.0020.00008
STC20.6300.6210.6460.0060.9020.8970.9060.0020.00023
STC30.6330.6190.6470.0070.8960.8890.9020.0020.00020
STC40.6350.6240.6510.0060.9000.8940.9050.0020.00030
STC50.6530.6410.6700.0050.8880.8830.8940.0020.00012
STC60.6640.6520.6850.0050.8790.8710.8850.0020.00012
STC70.6750.6560.6950.0080.8880.8800.8960.0030.00018
STC80.6880.6690.7060.0080.8720.8640.8820.0030.00018
Table 5. FID scores for noisy and reconstructed CT images across different test cases.
Table 5. FID scores for noisy and reconstructed CT images across different test cases.
Test CaseOriginal DoseSimulated DoseFID
NoisyRecon. (80 mA)
STC180 mA20 mA 1.4841 1.1341
STC280 mA20 mA 1.5997 1.1824
STC380 mA20 mA 1.5973 1.1895
STC480 mA20 mA 1.4461 0.9781
STC580 mA20 mA 1.7506 1.2513
STC640 mA20 mA 2.2662 1.7466
STC780 mA15 mA 2.2332 1.6235
STC840 mA10 mA 3.0284 2.5322
Table 6. Kernel Inception Distance (KID) mean, standard deviation, and 95% confidence interval computed for each test case.
Table 6. Kernel Inception Distance (KID) mean, standard deviation, and 95% confidence interval computed for each test case.
Test CaseDose/Tube CurrentKID
NoisyRecon. (80 mA)
MeanStd. Dev.MeanStd. Dev.95% CI (±)
STC120 mA 1.1614 0.4205 1.0352 0.6452 0.0259
STC220 mA 1.3145 0.6425 1.0571 0.6734 0.0757
STC320 mA 1.2368 0.5226 0.9996 0.7369 0.0729
STC420 mA 1.2320 0.5512 0.9155 0.5561 0.0822
STC520 mA 1.5137 0.5176 1.1812 0.5850 0.0343
STC620 mA 2.4250 0.7242 1.7895 0.7885 0.0462
STC715 mA 2.1443 0.7737 1.9666 0.6503 0.0381
STC810 mA 3.8554 1.5322 2.2006 0.9606 0.0563
Table 7. NIQE scores for simulated noisy and denoised CT images across different test cases and 95% confidence interval. All values are reported with two decimal places.
Table 7. NIQE scores for simulated noisy and denoised CT images across different test cases and 95% confidence interval. All values are reported with two decimal places.
Test CaseSimulatedDenoised
MeanMinMaxStd. Dev.MeanMinMaxStd. Dev.95% CI (±)
STC111.5610.0713.050.404.573.885.480.250.010
STC211.9910.3113.380.515.184.625.780.240.027
STC311.9910.4913.970.534.354.004.900.160.016
STC411.8510.9213.350.454.323.954.710.160.024
STC510.969.3712.660.464.053.574.750.190.011
STC610.409.0811.900.483.563.223.970.120.007
STC710.579.2711.880.453.993.364.650.240.014
STC89.778.5811.480.463.693.164.420.180.011
Table 8. Temporal Structural Similarity Index (tSSIM) demonstrating frame-to-frame consistency across test cases.
Table 8. Temporal Structural Similarity Index (tSSIM) demonstrating frame-to-frame consistency across test cases.
Test CaseTemporal SSIM
STC10.9870
STC20.9874
STC30.9851
STC40.9828
STC50.9771
STC60.9632
STC70.9776
STC80.9600
Table 9. Comparison of reported PSNR and SSIM values from recent LDCT denoising studies with those of the proposed method.
Table 9. Comparison of reported PSNR and SSIM values from recent LDCT denoising studies with those of the proposed method.
AuthorsMethodPSNRSSIM
Yang et al. [21]GAN24.06370.8090
Ma et al. [37]GAN29.65600.8596
Zhang et al. [46]Hformer33.44050.9163
Zhang et al. [47]ADCB29.30040.8659
Proposed MethodDnCNN-based39.96000.8906
Table 10. Comparison of image quality metrics (MSE, PSNR, SSIM) for noisy images and reconstructions without and with GAT.
Table 10. Comparison of image quality metrics (MSE, PSNR, SSIM) for noisy images and reconstructions without and with GAT.
Test CaseNoisyWithout GATWith GAT
MSEPSNRSSIMMSEPSNRSSIMMSEPSNRSSIM
STC137.5032.390.63717.8335.620.7827.6239.740.900
STC237.0332.450.63012.6237.120.7486.4540.590.902
STC338.0632.330.63321.0934.890.8257.7639.790.896
STC438.0132.330.63513.0336.980.7147.2340.010.900
STC535.4132.640.65311.7037.450.8616.0840.510.888
STC635.1932.670.66420.0935.100.7437.7739.440.879
STC732.2333.050.67515.5336.220.8036.2540.450.888
STC832.1433.060.68817.3035.750.8358.2539.210.872
Table 11. Qualitative scoring from human expert on a scale of 1–5.
Table 11. Qualitative scoring from human expert on a scale of 1–5.
QuestionSTC1STC2STC3STC4STC5STC6STC7STC8Mean Score
Does the image provide
sufficient detail to
support diagnosis confidently?
4.004.004.004.004.004.004.004.004.00
Would you consider this
image diagnostically equivalent
to a high-dose (80 mA) scan?
4.004.004.003.504.004.004.504.504.06
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hasan, M.; So, A.; El-Sakka, M.R. DnCNN-Based Denoising Model for Low-Dose Myocardial CT Perfusion Imaging. Electronics 2026, 15, 124. https://doi.org/10.3390/electronics15010124

AMA Style

Hasan M, So A, El-Sakka MR. DnCNN-Based Denoising Model for Low-Dose Myocardial CT Perfusion Imaging. Electronics. 2026; 15(1):124. https://doi.org/10.3390/electronics15010124

Chicago/Turabian Style

Hasan, Mahmud, Aaron So, and Mahmoud R. El-Sakka. 2026. "DnCNN-Based Denoising Model for Low-Dose Myocardial CT Perfusion Imaging" Electronics 15, no. 1: 124. https://doi.org/10.3390/electronics15010124

APA Style

Hasan, M., So, A., & El-Sakka, M. R. (2026). DnCNN-Based Denoising Model for Low-Dose Myocardial CT Perfusion Imaging. Electronics, 15(1), 124. https://doi.org/10.3390/electronics15010124

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop