Next Article in Journal
Machine Learning Methods in Predicting Patients with Suspected Myocardial Infarction Based on Short-Time HRV Data
Next Article in Special Issue
Enhanced Deep-Learning-Based Automatic Left-Femur Segmentation Scheme with Attribute Augmentation
Previous Article in Journal
Application of a Wireless and Contactless Ultrasonic System to Evaluate Optimal Sawcut Time for Concrete Pavements
Previous Article in Special Issue
Automated Precancerous Lesion Screening Using an Instance Segmentation Technique for Improving Accuracy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Recovery from Synthetic Noise Artifacts in CT Scans Using Modified U-Net

1
School of Science, Computing and Engineering Technologies, Swinburne University of Technology, Hawthorn, VIC 3122, Australia
2
Macquarie University Hearing (MU Hearing), Centre for Healthcare Resilience and Implementation Science, Macquarie University, Macquarie Park, NSW 2109, Australia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(18), 7031; https://doi.org/10.3390/s22187031
Submission received: 19 August 2022 / Revised: 11 September 2022 / Accepted: 11 September 2022 / Published: 16 September 2022
(This article belongs to the Special Issue Artificial Intelligence-Based Applications in Medical Imaging)

Abstract

:
Computed Tomography (CT) is commonly used for cancer screening as it utilizes low radiation for the scan. One problem with low-dose scans is the noise artifacts associated with low photon count that can lead to a reduced success rate of cancer detection during radiologist assessment. The noise had to be removed to restore detail clarity. We propose a noise removal method using a new model Convolutional Neural Network (CNN). Even though the network training time is long, the result is better than other CNN models in quality score and visual observation. The proposed CNN model uses a stacked modified U-Net with a specific number of feature maps per layer to improve the image quality, observable on an average PSNR quality score improvement out of 174 images. The next best model has 0.54 points lower in the average score. The score difference is less than 1 point, but the image result is closer to the full-dose scan image. We used separate testing data to clarify that the model can handle different noise densities. Besides comparing the CNN configuration, we discuss the denoising quality of CNN compared to classical denoising in which the noise characteristics affect quality.

Graphical Abstract

1. Introduction

The Global Cancer Observatory (GLOBOCAN) report estimates 19.3 million new cancer cases with almost 10 million deaths in 2020, lung cancer being the second most frequently diagnosed cancer after female breast cancer. However, lung cancer has the highest mortality at 18% [1] as the diagnosis often occurs at the late cancer stages. A randomized trial for lung cancer screening in the Netherlands and Belgium, which started in 2003, showed an increase in patient survivability as early cancer detection led to early treatment [2]. Early detection of cancer using cancer screening, targets people at high risk of lung cancer, such as heavy smokers and people with a family history of lung cancer. Screening is through CT scans, which utilize lower dose radiation while maintaining image quality just enough for the assessment [3]. The concept began in 1990 in which the image has enough detail to detect lung parenchyma but suffers image degradation from increased noise [4]. Nowadays, low radiation scans, known as Low Dose CT (LDCT), have become a popular choice for screening. These use approximately 25–30% of the standard dose radiation [5]. However, the LDCT has an increased noise level, leading to a risk of incorrect assessment. For example, it was found that cancer screening using LDCT of 53,454 people resulted in 24.2% positive, but 95% were false positive [6].
Tube current, collimation factor, bowtie filter, and exposure time are factors that can affect radiation dose during scanning. The radiation dose is proportional to the photon count, coupled with the absorption and scattering tendency during the scan, which may result in the detector receiving fewer photons. As a result, noise may appear in the projection data. The noise changes into a streak line after image reconstruction. There is also electronic noise on the image, such as shot noise, although in a modern CT scanner, the impact of electronic noise is less apparent due to a better electronic circuitry design [7].
Noise removal from low-dose CT scan images has become a growing area for research. Denoising methods have expanded from classical denoising such as Gaussian Smoothing, Non-Local Mean (NLM), to 3D Block Matching (known as BM3D) into using various CNN models, each with its own strength. Denoising CNN requires a learning process that depends on full-dose and low-dose radiation images, and it is impossible to get an exact copy of the image due to possible patient movement or even motion artifacts in between scans [8] and the risk of radiation, even though some researchers have attempted to have two runs within the scan, with a radiologist as the observer [9,10].
In classical denoising, there has been research to improve BM3D to address CT scan noise using a context-based method. The quality score was based on Contrast to Noise, which improved around 30% on average from the original BM3D [11]. There is various research on using CNN for denoising. A Generative Adversarial Network (GAN) method involved training denoising quality with a generator utilizing control of a discriminator on a 3D cardiac CT, resulting in four points of Peak Noise to Signal Ratio (PSNR) improvement [12]. There has been research using an offline dictionary (database of patches) to reconstruct low-dose CT by restoration using similar image patches. The quality measurement used radiologist scores [13,14]. Several modifications of U-Net for CT scan denoising have been used by reducing the filter size and adding an extra layers per stages [15], by removing a concatenated layer and adding a recurrent-residual connection [16], by using a stagnant filter size of 128 on all stages, and using a Leaky Rectified Linear Unit (ReLU) as the activation function [17]. There has been research using different network architecture and processes on targeting the residual noise to train in Residual Network (ResNet) [18].
As the main contribution of this research, we propose a denoising solution that stems from the understanding of several CNN models. The proposed model is a stacked modified U-Net model. The modification includes a transpose convolution layer for the decoder stage, increasing spatial learning using convolution layer stride-2 for contraction, improving noise detection using a high filter number at the early layer, maintaining the number of filters across all layers, and stacking two models together for improvement. These modifications have not been used together before to deal with the denoising problem. The dataset used synthetic noise obtainable by using a noise insertion tool at 10% of the standard dose. The approximate difference was 5–6% of the actual screening images. The structure of this paper is as follows: Section 2 covers the methodology of determining the CNN layers and the model. Section 3 describes results and comparative quality improvement before and after denoising, and comparative study of another method. Section 4 discusses the impact of denoising on a different dataset. The conclusion is provided in Section 5.

2. Materials and Methods

2.1. The Data Set

The data set used in this study was from the research of LDCT-and-Projection-data, stored in the Cancer Imaging Archive (TCIA) library. We utilized six patient datasets (out of 299 patients in the pool) with 1906 image pairs to train the neural network and test the result independently. Image pairs means each standard-dose image has a synthetic low-dose image copy. The sample dataset was the scan result of the Somatom Definition AS+ and Somatom Definition Flash CT scanner’s scan result from a patient with and without solid non-calcified nodules [19].

2.2. Noise Simulation

Noise simulation was obtained using a noise insertion tool that utilizes a validated photon-counting model applied to the post-processed projection data of a standard-dose scan. The images had a CT scan setting of 1.5 mm slice thickness, 120 kV, and a tube current between 650–750 mA for the full-dose scan. The projection data were extracted from the scanner to add the synthetic noise using a known noise model, then the modified projection data were returned to the scanner to utilize weighted, filtered back-projection (FBP). The simulation target was 10% of the full-dose radiation dose for the chest scan [12,20].
A theoretical approach for noise simulation, by altering projection data using Gaussian noise, was considered and tested. The projection data were calculated from a standard-dose image using Radon transformation.
ρ ω , θ = f x , y e i 2 ω x c o s θ + y s i n θ d x d y
where ρ ω , θ is projection data (sinogram) of the CT scan image f x , y (1). The detected photon count (attenuated intensity) was traceable using the Beer-Lambert Law for an non-homogenous material and a bowtie filter profile.
I = I 0 · P b · e μ x d x = I 0 · P b · e ρ ω , θ
where I is the detected photon count (count per second—cps), and I 0 is the original photon count (incident flux—cps), P b is the bowtie profile and μ x d x is the non-homogeneous target equivalent to the projection data ρ ω , θ (2).
I 0 = K · c · m A · s
The incident flux I 0 depends on the scanner constant K , collimation factor c , radiation dose m A , and exposure time s (3).
I 0 L = 10 % I 0
I L = P s I 0 L · P b · e ρ ω , θ
The simulation of low-dose radiation I 0 L was achievable by reducing the incident flux value to 10% of the standard dose (4). The implementation of random Poisson distribution P s generates noise in the new attenuated intensity I L (5).
ρ L ω , θ = l n I L / I 0 L
The synthetic low-dose projection data ρ L ω , θ is recalculated using Formula (6), with inverse Radon transformation to reconstruct the image [5]. This research focuses on neural networks, and there was no testing on the actual CT scanner. Thus, it was impossible to obtain the bowtie filter profile, exposure control, and electronic noise. Furthermore, back tracing the image into the projection data can cause deviation from the projection values, leading to greater error in the denoising results.

2.3. Selection of Denoising Technique

There are many approaches to address the image denoising problem, from the classical process to the recent utilization of neural networks. Classical denoising aims to remove noise by calculating the pixel grey value to the spatial or transform domain. In comparison, network training in neural networks aims to distinguish noise features and remove them [21]. We considered three classical denoising methods that use a spatial or transform domain to compare the denoising quality of the proposed CNN. The three methods were Gaussian smoothing, Non-Local Mean (NLM), and 3D block matching (BM3D). Gaussian smoothing applies a Gaussian function to the image to reduce noise using a smooth blur. The NLM method was introduced in 2005 for image denoising by weighting all pixels in the image with similarity to the target pixels [22]. While the BM3D method was introduced in 2007 using the same grouping method as NLM using thresholding, followed by applying the Wiener filter [23].
The proposed network stems from use of an autoencoder combined with CNN for image denoising. Several design layers fit into solving the denoising problem, each with a minor modification, even though the original development intention was for segmentation or classification. These designs became a reference for developing the new, proposed network. The first reference was U-Net, created in 2015 to deal with image segmentation and classification problems. The first half of the model consists of encoder layers that use a dual 3 × 3 kernel for the convolution process and a Rectified Linear Unit (ReLU) activation for each layer. The encoder layer uses a Max Pooling layer for contraction to move into deeper encoder layers. The second half consists of the decoder layer with the same dual 3 × 3 kernel for the convolution process and ReLU activation. This uses transpose convolution on stride 2 for expansion and skips the concatenation link from the corresponding encoder layer. There is a dropout layer before the last pooling and the first transpose convolution layer. Initially, the network used the SoftMax layer to assign probability distribution on the classification process [24]. Changing it into a Regression layer enabled the network to deal with the denoising problem. The number of filters in U-Net start at 64 and double every encoder layer with the deepest encoder layer having 1024 filters. Some researchers have modified the layer depth. The number of features to improve the training loss with each modification depends on the noise characteristic. The denoising target varies from an image with Gaussian noise [25] and a low-dose CT image (using noise simulation through Poisson distribution on the projection data) [26].
The second reference was the Segmentation Network (SegNet), developed in 2016 for dealing with the segmentation problem. It has a similar structure to U-Net, including a dual 3 × 3 kernel for the convolution process. The design utilizes the Batch Normalization layer before the ReLU activation layer and Un-polling skip connection [27]. The third reference is derived from the Deconvolution Network (DeConvNet) developed in 2016 for semantic segmentation. DeConvNet has the same characteristic as the SegNet, except the decoder layers utilize transpose convolution [28]. The original SegNet and DeConvNet have the same filter number of 64 across all layers. We increased the filter number to 128 for comparison purposes. Replacing the Softmax layer with a Regression Layer in both and removing the Batch Normalization layer enabled the SegNet or DeConvNet to deal with the denoising problem.
The fourth reference came from Residual Encoder-Decoder Network (RedNet), developed in 2016 for image restoration. RedNet uses encoder and decoder layers without a contraction or expansion process. The activation layer is ReLU, the decoder layer uses the transpose convolution process, and the summation layer is the skip connection. Without the contraction and expansion layer, the training can reduce spatial degradation and improve memory usage [29]. RedNet has the same number of filters (128) across all layers and has been used for CT scan denoising using Poisson noise on projection data [30]. The last reference came from stacking a model together, the Stacked U-Net, which was used to set-up two or more U-Net together for segmentation in 2018 [31] and 2019 [32], and for image reconstruction in 2021 [33].
The previous stacked U-Net models were used on different applications, and more modification was required to solve the denoising problem and to optimize the computational time, such as reducing the number of U-Net in the stack and increasing the convolutional operation in each block stage. The modification make the model different from the original, and for that reason there is no comparison to the existing stacked U-Net included in this paper. We only considered the stacked setting from those models without following the stack number in the existing models.
The proposed network contains two stacked U-Net layers, each with encoder and decoder layers with 128 as the number of filters on each layer. While the comparison network uses an encoder depth of 4, our model reduced it to 3. The encoder layers use a triple 3 × 3 kernel for the convolution process and Rectified Linear Unit (ReLU) activation for each layer; an extra convolution layer on stride 2 is for contraction to the deeper encoder layer.
This approach, which does not use a pooling layer, optimizes weight and biases and improves model accuracy [34]. The decoder layers use a triple 3 × 3 kernel for the transpose convolution process and Rectified Linear Unit (ReLU) activation for each layer. The expansion utilizes another transpose convolution on stride 2. The utilization of transpose convolution for decoder layers follows the idea from DeConvNet and RedNet. The skip connection uses the concatenation layer between the corresponding encoder and decoder on each network and between the stacked network (Figure 1). The input image layer has 512 × 512 pixels without a downscaling process to maintain the original spatial pixels. Training uses an Adaptive Moment Estimation (Adam) optimizer on its default settings to update the weight. Adam performs better than Scholastic Gradient Descent (SGD), Root Mean Square Propagation (RMSProp), Adaptive Gradient (AdaGrad), and Adaptive Delta (AdaDelta) [35]. Two previous studies used the same optimization for CNN image denoising [36] and MRI image denoising using the feature extraction method [20].
Each block before contraction or expansion layer may have different convolutional layers, one and two are the most common. Figure 2 shows the naming arrangement for each layer number in a block in this paper.

2.4. Performance Analysis

Network training used Root Mean Square Error (RMSE) to calculate the loss (difference) between the target image and the training output.
R M S E = i = 1 P x t i o i 2 / P x
If t i is the target pixel, o i is the output pixel after each training epoch, and P x is the number of pixels in the image (512 × 512 = 262,144), then the RMSE formula is as per Equation (7). A few analysis methods are available to determine the denoising quality; Peak Signal to Noise Ratio (PSNR) measures the ratio between peak signal against noise in the image. This can determine the amount of noise before and after denoising related to the maximum value of the image pixel. A higher PSNR score indicates less noise in the image, and score improvement can show the effectiveness of the denoising process.
P S N R = 20   log 10 f m a x / m s e
PSNR values are on a logarithmic scale between the maximum signal ( f m a x ) and the mean square error ( m s e ) (8). The mean square error is the average square of the difference between the target and the reference image. The aim is to find the improvement of PSNR score between the low-dose CT scan images and the denoising result.
S S I M t , r = l t , r α + c t , r β + s t , r γ
The second method uses a Structural Similarity Index (SSIM) to determine the image’s luminance (brightness), contrast, and structural information compared to the reference image [37]. This method allows checking the score improvement before and after the denoising process. A higher SSIM score indicates a close resemblance to the reference image. SSIM scores depend on the calculation of three components, the luminance l t , r ; the contrast c t , r and the structural information s t , r (9). The quality improvement follows the same process as PSNR for this method.
N I Q E = v 1 v 2 τ Σ 1 + Σ 2 / 2 1 v 1 v 2
The third method uses a Natural Image Quality Evaluator (NIQE), a blind image quality assessment utilizing measurable deviation from statistical regularity on the image patches [38]. The NIQE score is calculated based on the NIQE model. The model uses Natural Scene Statistic (NSS) obtainable from Generalized Gaussian Distribution (GGD) and Asynchronous GGD (AGGD) fitting of the Mean-Subtracted Contrast Normalized (MSCN) image collection [39]. A lower NIQE score indicates a better-quality image. The calculation is based on the vector and covariance of the NIQE model ( v 1 and Σ 1 ) and the target image ( v 2 and Σ 2 ), with the sharpness threshold of the patch T (10). The patch size is 48 × 48 and the sharpness threshold is 0.5. Because this method uses statistical modeling, the first approach to find image quality improvement is to obtain the score for low-dose images (noisy references).
N I Q E I = N I Q E L D N I Q E D
As the scores become lower with better image quality, the improvement calculation ( N I Q E I ) is obtained from the difference between the noisy reference NIQE score ( N I Q E L D ) and the denoising result score ( N I Q E D ) (11).

3. Results

3.1. Noise Synthesis

The conversion from full-dose and low-dose of Digital Imaging and Communications in Medicine (DICOM) images into a 512 × 512 array was done using MATLAB software. An attempt to create synthetic noise from the formula did not provide good results because of the lack of Bowtie profile information. Utilizing the noise model and Bowtie profile provides clearer streak lines focusing on the area of interest with maximum width as seen in the sample images.
Without the model and Bowtie profile, the noise had fewer lines with a broader noise effect (Figure 3). This Bowtie effect created focused noise on the target area, which affects the selection of denoising methods.

3.2. Observed Network Setting

The dataset was split into two parts. The first consisted of three patients with 986 image pairs, the training utilized a random 50% of the image pairs (493 images of low-dose and 493 images of standard-dose), validation used 30% of the image pairs, and testing used 20% of the image pairs. Training used a supercomputer with a 64 GB GPU and 80 GB CPU memory and MATLAB 2021b software. The performance scores were derived from the training time of each CNN model, including the reference models for comparison. The MATLAB version was used since it provides a new function for capturing the lowest RMSE validation loss, which negates the need to use training early stopping.
The network used the ADAM optimizer setting with default parameters of β1 = 0.9, β2 = 0.999, and ε = 10−8 [31]. There were 300 training epochs in a small batch of two images. Having the entire image of 512 × 512 as an input utilized a lot of memory, and it was impossible to train with a larger batch. Training used a learning rate of 10−4, with a reduction of 0.8 every ten epochs. At 300 epochs, the final learning rate was 1.5 × 10−7.
The proposed model used two stacked networks, considering training improvement and efficiency of training time. The selection stacked network over a single network reduced the RMSE loss by 310 points and doubled the training time from 544 to 1229 min using the same 300 epochs. Considering time efficiency, when the single network stopped the training at the same 544 min, the stacked network still had a lower RSME loss by 237 points (Figure 4).
Filter size determines the observation area. A smaller filter of single 3 × 3 can run fast but has limited coverage. Tests on larger size 5 × 5 and 7 × 7 filters produced excellent RMSE losses, at the cost of long training times. Since a dual 3 × 3 filter has the same coverage as a single 5 × 5, while a triple 3 × 3 filter can cover a single 7 × 7 at a reduced number of weight parameters, both were tested to check the quality and training time of the proposed model.
The training result showed an increase in the training time from 816 min for a single 3 × 3 kernel to 1229 min for a double 3 × 3 kernel and 1564 min for a triple 3 × 3 kernel. There was an improvement in RMSE loss when using more filter layers in one encoder/decoder stage, indicating layer saturation. The RMSE loss improved by 363 points on a double 3 × 3 and improved by 60 points on a triple 3 × 3 (Figure 5).
At the same time, this shows that a dual 3 × 3 kernel had the same result as a single 5 × 5 kernel with almost half of the training time of 2191 min. At the bigger kernel size of 5 × 5, the learning rate no longer fit, as the training result showed higher oscillation up to the 70 epoch (500 min) (Figure 6).
The proposed model used a dual 3 × 3 instead of a triple 3 × 3 or a single 5 × 5/7 × 7 kernel due to the slight increase in PSNR and SSIM improvement shown in the result but utilized more memory and training time (Table 1). AV status in the header shows the average score of 174 testing images, each representing quality components NIQE, PSNR, and SSIM. The green highlight indicates the proposed network setting in quality comparison to different network setting.
The final piece of the network setting for testing was the number of filters. The number of filters was set at 128 for all layers as this provides the most optimum setting for training. The doubling up method on the filter numbers, as seen in U-Net, did not provide an improved result. Crux lies in the noise focus needed a high filter number in the early layer. The 64 filters provided a short training time (510 min) but with a higher RMSE loss of 895 points compared to the 128 filters. When the filter was increased to 256 it had a worse RMSE loss by 179 points while increasing the training time substantially (3013 min) (Figure 7).

3.3. Comparison between Denoising Techniques

Three classical denoising methods (Gaussian smoothing, NLM, and BM3D) were tested on the same image samples for quality comparison of the CNN methods, and the quality between CNN models. The proposed network model has a label of “S-Dual3” in the following bar charts, indicating the use of a dual 3 × 3 layer at each encoder/decoder stage. The other CNN models are a RedNet with 20 layers, a modified U-Net, a modified SegNet, and a modified DeConvNet. These networks retained the model structure with modifications, as stated earlier.
Table 2 indicates the training time difference between CNN models with the proposed network run for the longest time. The figures does not include classical denoising as they do not need any training.
A comparison chart showed average scores on three different quality measurements, NIQE, PSNR, and SSIM. The mean score was derived by averaging the quality score of 197 images. The NIQE comparison chart (Figure 8) shows the score of low-dose images as the X-axis, which are presented as a group to understand the chart better. The X-axis scores were obtained by measuring the low dose image against the full-dose image quality modeling result. As the chart does not accommodate the number of images in one group, there is a discrepancy between the improvement score and the total mean score of each denoising method. The Y-axis represents the NIQE improvement score which is the amount of reduction from the NIQE score of the low-dose image. There is a cap on the lower end of low-dose image NIQE score at 16; any score at or below 16 causes the denoising process to exhibit a higher score. A higher score indicates a worsening quality. Tracing back to the comparison score between low-dose ( N I Q E L D ) and full-dose image ( N I Q E F D ) shows that the latter has a higher score than the former ( N I Q E L D < N I Q E F D ).
The bar chart is configured with average NIQE scores from low (left) to high (right). The average score was calculated over all tested images, unlike in Figure 8 which shows classes of noise level. The S-Dual3 method was best with 14.89 points of improvement, followed by SegNet (14.80), DeConvNet (14.73), UNet (14.25), RedNet (14.00), BM3D (13.60), NLM (12.42), and Gaussian (11.47).
The PSNR comparison chart (Figure 9) shows the PSNR score of low-dose images on the X-axis, while the Y-axis represents the PSNR improvement score. The X-axis grouping and bar arrangement follows the same configuration of low (left) to high (right) as the NIQE chart. The S-Dual3 method was best with 10.55 points of improvement, followed by U-Net (10.01), RedNet (9.85), DeConvNet (9.72), SegNet (9.67), BM3D (6.33), Gauss (4.77), and NLM (2.28).
The SSIM comparison chart (Figure 10) has the same presentation as the previous two charts. The S-Dual3 method was best with 0.00895 points of improvement, followed with RedNet (0.00873), DeConvNet (0.00872), U-Net (0.0087), SegNet (0.0087), BM3D (0.00731), Gauss (0.00452), and NLM (0.00365).
Both PSNR and SSIM charts show a similar equalizing trend with a higher denoising effect on higher noise density. There was a slight difference in these two quality scores between CNN models, with the proposed model showing the highest score and observable difference. The classical denoising techniques had lesser scores with significant discrepancies than the CNN techniques. This result indicates the inability of classical denoising to address CT scan noise that occurs with different densities (focused noise).

4. Discussion

This study compared the visual observation of a denoising image against a quality score to obtain a relationship between those two. A suitable location would be either the upper thorax or diaphragm area, in which the area of interest contains a rich contrast variation.
Figure 11 shows the full-dose image and its counterpart of the synthetic low-dose image from the upper thorax area from one of the data samples. The number indicates a PSNR score of 50.7883, with only one quality score for comparison. A noise mask was created by taking the difference between those two images and applying the contrast setting at 10% of the maximum pixel value. The noise mask shows a higher noise density in the area of interest (patient’s body) that appears as a short line compared to the other area, which resembles “shot” noise. The denoising result from various techniques shows image quality with an accompanying noise mask to better understand the amount of noise that has been removed. The denoising noise mask is derived from the difference between the result and full-dose image, applied the same contrast setting as the previous mask to retain the same reference (Figure 12).
Classical denoising had less impact on removing noise. The NLM technique reduced some noise outside the area of interest and could not deal with focused noise, with a slight increase of PSNR score less than 1 point. Gaussian smoothing reduced a lot of noise; thus, the high PSNR score improvement at 6.7112. The downside was the introduction of blur to the image, which had a detrimental effect on assessment. The BM3D technique provided better noise removal than NLM, even reducing the impact of focused noise, although the limitation appeared at higher noise density, with a PSNR score improvement at 2.9458. The SegNet and DeConvNet techniques had close PSNR improvement scores of 10.6364 and 10.7092; the denoising result and noise mask were similar. Both models had differences in the decoder stage between a standard convolution layer and transpose convolution layer. The slight improvement indicated that the application of transpose convolution into the decoder stage gave a better result.
The RedNet technique was slightly worse than SegNet, DeConvNet, U-Net, and the proposed model, with a PSNR improvement score of 10.3974. The difference was the missing contraction + expansion layer in the model, creating lesser quality. The U-Net had the second highest PSNR improvement score at 10.8335. The noise mask showed less noise, but the image showed higher contrast than the full-dose image. This contrast is because of the Max Pooling layer. The same effect appeared using the RedNet technique. The proposed model had an improvement score of 11.5187, and showed slightly more noise in the mask than U-Net, but resembled the full-dose image.
Two additional image set (Figure 13 and Figure 14) were added to confirm better improvement resulting from the proposed network.
During repeated training, there is a chance that the 30% validation data set (295 images) may lean into a group of images with lower or higher noise density. The same can happen on the remaining 20% testing data set (197 images), making the NIQE, PSNR, and SSIM scores vary during each test. Furthermore, the training dataset had different areas of interest due to the patient size, creating a difference in noise density. Confirming the CNN network can address CT scan images in general, so a completely new dataset was used for comparison to guarantee the best denoising result. The second part of the dataset was used for independent testing, and consisted of three patients with 920 image pairs.
The testing phase used only 50% of the synthetic low-dose image in the pool (460 images), and their standard dose pairs were used for quality score calculation. We used a different CT scan image set to confirm the network quality was still working well with different sizes in the area of interest. This comparison targeted only on CNN models while excluding the classical denoising method. The NIQE, PSNR, and SSIM comparison charts (Figure 15, Figure 16 and Figure 17) show a similar result. NIQE scores were worse when dealing with noise below 16 on the X-axis. The PSNR chart shows a surprising improvement for the proposed model with a greater PSNR score (X-axis—less noise), better than the other models. The proposed model led the improvement score for SSIM as well. The equalizing trends are visible on both PSNR and SSIM charts; more noisy images improved than less noisy ones.
This training result was tested against the actual screening image showing each method’s denoising quality. The screening dataset was obtained from National Lung Screening Trial research [40], stored in the Cancer Imaging Archive (TCIA) library [41]. We run the test using seven methods, since the BM3D requires a good image to calculate the standard deviation. Figure 18 shows that the NLM method left a trace of noise, the Gaussian caused a little image blur, and the SegNet, and DeConvNet shared the same spotty pixels i the solid area upper left and right and had contrast enhancement. The last three images had similar results to the RedNet, and UNet had higher contrast enhancement than the S-Dual3.

5. Conclusions

The proposed CNN model was developed to denoise low-dose CT scan images and compare various CNN models. This research proved that using particular layers and having specific models can improve the denoising outcome. The denoising result showed a close resemblance to the standard-dose image compared to other models. It illustrated that quality scores could guide quality improvement, but could not determine the same perceptual quality.
The proposed model has a suitable depth and optimum setting considering memory usage, training time, and denoising quality. A possible direction for future improvement is utilizing different layer combinations and depth arrangements. Because of resemblance of results to the standard-dose image, the model is suitable for segmentation and classification work. Caution is recommended because of the slight noise variation from the low-dose image during training, and deviation of the denoising result from the standard-dose image.

Author Contributions

Conceptualization, R.G. and R.C.; methodology, R.G.; software, R.G.; validation, R.G., Y.T., H.N. and R.C.; formal analysis, R.G., Y.T., H.N. and R.C.; investigation, R.G., Y.T., J.Z., H.N. and R.C.; resources, R.G., J.Z., H.N. and R.C.; writing—original draft preparation, R.G. and R.C.; writing—review and editing, R.G., Y.T., J.Z., H.N. and R.C.; visualization, R.G. and R.C.; supervision, J.Z., H.N. and R.C.; project administration, R.G. and R.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. Data were obtained from The Cancer Imaging Archive (TCIA) hosted by The National Cancer Institute (NCI), Washington University in St. Louis, and the University of Arkansas for Medical Sciences (UAMS), Low Dose CT Image and Projection Data (LDCT-and-Projection-data) (Version 4) [Data set] at https://doi.org/10.7937/9NPB-2637, accessed on 21 November 2021. Data from the National Lung Screening Trial (NLST) [Data set] at https://doi.org/10.7937/TCIA.HMQ8-J677, accessed on 21 November 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  2. de Koning, H.J.; van der Aalst, C.M.; de Jong, P.A.; Scholten, E.T.; Nackaerts, K.; Heuvelmans, M.A.; Lammers, J.J.; Weenink, C.; Yousaf-Khan, U.; Horeweg, N.; et al. Reduced Lung-Cancer Mortality with Volume CT Screening in a Randomized Trial. N. Engl. J. Med. 2020, 382, 503–513. [Google Scholar] [CrossRef] [PubMed]
  3. Genevois, P.A.; Tack, D. Dose Reduction and Optimization in Computed Tomography of the Chest. In Radiation Dose from Adult and Pediatric Multidetector Computed Tomography, 1st ed.; Springer: Berlin, Germany, 2007; Volume 10, pp. 153–159. [Google Scholar]
  4. Naidich, D.P.; Marshall, C.H.; Gribbin, C.; Arams, R.S.; McCauley, D.I. Low-dose CT of the lungs: Preliminary observations. Radiology 1990, 175, 729–731. [Google Scholar] [CrossRef] [PubMed]
  5. Massoumzadeh, P.; Don, S.; Hildebolt, C.F.; Bae, K.T.; Whiting, B.R. Validation of CT dose-reduction simulation. Med. Phys. 2008, 36, 174–189. [Google Scholar] [CrossRef] [PubMed]
  6. Aberle, D.R.; Abtin, F.; Brown, K. Computed Tomography Screening for Lung Cancer: Has It Finally Arrived? Implications of the National Lung Screening Trial. J. Clin. Oncol. 2013, 31, 1002–1008. [Google Scholar] [CrossRef]
  7. Duan, X.; Wang, J.; Leng, S.; Schmidt, B.; Allmendinger, T.; Grant, K.; Flohr, T.; McCollough, C.H. Electronic noise in CT detectors: Impact on image noise and artifacts. Am. J. Roentgenol. 2013, 201, W626–W632. [Google Scholar] [CrossRef]
  8. Nieuwenhove, V.V.; Beenhouwer, J.D.; Schryver, T.D.; Hoorebeke, L.V.; Sijbers, J. Data-Driven Affine Deformation Estimation and Correction in Cone Beam Computed Tomography. IEEE Trans. Image Process. 2017, 26, 1441–1451. [Google Scholar] [CrossRef]
  9. Prasad, S.R.; Wittram, C.; Shepard, J.; McLeaod, T.; Rhea, J. Standard-Dose and 50%—Reduced-Dose Chest CT: Comparing the Effect on Image Quality. Am. J. Roentgenol. 2002, 179, 461–465. [Google Scholar] [CrossRef]
  10. Kubo, T.; Ohno, Y.; Takenaka, D.; Nishino, M.; Gautam, S.; Sugimura, K.; Kauczor, H.U.; Hatabu, H. Standard-dose vs. low-dose CT protocols in the evaluation of localized lung lesions: Capability for lesion characterization—iLEAD study. Eur. J. Radiol. Open 2016, 3, 67–73. [Google Scholar] [CrossRef]
  11. Chen, L.L.; Gou, S.P.; Yao, Y.; Bai, J.; Jiao, L.; Sheng, K. Denoising of Low Dose CT Image with Context-Based BM3D. In Proceedings of the IEEE Region 10 Conference (TENCON), Singapore, 22–25 November 2016. [Google Scholar]
  12. Wolterink, J.M.; Leiner, T.; Viergever, M.A.; Isgum, I. Generative Adversarial Networks for Noise Reduction in Low-Dose CT. IEEE Trans. Med. Imaging 2017, 36, 2536–2545. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Rong, J.; Lu, H.; Xing, Y.; Meng, J. Low-Dose Lung CT Image Restoration Using Adaptive Prior Features From Full-Dose Training Database. IEEE Trans. Med. Imaging 2017, 36, 2510–2523. [Google Scholar] [CrossRef] [PubMed]
  14. Yu, F.; Chen, Y.; Luo, L. CT image denoising based on sparse representation using global dictionary. In Proceedings of the ICME International Conference on Complex Medical Engineering, Beijing, China, 25–28 May 2013. [Google Scholar]
  15. Alsamadony, K.L.; Yildirim, E.U.; Glatz, G.; Waheed, U.B.; Hanafy, S.M. Deep Learning Driven Noise Reduction for Reduced Flux Computed Tomography. Sensors 2021, 21, 1921. [Google Scholar] [CrossRef] [PubMed]
  16. Nasrin, S.; Alom, M.Z.; Burada, R.; Taha, T.M.; Asari, V.K. Medical Image Denoising with Recurrent Residual U-Net (R2U-Net) base Auto-Encoder. In Proceedings of the IEEE National Aerospace and Electronics Conference (NAECON), Dayton, OH, USA, 15–19 July 2019. [Google Scholar]
  17. Huang, S.C.; Hoang, Q.V.; Le, T.H.; Peng, Y.T.; Huang, C.C.; Zhang, C.; Fung, B.C.M.; Cheng, K.H.; Huang, S.W. An Advanced Noise Reduction and Edge Enhancement Algorithm. Sensors 2021, 21, 5391. [Google Scholar] [CrossRef] [PubMed]
  18. Yang, W.; Zhang, H.; Yang, J.; Wu, J.; Yin, X.; Chen, Y.; Shu, H.; Luo, L.; Coatrieux, G.; Gui, Z.; et al. Improving Low-Dose CT Image Using Residual Convolutional Network. IEEE Access 2017, 5, 24698–24705. [Google Scholar] [CrossRef]
  19. McCollough, C.H.; Chen, B.; Holmes, D.R.I.; Duan, X.; Yu, Z.; Yu, L.; Leng, S.; Fletcher, J.G. Low Dose CT Image and Projection Data (LDCT-and-Projection-data) (Version 4) [Data set]. In The Cancer Imaging Archive; TCIA: Manchester, NH, USA, 2020. [Google Scholar] [CrossRef]
  20. Yao, L. A Multifeature Extraction Method Using Deep Residual Network for MR Image Denoising. Comput. Math. Methods Med. 2020, 2020, 8823861. [Google Scholar] [CrossRef]
  21. Fan, L.; Zhang, F.; Fan, H.; Zhang, C. Brief review of image denoising techniques. Vis. Comput. Ind. Biomed. Art 2019, 2, 7. [Google Scholar] [CrossRef]
  22. Buades, A.; Coll, B.; Morel, J.M. A non-local algorithm for image denoising. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–26 June 2005. [Google Scholar]
  23. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  24. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  25. Gurrola-Ramos, J.; Dalmau, O.; Alarcon, T.E. A Residual Dense U-Net Neural Network for Image Denoising. IEEE Access 2021, 9, 31742–31754. [Google Scholar] [CrossRef]
  26. Heinrich, M.P.; Stille, M.; Buzug, T.M. Residual U-Net Convolutional Neural Network Architecture for Low-Dose CT Denoising. Curr. Dir. Biomed. Eng. 2018, 4, 297–300. [Google Scholar] [CrossRef]
  27. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  28. Noh, H.; Hong, S.; Han, B. Learning Deconvolution Network for Semantic Segmentation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015. [Google Scholar]
  29. Mao, X.J.; Shen, C.; Yang, T.B. Image Restoration Using Convolutional Auto-encoders with Symmetric Skip Connections. arXiv 2016, arXiv:1606.08921. [Google Scholar]
  30. Chen, H.; Zhang, Y.; Kalra, M.K.; Lin, F.; Chen, Y.; Liao, P.; Zhou, J.; Wang, G. Low-Dose CT With a Residual Encoder-Decoder Convolutional Neural Network. IEEE Trans. Med. Imaging 2017, 36, 2524–2535. [Google Scholar] [CrossRef] [PubMed]
  31. Ghosh, A.; Ehrlich, M.; Shah, S.; Davis, L.; Chellappa, R. Stacked U-Nets for Ground Material Segmentation in Remote Sensing Imagery. In Proceedings of the Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake CIty, UT, USA, 18–22 June 2018. [Google Scholar]
  32. Sevastopolsky, A.; Drapak, S.; Kiselev, K.; Snyder, B.; Keenan, J.; Georgievskaya, A. Stack-U-Net: Refinement network for improved optic disc and cup image segmentation. In Proceedings of the Image Processing, San Diego, CA, USA, 16–21 February 2019; pp. 1–9. [Google Scholar]
  33. Mizusawa, S.; Sei, Y.; Orihara, R.; Ohsuga, A. Computed tomography image reconstruction using stacked U-Net. Comput. Med. Imaging Graph. 2021, 90, 101920. [Google Scholar] [CrossRef] [PubMed]
  34. Ayachi, R.; Afif, M.; Said, Y.; Atri, M. Strided Convolution Instead of Max Pooling for Memory Efficiency of Convolutional Neural Networks. In Proceedings of the SETIT 2018, Maghreb, Tunisia, 18–20 December 2018. [Google Scholar]
  35. Kingma, D.P.; Ba, J.L. ADAM: A Method for Stochastic Optimization. In Proceedings of the ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  36. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a Fast and Flexible Solution for CNN based Image Denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef]
  37. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  38. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2013, 20, 209–212. [Google Scholar] [CrossRef]
  39. Athar, S.; Wang, Z. A Comprehensive Performance Evaluation of Image Quality Assessment Algorithms. IEEE Access 2019, 7, 140030–140070. [Google Scholar] [CrossRef]
  40. Clark, K. Data from the National Lung Screening Trial (NLST) [Data set]. In The Cancer Imaging Archive; TCIA: Manchester, NH, USA, 2013. [Google Scholar] [CrossRef]
  41. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and Operating a Public Information Repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Comparison diagram between the model reference and the proposed model, including a slight modification to fit the denoising problem. The convolution and transpose convolution stride one blocks can have more than one layer (n Layers) (i.e., single layer with 3 × 3 filter, dual layers with 3 × 3 filter).
Figure 1. Comparison diagram between the model reference and the proposed model, including a slight modification to fit the denoising problem. The convolution and transpose convolution stride one blocks can have more than one layer (n Layers) (i.e., single layer with 3 × 3 filter, dual layers with 3 × 3 filter).
Sensors 22 07031 g001
Figure 2. Each block in Figure 1 can become one of the options during testing. The naming in this paper is defined as Single N, Dual N, and Triple N.
Figure 2. Each block in Figure 1 can become one of the options during testing. The naming in this paper is defined as Single N, Dual N, and Triple N.
Sensors 22 07031 g002
Figure 3. Synthetic noise mask from the sample (a) with the Bowtie filter; (b) without the Bowtie filter.
Figure 3. Synthetic noise mask from the sample (a) with the Bowtie filter; (b) without the Bowtie filter.
Sensors 22 07031 g003
Figure 4. Training result between a normal-modified U-Net and the double stacked-modified U-Net on the same double layers with 3 × 3 filters. Rather than the epoch number, the X-axis represents time in minutes to show the training length for 300 epochs.
Figure 4. Training result between a normal-modified U-Net and the double stacked-modified U-Net on the same double layers with 3 × 3 filters. Rather than the epoch number, the X-axis represents time in minutes to show the training length for 300 epochs.
Sensors 22 07031 g004
Figure 5. Training results of the double stacked-modified U-Net on different layers: single, dual, and triple layers with 3 × 3 filters. Rather than the epoch number, the X-axis represents time in minutes to indicate the training length for 300 epochs.
Figure 5. Training results of the double stacked-modified U-Net on different layers: single, dual, and triple layers with 3 × 3 filters. Rather than the epoch number, the X-axis represents time in minutes to indicate the training length for 300 epochs.
Sensors 22 07031 g005
Figure 6. Training results of the double stacked-modified U-Net compare a dual layer with 3 × 3 filters and a single layer with 5 × 5 filters. Rather than the epoch number, the X-axis represents time in minutes to show the training length for 300 epochs.
Figure 6. Training results of the double stacked-modified U-Net compare a dual layer with 3 × 3 filters and a single layer with 5 × 5 filters. Rather than the epoch number, the X-axis represents time in minutes to show the training length for 300 epochs.
Sensors 22 07031 g006
Figure 7. Training results between different filter numbers for 300 epochs with the X-axis representing training time.
Figure 7. Training results between different filter numbers for 300 epochs with the X-axis representing training time.
Sensors 22 07031 g007
Figure 8. Comparison chart on NIQE scores between denoising methods. The X-axis represents the NIQE scores of the low-dose images in groups, while the Y-axis represents the amount of NIQE score improvement. Greater improvement indicates a better image quality.
Figure 8. Comparison chart on NIQE scores between denoising methods. The X-axis represents the NIQE scores of the low-dose images in groups, while the Y-axis represents the amount of NIQE score improvement. Greater improvement indicates a better image quality.
Sensors 22 07031 g008
Figure 9. Comparison chart on PSNR scores between denoising methods. The X-axis represents the PSNR scores of the low-dose images in groups, while the Y-axis represents the amount of PSNR score improvement. Greater improvement indicates a better image quality.
Figure 9. Comparison chart on PSNR scores between denoising methods. The X-axis represents the PSNR scores of the low-dose images in groups, while the Y-axis represents the amount of PSNR score improvement. Greater improvement indicates a better image quality.
Sensors 22 07031 g009
Figure 10. Comparison chart on SSIM scores between denoising methods. The X-axis represents the SSIM scores of the low-dose images in groups, while the Y-axis represents the amount of PSNR score improvement. Greater improvement indicates a better image quality.
Figure 10. Comparison chart on SSIM scores between denoising methods. The X-axis represents the SSIM scores of the low-dose images in groups, while the Y-axis represents the amount of PSNR score improvement. Greater improvement indicates a better image quality.
Sensors 22 07031 g010
Figure 11. (a) Full dose scan image, (b) synthetic low-dose image at PSNR 50.7883 dB, and (c) noise mask for the low-dose image at 10% of the maximum pixel value.
Figure 11. (a) Full dose scan image, (b) synthetic low-dose image at PSNR 50.7883 dB, and (c) noise mask for the low-dose image at 10% of the maximum pixel value.
Sensors 22 07031 g011
Figure 12. Image comparison between denoising methods on the upper thorax. The first row is the denoising result while the second row is the noise mask. (a) NLM, 51.3995 dB; (b) Gaussian, 57.4996 dB; (c) BM3D, 53.7341 dB; (d) SegNet, 61.4247 dB; (e) DeConvNet, 61.4975 dB; (f) RedNet, 61.1857 dB; (g) U-Net, 61.6218 dB; and (h) S-Dual 3, 62.3070 dB.
Figure 12. Image comparison between denoising methods on the upper thorax. The first row is the denoising result while the second row is the noise mask. (a) NLM, 51.3995 dB; (b) Gaussian, 57.4996 dB; (c) BM3D, 53.7341 dB; (d) SegNet, 61.4247 dB; (e) DeConvNet, 61.4975 dB; (f) RedNet, 61.1857 dB; (g) U-Net, 61.6218 dB; and (h) S-Dual 3, 62.3070 dB.
Sensors 22 07031 g012
Figure 13. Image comparison between denoising methods in the middle lung area. The first row is the denoising result while the second row is the noise mask. (a) NLM, 59.7933 dB; (b) Gaussian, 55.5066 dB; (c) BM3D, 59.8350 dB; (d) SegNet, 61.3051 dB; (e) DeConvNet, 61.3086 dB; (f) RedNet, 61.6290 dB; (g) U-Net, 61.6312 dB; and (h) S-Dual 3, 62.0255 dB.
Figure 13. Image comparison between denoising methods in the middle lung area. The first row is the denoising result while the second row is the noise mask. (a) NLM, 59.7933 dB; (b) Gaussian, 55.5066 dB; (c) BM3D, 59.8350 dB; (d) SegNet, 61.3051 dB; (e) DeConvNet, 61.3086 dB; (f) RedNet, 61.6290 dB; (g) U-Net, 61.6312 dB; and (h) S-Dual 3, 62.0255 dB.
Sensors 22 07031 g013
Figure 14. Image comparison between denoising methods around the diaphragm area. The first row is the denoising result while the second row is the noise mask. (a) NLM, 54.2858 dB; (b) Gaussian, 56.2689 dB; (c) BM3D, 58.3729 dB; (d) SegNet, 61.1182 dB; (e) DeConvNet, 61.1210 dB; (f) RedNet, 61.3288 dB; (g) U-Net, 61.2637 dB; and (h) S-Dual 3, 61.8016 dB.
Figure 14. Image comparison between denoising methods around the diaphragm area. The first row is the denoising result while the second row is the noise mask. (a) NLM, 54.2858 dB; (b) Gaussian, 56.2689 dB; (c) BM3D, 58.3729 dB; (d) SegNet, 61.1182 dB; (e) DeConvNet, 61.1210 dB; (f) RedNet, 61.3288 dB; (g) U-Net, 61.2637 dB; and (h) S-Dual 3, 61.8016 dB.
Sensors 22 07031 g014
Figure 15. Comparison chart of NIQE scores from different CNN models with a new data set.
Figure 15. Comparison chart of NIQE scores from different CNN models with a new data set.
Sensors 22 07031 g015
Figure 16. Comparison chart of PSNR scores from different CNN models with a new data set.
Figure 16. Comparison chart of PSNR scores from different CNN models with a new data set.
Sensors 22 07031 g016
Figure 17. Comparison chart of SSIM scores from different CNN models with a new data set.
Figure 17. Comparison chart of SSIM scores from different CNN models with a new data set.
Sensors 22 07031 g017
Figure 18. Image comparison between the screening image and after denoising. (a) Screening; (b) NLM; (c) Gaussian; (d) SegNet; (e) DeConvNet; (f) RedNet; (g) U-Net; and (h) S-Dual 3.
Figure 18. Image comparison between the screening image and after denoising. (a) Screening; (b) NLM; (c) Gaussian; (d) SegNet; (e) DeConvNet; (f) RedNet; (g) U-Net; and (h) S-Dual 3.
Sensors 22 07031 g018
Table 1. Quality scores between various network setting on 300 training epochs.
Table 1. Quality scores between various network setting on 300 training epochs.
NetworkFilterNIQEAVPSNRAVSSIMAVTraining Time (min)
Single2 × [3 × 3]7.67210.3770.00890544
Stacked1 × [3 × 3]7.69510.4320.00892816
Stacked2 × [3 × 3]7.66710.5570.008951229
Stacked3 × [3 × 3]7.72710.5700.008952191
Stacked1 × [5 × 5]7.88610.5490.008951564
Stacked1 × [7 × 7]7.92810.5820.008963528
Table 2. Training time (in minutes) comparison between CNN models at 300 epochs.
Table 2. Training time (in minutes) comparison between CNN models at 300 epochs.
S-Dual [3 × 3]SegNetDeConvNetRedNetUNet
Training1229644576764406
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gunawan, R.; Tran, Y.; Zheng, J.; Nguyen, H.; Chai, R. Image Recovery from Synthetic Noise Artifacts in CT Scans Using Modified U-Net. Sensors 2022, 22, 7031. https://doi.org/10.3390/s22187031

AMA Style

Gunawan R, Tran Y, Zheng J, Nguyen H, Chai R. Image Recovery from Synthetic Noise Artifacts in CT Scans Using Modified U-Net. Sensors. 2022; 22(18):7031. https://doi.org/10.3390/s22187031

Chicago/Turabian Style

Gunawan, Rudy, Yvonne Tran, Jinchuan Zheng, Hung Nguyen, and Rifai Chai. 2022. "Image Recovery from Synthetic Noise Artifacts in CT Scans Using Modified U-Net" Sensors 22, no. 18: 7031. https://doi.org/10.3390/s22187031

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop