Next Article in Journal
Machine Learning in Evolving Art Styles: A Study of Algorithmic Creativity
Previous Article in Journal
Linear Quadratic Regulator Control of Rotary Inverted Pendulum Using Elvis III Embedded Platform
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Enhanced Adaptive Wiener Filtering for Frequency-Varying Noise with Convolutional Neural Network-Based Feature Extraction †

Graduate Institute of Communication Engineering, National Taiwan University, Taipei 106319, Taiwan
*
Author to whom correspondence should be addressed.
Presented at the 2024 IEEE 6th Eurasia Conference on IoT, Communication and Engineering, Yunlin, Taiwan, 15–17 November 2024.
Eng. Proc. 2025, 92(1), 47; https://doi.org/10.3390/engproc2025092047
Published: 2 May 2025
(This article belongs to the Proceedings of 2024 IEEE 6th Eurasia Conference on IoT, Communication and Engineering)

Abstract

:
Denoising has long been a challenge in image processing. Noise appears in various forms, such as additive white Gaussian noise (AWGN) and Poisson noise across different frequencies. This study aims to denoise images without prior knowledge of the noise distribution. First, we estimate the noise power in the frequency domain to approximate the local signal-to-noise ratio (SNR) and guide an adaptive Wiener filter. The initial denoised result is obtained by assembling the locally filtered patches. However, since the Wiener filter is a low-pass filter, it can remove fine details along with the noise. To overcome this limitation, we post-process the noise and interpolate it between the denoised and original noisy patches to enhance the denoised image. We also mask the frequency domain to avoid grid-like artifacts. Additionally, we introduce a convolutional neural network-based refinement technique to the spatial domain to recover latent textures lost during denoising. The method presents the effectiveness of masking and feature extraction.

1. Introduction

The ultimate goal of the mathematical model for image denoising [1,2,3,4,5,6] is to recover the original image x m , n based on the observed noisy image B m , n , as in (1), or based on some learning models [7].
B m , n = x m , n + N m , n
First, we provide a brief review of the adaptive Wiener filter [8,9], with the corresponding flowchart (Figure 1). The method can be roughly divided into three stages.
The first stage is “smoother patch determination”, where relatively smoother patch gradients are collected and transformed to the frequency domain using the fast Fourier transform (FFT). In the second stage, the Fourier-transformed patch gradients are regarded as noise power in the frequency domain. The estimated noise E p , q is obtained by weighting these patches, a process known as sparse band determination. In the final stage, a 2D first-order function N e s t p , q is used to approximate E p , q . With N e s t p , q , the frequency-varying adaptive Wiener filter can be determined for all patches in the noisy image.
However, since the Wiener filter is a signal-to-noise ratio (SNR)-based low-pass filter, it removes texture and details that have less power than the estimated noise. Figure 2, Figure 3, Figure 4 and Figure 5 illustrate the possible drawbacks of the method in Ref. [9]. For denoised patches, the number of nonzero components in a 9 × 9 FFT is typically less than 10, regardless of patch variance, whereas for sharp image patches, the number of nonzero components exceeds 40. Additionally, since the adaptive filter makes several components equal zero, the resulting patches can exhibit grid-like noise. Therefore, we address these problems and refine the details removed by the adaptive Wiener filter in this study.

2. Proposed Algorithm

2.1. Adaptive Masking in Frequency Domain

In this phase, features are recovered by interpolating the result with the noisy patch [9]. The interpolation ratio is determined as
P n e w = M I · P d e n o i s e d + 1 M I · P n o i s y
where M I = e i d x 81 .
We specify the top-left coordinate of a patch p 1 ,   q 1 as (1,1), with q i increasing from left to right and p i increasing from top to bottom. The parameter idx in (1) is calculated as idx = p i · q i . The Wiener-filtered patch takes a larger proportion in the low-frequency component. In other words, we preserve information in the high-frequency components from the noisy patch. After the interpolation, we apply an overall mask M O to the new patch P n e w . Ideally, P n e w is smoothed in the low-variance area, and its details are preserved in the high-variance area. However, when dealing with noisy image patches, the variance deviates from the true variance. Therefore, we use the number of nonzero coefficients in the Wiener filter H p , q , a , b [9] as a lookup parameter, denoted as n z . Here p , q are indices in the frequency domain, while a , b represent the spatial location of current patch P in the noisy image. When the number of nonzero components is small, we classify the patch as a smooth area, and most high-frequency components are forced to zero by M O (3). Conversely, if the Wiener filter H has few zero components, the filter performs well, making M O an all-pass filter (Figure 6, Figure 7 and Figure 8).
P o u t = M O · P n e w   where   M O = e idx 2 n z 3

2.2. Detail Refinement

After recovering high-frequency components via masking in the Fourier domain, the remaining challenge is preserving fine patch details spatially. For instance, in the case of a brick wall, the gaps between bricks cannot be captured by masking alone. These details are inherently spatial. Therefore, we aim to extract subtle features by training a convolutional neural network (CNN) based on the following architecture.
The CNN model takes a 50 × 50 × 2 input, with the target being a 50 × 50 × 2 patch. The first channel of the input is the noisy image patch, and the second channel is the result of the adaptive Wiener filter. We train the combination against the true image patch. In training, we use MSE as the loss function (4).
l n = 1 N i = 1 N f θ x i y i 2
The ground truths of the training data are collected from ImageNet-A dataset [7]. For each image Y i , we apply AWGN with a standard deviation uniformly sampled in the range of [0, 30]. After applying the proposed blind denoising, we randomly select patches to form x i and the corresponding patch in Y i is y i . The idea here is to consider both the MMSE-denoised result—i.e., the adaptive Wiener filter—and the original noisy image patch with possible latent details. This model extracts spatial features from both the Wiener-filtered result and the original noisy image patch (Figure 9).

3. Results

We assessed the performance of the developed model in this study in four different scenarios: AWGN, Poisson noise, and two frequency-varying noise distributions. We compared the results with those obtained using IRCNN [6], BM3D, and the adaptive Wiener filter. For a restored image y m , n and its corresponding ground truth x m , n , the performance metric used was peak SNR (PSNR) (5).
P S N R = 10   log 10 255 2 1 M N m = 1 M n = 1 N y m , n x m , n 2
The test set comprises five real-world images. For each scenario, we present the estimated noise spectrum and enlarged views of the results.

3.1. AWGN

The noise N m , n is normally distributed with a mean of 0 and a standard deviation of 15, which is represented as
N m , n   ~   N 0 ,   15 2
The noise distribution for the model as in (6) is plotted as in Figure 10. Then, in Table 1, the restoration performance for the image degraded by the noise in (6) is shown. Some visual reconstruction results are shown in Figure 11.

3.2. Poisson Noise

When the noise follows a Poisson distribution, (7) is used.
N m , n   ~   P o i s s o n x m , n
The distribution of the Poisson noise in (7) is shown in Figure 12. The restoration performance for the images degraded by the noise in (7) is shown in Table 2 and Figure 13 show some visual results of image restoration.

3.3. Frequency-Varying Noise (Peak-like)

Given the noise spectrum of AWGN in Section A, denoted as N 0 / 2 , we apply a peak-like frequency-varying noise characterized by
E N p , q = N 0 ( p 1 M + q 1 N )
where the M and N denote the height and width of the image, and N p , q represents the M × N point FFT of the AWGN. The indices are symmetric, with p1 = min(p, M−p), and q1 = min(q, N−q). The noise model in (8) is plotted in Figure 14. The restoration performance for the images interfered by the noise in (8) is shown as in Table 3. In Figure 15, some visual results of the reconstructed images are shown.

3.4. Frequency-Varying (Valley-like) Noise

Using the same notation as in Section C, the noise distribution exhibits a valley-like surface in the frequency domain, described by:
E N p , q = N 0 ( 1 0.4 · p 1 M 0.4 · q 1 N )
The noise model in (9) is plotted in Figure 16. The restoration performance for the images interfered by the noise in (9) is shown as in Table 4. In Figure 17, some visual results of the reconstructed images are shown.

3.5. Average Runtime per Image

Taking the image ‘Castle’ as an example, with a size of 1006 × 782 × 3 , BM3D [5] takes 47 s to restore the image, while IRCNN [6] takes only 2 s, and the adaptive Wiener filter [9] requires 5 s. The proposed postprocessing takes 15 s.

4. Conclusions

In this study, we improved the adaptive Wiener filter [9]. For the original adaptive Wiener filter, a noise-free image must be sparse in some space-frequency domain for the determination of the noise distribution based on smooth patches. The noise model is then derived from this distribution, enabling the automatic design of an adaptive Wiener filter for frequency-varying noise spectra. However, the original adaptive Wiener filter tends to erase most of the details in the denoised image and may introduce grid-like noise. To address these issues, we interpolated the filtered result with the original patch to recover a certain amount of high-frequency components. After interpolation, a mask was applied to remove the blocking effect. To further enhance performance, we trained a CNN to retrieve the details lost during filtering. This method represents a trade-off between processing time and performance. Although the proposed detail refinement CNN was specifically trained to extract features under AWGN conditions, the developed method demonstrates the best average performance under Poisson noise and frequency-varying noise when compared to other approaches. The method effectively processes various noise distributions and achieves these results within a reasonable time frame.

Author Contributions

Conceptualization, C.-L.L. and J.-J.D.; methodology, C.-L.L.; software, C.-L.L.; validation, C.-L.L.; formal analysis, C.-L.L. and D.-Y.L.; investigation, J.-J.D.; resources, C.-L.L.; data curation, C.-L.L.; writing—original draft preparation, C.-L.L.; writing—review and editing, D.-Y.L. and J.-J.D.; visualization, C.-L.L. and D.-Y.L.; supervision, J.-J.D.; project administration, J.-J.D.; funding acquisition, J.-J.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science and Technology Council, Taiwan, under the contract of NSTC 113-2221-E-002-146.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lebrun, M. An analysis and implementation of the BM3D image denoising method. Image Process. Line 2012, 2, 175–213. [Google Scholar] [CrossRef]
  2. Pyatykh, S.; Hesser, J.; Zheng, L. Image noise level estimation by principal component analysis. IEEE Trans. Image Process. 2012, 22, 687–699. [Google Scholar] [CrossRef] [PubMed]
  3. Liu, X.; Tanaka, M.; Okutomi, M. Single-image noise level estimation for blind denoising. IEEE Trans. Image Process. 2013, 22, 5226–5237. [Google Scholar] [CrossRef]
  4. Ding, J.J.; Chang, J.Y.; Liao, C.L.; Tsai, Z.H. Image deblurring using local Gaussian models based on noise and edge distribution estimation. In Proceedings of the TENCON 2021—2021 IEEE Region 10 Conference, Auckland, New Zealand, 7–10 December 2021; pp. 714–719. [Google Scholar]
  5. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-d transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  6. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning deep CNN denoiser prior for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3929–3938. [Google Scholar]
  7. Hendrycks, D.; Zhao, K.; Basart, S.; Steinhardt, J.; Song, D. Natural adversarial examples. In Proceedings of the IEEE/CVT Conf. Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 15262–15271. [Google Scholar]
  8. Zhang, X. Image denoising using local Wiener filter and its method noise. Optik 2016, 127, 6821–6828. [Google Scholar] [CrossRef]
  9. Ding, J.J.; Liao, C.L. Image denoising based on the noise prediction model using smooth patch and sparse domain priors. In Proceedings of the International Workshop on Advanced Imaging Technology (IWAIT), Jeju, Republic of Korea, 9–11 January 2023; Volume 12592, pp. 170–175. [Google Scholar]
Figure 1. Flowchart of original adaptive Wiener filter.
Figure 1. Flowchart of original adaptive Wiener filter.
Engproc 92 00047 g001
Figure 2. Grid-like effect. The left subfigure is the output of patch smoothing from the method in [9] and the right one is the result after masking.
Figure 2. Grid-like effect. The left subfigure is the output of patch smoothing from the method in [9] and the right one is the result after masking.
Engproc 92 00047 g002
Figure 3. A limitation of the adaptive Wiener filter. The red box indicates the gap between the distributions of filtered patches and sharp patches.
Figure 3. A limitation of the adaptive Wiener filter. The red box indicates the gap between the distributions of filtered patches and sharp patches.
Engproc 92 00047 g003
Figure 4. The flowchart of the proposed algorithm. The red circle indicates the post-processing methods of adaptive filtering and detail refinement.
Figure 4. The flowchart of the proposed algorithm. The red circle indicates the post-processing methods of adaptive filtering and detail refinement.
Engproc 92 00047 g004
Figure 5. Interpolation mask M I . A warm color indicates a larger value.
Figure 5. Interpolation mask M I . A warm color indicates a larger value.
Engproc 92 00047 g005
Figure 6. CNN architecture of detail refinement.
Figure 6. CNN architecture of detail refinement.
Engproc 92 00047 g006
Figure 7. M O coefficient of n z number (left) 2 , (middle) 10 , and (right) 70 . A warm color indicates a larger value.
Figure 7. M O coefficient of n z number (left) 2 , (middle) 10 , and (right) 70 . A warm color indicates a larger value.
Engproc 92 00047 g007
Figure 8. Training for detail refinement.
Figure 8. Training for detail refinement.
Engproc 92 00047 g008
Figure 9. A visualization of the effect of detail refinement.
Figure 9. A visualization of the effect of detail refinement.
Engproc 92 00047 g009
Figure 10. N e s t p , q for AWGN. A warm color means a larger value.
Figure 10. N e s t p , q for AWGN. A warm color means a larger value.
Engproc 92 00047 g010
Figure 11. Enlarged visualization of restoration for AWGN.
Figure 11. Enlarged visualization of restoration for AWGN.
Engproc 92 00047 g011
Figure 12. N e s t p , q for Poisson noise. A warm color means a larger value.
Figure 12. N e s t p , q for Poisson noise. A warm color means a larger value.
Engproc 92 00047 g012
Figure 13. Enlarged visualization of restoration for Poisson noise.
Figure 13. Enlarged visualization of restoration for Poisson noise.
Engproc 92 00047 g013
Figure 14. N e s t p , q for (8). A warm color means a larger value.
Figure 14. N e s t p , q for (8). A warm color means a larger value.
Engproc 92 00047 g014
Figure 15. Enlarged visualization of restoration for peak-like frequency varying noise.
Figure 15. Enlarged visualization of restoration for peak-like frequency varying noise.
Engproc 92 00047 g015
Figure 16. N e s t p , q for (9). A warm color means a larger value.
Figure 16. N e s t p , q for (9). A warm color means a larger value.
Engproc 92 00047 g016
Figure 17. Enlarged visualization of restoration for valley-like frequency-varying noise.
Figure 17. Enlarged visualization of restoration for valley-like frequency-varying noise.
Engproc 92 00047 g017
Table 1. The restoration performance of images degraded by the noise model in (6).
Table 1. The restoration performance of images degraded by the noise model in (6).
PSNRCastlePillar Bridge Cottage Snow Field Average
Adaptive Wiener [9]29.27331.70329.88630.63928.76629.273
BM3D [5]27.01028.70427.20027.71427.00327.010
IRCNN [6]26.79927.87427.05927.25527.04826.799
Proposed30.31031.96530.64030.98129.38130.310
Table 2. The restoration performance of images degraded by the noise model in (7).
Table 2. The restoration performance of images degraded by the noise model in (7).
PSNRCastlePillar Bridge Cottage Snow Field Average
Adaptive Wiener [9]32.88934.54832.65832.28631.66232.809
BM3D [5]32.66535.10832.70932.68432.14233.062
IRCNN [6]31.86634.43632.05532.22731.52432.421
Proposed33.09535.09533.15432.91032.09933.271
Table 3. The restoration performance of images degraded by the noise model in (8).
Table 3. The restoration performance of images degraded by the noise model in (8).
PSNRCastlePillar Bridge Cottage Snow Field Average
Adaptive Wiener [9]27.77829.17628.49128.41527.27828.227
BM3D [5]21.51922.80721.83921.96521.99022.024
IRCNN [6]23.92724.84124.25124.32324.26024.321
Proposed28.63728.90028.81628.18126.88828.285
Table 4. The restoration performance of images degraded by the noise model in (9).
Table 4. The restoration performance of images degraded by the noise model in (9).
PSNRCastlePillar Bridge Cottage Snow Field Average
Adaptive Wiener [9]31.33833.56131.48832.29330.64431.865
BM3D [5]31.86533.61231.96132.59431.10432.227
IRCNN [6]30.05131.09730.32830.57530.07530.425
Proposed32.03233.51332.06832.49831.21532.265
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liao, C.-L.; Ding, J.-J.; Lu, D.-Y. Enhanced Adaptive Wiener Filtering for Frequency-Varying Noise with Convolutional Neural Network-Based Feature Extraction. Eng. Proc. 2025, 92, 47. https://doi.org/10.3390/engproc2025092047

AMA Style

Liao C-L, Ding J-J, Lu D-Y. Enhanced Adaptive Wiener Filtering for Frequency-Varying Noise with Convolutional Neural Network-Based Feature Extraction. Engineering Proceedings. 2025; 92(1):47. https://doi.org/10.3390/engproc2025092047

Chicago/Turabian Style

Liao, Chun-Lin, Jian-Jiun Ding, and De-Yan Lu. 2025. "Enhanced Adaptive Wiener Filtering for Frequency-Varying Noise with Convolutional Neural Network-Based Feature Extraction" Engineering Proceedings 92, no. 1: 47. https://doi.org/10.3390/engproc2025092047

APA Style

Liao, C.-L., Ding, J.-J., & Lu, D.-Y. (2025). Enhanced Adaptive Wiener Filtering for Frequency-Varying Noise with Convolutional Neural Network-Based Feature Extraction. Engineering Proceedings, 92(1), 47. https://doi.org/10.3390/engproc2025092047

Article Metrics

Back to TopTop