Next Article in Journal
Developing a Novel Prosthetic Hand with Wireless Wearable Sensor Technology Based on User Perspectives: A Pilot Study
Previous Article in Journal
Deep Learning Models to Reduce Stray Light in TJ-II Thomson Scattering Diagnostic
Previous Article in Special Issue
COVID-19 Hierarchical Classification Using a Deep Learning Multi-Modal
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multibranch Wavelet-Based Network for Image Demoiréing

1
Department of Electrical Engineering, National Taiwan Normal University, Taipei 10610, Taiwan
2
Department of Electrical Engineering, National Sun Yat-sen University, Kaohsiung 80424, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(9), 2762; https://doi.org/10.3390/s24092762
Submission received: 20 March 2024 / Revised: 24 April 2024 / Accepted: 24 April 2024 / Published: 26 April 2024
(This article belongs to the Special Issue Advanced Deep Learning for Biomedical Sensing and Imaging)

Abstract

:
Moiré patterns caused by aliasing between the camera’s sensor and the monitor can severely degrade image quality. Image demoiréing is a multi-task image restoration method that includes texture and color restoration. This paper proposes a new multibranch wavelet-based image demoiréing network (MBWDN) for moiré pattern removal. Moiré images are separated into sub-band images using wavelet decomposition, and demoiréing can be achieved using the different learning strategies of two networks: moiré removal network (MRN) and detail-enhanced moiré removal network (DMRN). MRN removes moiré patterns from low-frequency images while preserving the structure of smooth areas. DMRN simultaneously removes high-frequency moiré patterns and enhances fine details in images. Wavelet decomposition is used to replace traditional upsampling, and max pooling effectively increases the receptive field of the network without losing the spatial information. Through decomposing the moiré image into different levels using wavelet transform, the feature learning results of each branch can be fully preserved and fed into the next branch; therefore, possible distortions in the recovered image are avoided. Thanks to the separation of high- and low-frequency images during feature training, the proposed two networks achieve impressive moiré removal effects. Based on extensive experiments conducted using public datasets, the proposed method shows good demoiréing validity both quantitatively and qualitatively when compared with the state-of-the-art approaches.

1. Introduction

Moiré pattern is a type of visual artifact that is noted when interference occurs between overlapping patterns, such as the grid of display elements and camera sensors, greatly reducing the overall perceptual quality of the image. With the development of digital imaging technology and the widespread use of digital cameras and displays, images with moiré patterns have become increasingly common. The appearance of moiré patterns is quite diverse and complex, forming different stripes, curves, or ripples in space, and the patterns are also accompanied by color changes superimposed on the image. Due to its complexity, moiré pattern removal remains a challenging problem that is attracting increasing attention from the research community. Unlike other image restoration tasks, such as denoising, dehazing, and super-resolution, is difficult to synthesize the moiré effect it into a formation model due to its complex nature. Moiré patterns exhibit irregularities in shape and color over a wide range of frequencies. Figure 1 shows moiré textures of different scales, frequencies, and colors.
Moiré patterns are caused by frequency aliasing. Anti-aliasing optical filters (low-pass filters) can solve this problem; however, it is easy to over-smooth the image, resulting in a blurred image [1]. Several studies have used signal processing techniques to remove moiré patterns [2]. Wei et al. used local median filters and Gaussian notch filters to remove moiré patterns [3]; however, the process is a semi-automatic and requires manual parameter adjustment. Liu et al. proposed a low-rank and sparse matrix decomposition model for image demoiréing; however, the complex pattern still cannot be effectively eliminated. Yang et al. [4] used a low-rank sparse matrix decomposition model to eliminate moiré patterns in the green channel; guided filtering is applied to the extracted green channel texture layer to remove moiré patterns in the red and blue channels. However, it is unable to process screen captured images. Yang et al. [5] proposed a demoiréing method by applying Layer Decomposition Polyphase Component (LDPC) to screen capture images, which results in high computational complexity and excessive smoothing of image details.
In order to achieve better moiré removal performance, a learning framework for single image moiré removal has been presented in [6,7,8,9,10,11,12,13,14,15,16] based on the deep study of moiré patterns from a large dataset. Relying on the rapid development of deep learning with great success in numerous perceptual tasks, especially in image restoration, several deep learning-based single image moiré removal frameworks have been presented. They have shown that moiré-relevant features are automatically learned using a deep convolutional neural network to achieve better results in image moiré removal.
This paper proposes to remove moiré patterns in the frequency domain by initially decomposing the moiré image into distinct frequency bands using discrete wavelet transform (DWT). Moiré patterns come in many forms and exist in both high- and low-frequency images, so different methods are required to effectively remove moiré patterns. Sub-band images, the high- and low-frequency images obtained through wavelet decomposition, are then fed to different branches, which use networks for end-to-end learning to effectively remove different levels of moiré patterns. The low-frequency sub-images at each decomposition level are processed using the moiré removal network (MRN) to eliminate moiré effects in the low-frequency band, while preserving background details. The high-frequency sub-images are treated with detail-enhanced moiré removal network (DMRN) to remove moiré patterns, effectively retaining high-frequency details. The inherent reversibility of the wavelet transformation can prevent the loss of image information during the upsampling and downsampling in learning procedures, thus ensuring a sufficient receptive field. Therefore, the proposed method is expected to better learn moiré patterns than current deep learning-based image demoiréing approaches. Furthermore, with the integration of a multibranch process, better demoiréing results will be produced. Experimental results show that our method outperforms the state-of-the-art moiré removal methods in terms of quantitative and qualitative comparisons.
The remainder of the paper is organized as follows. In Section 2, we provide an overview and discuss work related to moiré pattern removal. Section 3 explains the proposed framework of the moiré removal network. In Section 4, we present the experimental results. Finally, concluding remarks are made in Section 5.

2. Background Review

This section provides an overview of the most relevant work on image demoiréing. First, image restoration using deep learning techniques is briefly reviewed. Then, we review some deep learning-based demoiréing methods that inspired the design of our proposed deep model. In addition, frequency-based demoiréing methods are also studied in this section.

2.1. Image Restoration

Traditional image restoration methods are usually based on manual feature extraction and models, but these methods may be limited when dealing with complex image problems. Recently, deep learning methods have received a great deal of attention in almost all fields of image processing, especially for image restoration tasks such as image denoising, deblurring, dehazing, and super-resolution. In deep learning, some famous structures such as U-Net [17], recurrent neural networks [18], multi-stage network [19], MIMO-UNet [20], and encoder-decoder transformer [21] have performed well in image restoration tasks. These methods are often able to learn effective feature representations from large amounts of data, enabling them to better handle image restoration problems [22,23,24,25,26,27,28,29]. In addition, mechanisms such as attention modules, residual learning, and multiscale representation are introduced into the deep learning architecture to further improve the efficiency of the restoration model. Through deep learning, the model can better understand and learn the internal structure of the image, thereby achieving more accurate and efficient image restoration.

2.2. Moiré Pattern Removal Using Deep Learning

Several image demoiréing methods have been proposed to treat image demoiréing as an image restoration problem, capable of solving a wider range of moiré pattern types. Sun et al. [6] proposed a multiscale CNN (DMCNN) to remove multi-frequency moiré patterns and established a large-scale benchmark dataset based on ImageNet [30] called the TIP dataset. Liu et al. [7] constructed a convolutional neural network with coarse- and fine-scale options. In the coarse-scale network, the input image is first downsampled and passed through stacked residual blocks to remove moiré patterns. The fine network upsamples the processed low-resolution image back to its original resolution. He et al. [8] proposed MopNet, which was specifically designed for moiré patterns, including multiscale feature aggregation for resolving complex frequencies, edge detection to explore the degree of imbalance between color channels, and a perceptual classifier designed to represent the appearance of moiré stripe diversity. Yu et al. [9] introduced a method to demonstrate moiré patterns in 4k ultra high-definition images (UHDM). Yang et al. [10] proposed a parallel multiscale architecture with residual blocks. To efficiently exploit the relationship between different feature maps, this architecture employs an information exchange module (IEM) and a final feature fusion layer (FFF) to continuously exchange information throughout the network. To eliminate multiscale moiré patterns, they built a semantically aligned scale-aware module. Yang et al. [11] proposed a new strategy to flexibly model moiré patterns, enabling the construction of well-aligned training triplets. These triplets were utilized by the translation network MoireDet for moiré pattern detection. Niu et al. [12] proposed a progressive moiré removal and texture complementation network. Their method used a multi-stream framework to gradually remove moiré patterns from low- to high-resolution images. They proposed a progressive texture complementation block to fuse information from different streams, improving moiré removal and detail complementation. Nguyen et al. [13] proposed a multiscale guided screenshot demoiréing algorithm to understand moiré frequency correlations. Initially, they extracted multiscale features from the input image. Then, they introduced FTB and MGRB to capture and remove moiré artifacts at each scale by modulating features. Although some published algorithms perform well, the problem remains largely unresolved due to significant changes in the frequency, shape, and color of moiré patterns.

2.3. Frequency Domain Learning

Frequency analysis has always been a powerful tool in image processing that can effectively use frequency domain information to significantly improve the performance of image restoration. Transform operations are often used as image pre- or post-processing, and downsampling or upsampling processes are also used to design deep networks. Some methods embed frequency information into the network structure to exploit its availability [31,32,33,34,35,36,37]. Liu et al. [36] proposed a framework based on Generative Adversarial Networks (GAN) to embed facial attribute vectors in the generator and discriminator. This method guides the model to generate facial images with corresponding input attributes. To enhance aging details, they employed wavelet packet transform to effectively extract features at multiple scales in frequency space. Li et al. [37] proposed a method to transform DWT and IDWT into a general network layer, replacing the traditional upsampling and downsampling operations. The wavelet integrated convolutional network (WaveCNet) was introduced for image classification and achieves higher image classification accuracy and better noise suppression. Several works focus on moiré pattern removal using frequency domain processes. Zheng et al. [14] proposed a learnable bandpass filter (MBCNN) to first learn the frequency to eliminate moiré patterns and designed a tone mapping strategy for color restoration. Although removing moiré patterns using frequency domain processing has achieved good results in deep learning networks, over-smoothing high frequency details and distinguishing low frequency patterns from background colors in complex background images remain a challenge. Liu et al. [15] proposed a two-branch wavelet network with spatial attention mechanism for eliminating moiré patterns, and they utilized dense convolution and dilated convolution to achieve large-scale sensing fields. Sun et al. [16] replaced conventional downsampling and upsampling with DWT and IDWT and introduced an efficient attention fusion module to enable the network to dynamically prioritize feature information, with a specific focus on moiré details. Wavelet transform has several advantages, including time–frequency localization, multi-resolution analysis, and noise reduction. For time–frequency localization, wavelets can be used to identify specific features in a signal, such as transient events or harmonics, that may not be visible with other time–frequency localization methods. Moiré patterns are characterized by higher harmonics and transient events, and wavelet transform is useful for identifying these specific features. The wavelet transform provides a multi-resolution representation of the signal, which means that the moiré signal can be analyzed at different resolution levels so that we can use different strategies to process it. Moreover, wavelet transform can separate signals and noise in different frequency bands, making it easier to suppress noise in moiré patterns. The proposed multibranch network combined with wavelet transform is able to extract more meaningful and sparse moiré pattern representations.

3. Proposed Method

In this section, we first explain the embedding method used for the wavelet transform and its inverse. We then introduce two key components: moiré removal network (MRN) and detail-enhanced moiré removal network (DMRN). These two networks play a crucial role in moiré removal and image detail enhancement. Finally, we define the optimization model tailored to our specific loss function. This section aims to provide a concise yet informative overview of the key techniques and components used in our image restoration approach.

3.1. Wavelet-Based Image Decomposition

Wavelet transform decomposes an image into multiple sub-images, which are also wavelet coefficients and can be used as features for different tasks. The traditional DWT is translation variant; therefore, its effectiveness may be negatively affected. We select translation-invariant wavelet transform to ensure the robustness of sub-images. Given an image X , X is decomposed into four sub-band images, called X L L ,   X L H ,   X H L , and X H H using the 2D DWT [38], which consists of four filters, namely f L L ,   f L H ,     f H L ,   and   f H H . During the transformation, these four filters have fixed parameters and use a convolution step size of 2. Taking the Haar wavelet as an example, these four filters are defined as follows:
f L L = 1 1 1 1 ,   f L H = 1 1 1 1 , f H L = 1 1 1 1 ,   f H H = 1 1 1 1 .
DWT reduces the resolution of the input images, making it easier to extract features of image structure and texture. If the size of the image is n × n × 3 , it becomes ( n / 2 ) × ( n / 2 ) × 12 after DWT, and downsampling is also performed to make the feature channel four times the original one. We decompose the moiré image M using the Haar DWT to obtain wavelet bands at three different levels. The sub-images of each level can be denoted as follows:
D W T M = I L L 1 ,   I L H 1 ,   I H L 1 ,   I H H 1 , D W T I L L n = I L L n + 1 ,     I L H n + 1 ,   I H L n + 1 ,   I H H n + 1 ,     n ( 1 , 2 , 3 )
where D W T is the wavelet decomposition operation, and I is the sub-image obtained through D W T decomposition. I L L n ,     I L H n ,   I H L n , and I H H n are the sub-images bands L L ,   L H ,   H L , and H H at level n , respectively. Specifically, L L has the overall context of the image, while L H , H L , and H H encapsulate the detailed information of the image.

3.2. Multibranch Network Modules

As shown in Figure 2, the moiré pattern image is decomposed using DWT into three levels of sub-images. At each level, low-frequency and high-frequency sub-images are processed by the moiré removal network (MRN) and a detail-enhanced moiré removal network (DMRN), respectively, in a process called one branch learning. To enable the network to learn more complex mappings, the high-frequency and low-frequency sub-images processed using DMRN and MRN of the current branch learning will be combined through IDWT, and the results will be sent to the next branch for learning. Here, the number of the branch is equal to that of the DWT decomposition level. For example, at the level n, both high-frequency and low-frequency sub-images are processed using the pipeline process, branch n. The output of branch n includes the low-frequency sub-image of level n − 1, which is processed by branch n − 1 together with the high-frequency sub-image of level n − 1. This looping process continues on all branches until the output size matches the original input, thereby enhancing the network’s ability to capture complex features and patterns. The multibranch learning process can be represented as follows:
R L L n = M R N ( I L L n ) , R L H n ,   R H L n ,   R H H n = D M R N   I L H n ,   I H L n ,   I H H n ,   O u t n = I D W T ( R L L n ,   R L H n ,   R H L n ,   R H H n ) ,   n ( 1 , 2 , 3 )
where R L L n and ( R L H n ,   R H L n ,   R H H n ) are the demoiréd low-frequency and high-frequency images obtained by the n th branch learning, and O u t n is the output image of the n th branch learning through IDWT.

3.3. Moiré Removal Network

The objective of MRN is to remove moiré patterns from the low-frequency image decomposed by DWT while preserving the structure of smooth areas. We extract features through the hierarchical structure, which has always been a common method to solve problems at different scales. The hierarchical structure, which includes two convolutional layers with 3 × 3 filters and 32, 64, and 128 output channels for branches 1, 2, and 3 learning, aims to extract multiscale features within the same level and allow them to interact and dynamically fuse, significantly improving the model’s ability to handle moiré patterns at various scales. Then, eight residual blocks (ResBlocks) [39] are used to prevent the problem of gradient vanish and gradient explosion in deep networks. Squeeze-and-excitation block (SEBlock) [40], the attention module we used, consists of squeeze, excitation, and scale, which is simple yet effective in helping the model learn to focus more on important features. Finally, after passing through eight ResBlocks again, the channel size gradually decreases, finally returning to three channels.

3.4. Detail-Enhanced Moiré Removal Network

The objective of DMRN is to remove moiré patterns from the high-frequency image decomposed by DWT while enhancing the fine details in the image. DenseNet [41] has demonstrated excellent performance in various image restoration tasks; specifically, it effectively exploits hierarchical features by densely connecting multiple residual dense blocks (RDBs). We leverage four RDBs to efficiently train the model simultaneously to learn high-frequency details and eliminate moiré patterns. These RDBs embody residual connections and dense feature extraction, which help capture complex patterns and enhance the model’s ability to process high-frequency information. The combination of multiple RDBs further enables the network to effectively learn and represent complex visual features, ultimately helping to eliminate high-frequency moiré patterns during training as shown in Figure 3. To guide the network to efficiently discern high-frequency details from high-frequency moiré patterns, our network focuses on directly connecting inputs and outputs through skip connections, learning only the residual information between input and output.

3.5. Loss Function

To learn the proposed network, the problem is formulated as a mapping function between a moiré image component and its corresponding moiré-free component. To solve the problem, we intend to learn a mapping function subject to the pixel-wise L 1 loss. The loss function L 1 is minimized using the Adam optimization algorithm [42], which is designed for first-order gradient-based optimization of stochastic objective functions. In the proposed framework, there are three branch learning processes. In each branch, the low-frequency image is denoted by R L L n ,   and the high-frequency image components are denoted by R L H n ,   R H L n , and R H H n . The corresponding moiré-free low-frequency image is denoted by G L L n ,   and the moiré-free high-frequency image is denoted by G L H n ,   G H L n , and G H H n . The loss of all branch learning is expressed as follows:
L b r a n c h = L 1 ( R L L n ,   G L L n ) + n = 1 3 ( L 1 R L H n ,   G L H n + L 1 R H L n ,   G H L n + L 1 R H H n ,   G H H n ) .  
For the demoiréing output image O ^ processed by three branch learning processes, we apply L 1 loss, feature-based perceptual loss L p [43], and Fourier space loss L f t [20]. The demoiréing loss function is denoted as follows:
L D = L 1 O ^ ,   O + L p O ^ ,   O + L f t O ^ ,   O ,   .  
where O is the ground-truth image.
Then, the total loss of the proposed framework is formulated as follows:
L t o t a l = L b r a n c h + L D .  

4. Experimental Result

To quantitatively and qualitatively evaluate the performance of the proposed multibranch wavelet-based single image demoiréing method, the five well-known methods, including the latest learning-based moiré removal methods such as DMCNN [6], WDNet [15], MBCNN [14], EAFM [16], and UHDM [9], are compared. Finally, to investigate the effectiveness of different components in the proposed framework, we conducted an ablation study.

4.1. Experimental Details and Datasets

The proposed method was implemented in Python programming language with PyTorch 1.7.1 [44] on a computer equipped Intel i7-10700F CPU, 2.9 GHz, 32 GB memory (Intel, Santa Clara, CA, USA), and NVIDIA GeForce RTX 3090 GPU (Nvidia, Santa Clara, CA, USA). The input image is 256 × 256 with a batch size of 4. The learning rate in our experiments is initially set to 0.0002 and decayed using cyclic cosine annealing [45], and the model is optimized using Adam method. To train the proposed MBWDN, the TIP dataset with a large diversity of moiré images [6] was employed. TIP includes 135,000 moiré images acquired from screens and captured under various imaging conditions, where each image pair contains one moiré image and its moiré-free ground-truth. We picked 90% images from the dataset for training, and the remaining 10% were used for testing.

4.2. Quantitative Evaluation

To quantitatively evaluate the performance of multibranch wavelet-based single image demoiréing method, the five moiré removal networks, including DMCNN, WDNet, MBCNN, EAFM, and UHDM, were used for comparisons, where all methods are also deep learning-based approaches. The three well-known metrics, peak signal-to-noise ratio (PSNR) [46], structural similarity index (SSIM) [47], and learning perceptual image block similarity (LPIPS) [48], were used for quality assessment of moiré-removed images. Quantitative comparison results are shown in Table 1. As shown in Table 1, the proposed method significantly outperforms the DMCNN, WDNet, and EAFM and outperforms the MBCNN and UHDM methods. Our method outperforms all methods, achieving a PSNR improvement of 1.11dB when compared to the state-of-the-art method, UHDM. Furthermore, the proposed method produces structures similar to moiré-free images with an SSIM value of 0.951. Among all compared methods, the number of parameters of the proposed network is slightly higher than that of DMCNN and WDNet; however, DMCNN and WDNet fail to effectively eliminate moiré patterns, resulting in the much lower PSNR values.
Most existing methods usually only address moiré effect in the frequency or spatial domain. However, the variety and complexity of moiré patterns makes it extremely challenging to completely remove them from a domain while preserving the original texture. Moiré patterns span low and high frequencies in the frequency domain and are intertwined with image texture in the spatial domain. In our multibranch architecture, the wavelet decomposes an image into different scales of details and approximations, allowing for multi-resolution analysis of the image. In general, the wavelet does not result in the loss of spatial domain information, so our proposed method can simultaneously remove moiré patterns both in the frequency domain and the spatial domain. Furthermore, we train two sub-networks to handle high and low frequencies, allowing each sub-network to focus on its designated task to obtain better demoiréing results.

4.3. Qualitative Evaluation

Figure 4 and Figure 5 illustrate the qualitative comparison of the different methods. We can observe that the input image is heavily contaminated by moiré patterns, showing obvious color changes and curve- or stripe-shaped patterns superimposed on the clear image. DMCNN does not effectively remove moiré patterns and has color distortion issues. UHDM shows better results in moiré removal and color restoration; however, in some cases, certain patterns still cannot be eliminated. In contrast, our proposed method is proven to be more effective in removing moiré patterns. Moreover, our proposed network performs well in recovering high-frequency details and edges because we separate high-frequency images in DMRN for training. Furthermore, in order to reveal the effectiveness of the proposed wavelet decomposition-guided image demoiréing framework in benefiting image detail restoration, we highlight some regions (highlighted by the red boxes) in the restored image, as shown in Figure 5. Based on the corresponding highlighted areas, the proposed method is more effective in reconstructing areas with rich texture, retaining more details, and providing better color recovery and better contrast than the compared methods. Furthermore, in order to demonstrate the practicality and feasibility of our proposed method, we used our proposed method on real moiré images produced by capturing LCD screens using mobile phone cameras, as shown in Figure 6. The recaptured screen images contain various types of moiré patterns. Although these moiré images are not included in the datasets used for evaluation, our proposed method demonstrates its capability to effectively eliminate various types of moiré patterns. The proposed method demonstrates its credibility by effectively enhancing the visibility of the recaptured screen image by removing moiré patterns.

4.4. Ablation Experiment

To investigate the effectiveness of the different components in the proposed framework, an ablation study is performed and described as follows. First, to evaluate the effectiveness of the multibranch structure, we remove two branches from the original structure. Second, to evaluate the effectiveness of the two networks used in the multibranch structure, we replace DMRN with MRN, and both sub-images are processed using MRN. Third, to evaluate the effectiveness of the loss function, we removed the loss of each branch L D , and only L 1 , L p ,   and L f t are used between the final demoiréd image and the ground-truth image. Fourth, we removed L f t to verify its effectiveness. As shown in Table 2, replacing or removing these key components of the proposed results in poor image restoration quality.

5. Conclusions

This paper proposes a new multibranch wavelet-based image demoiréing network (MBWDN) embedded into wavelet image decomposition-guided framework for single image moiré pattern removal. Our method first decomposes an input moiré image into the low-frequency and high-frequency components. The low-frequency component is fed into our moiré removal network (MRN) with multiscale CNN, an attention module, and residual blocks to obtain the moiré-removed component while extracting the moiré-relevant structural features for preserving the structure and color information of the image. The high-frequency component is enhanced using residual dense blocks while preserving the original image details. Through multibranch learning, the moiré patterns of the decomposed images at all levels can be effectively removed while retaining the structure and color of the original image. The presented experimental results show that the proposed method achieves better moiré removal performance and lower (or comparable) computational complexity with the state-of-the-art single-image demoiréing algorithms.
The images in the test dataset contain moiré patterns throughout the image, but typically only a portion of the image is affected by moiré artifacts. Occasionally, the network cannot accurately identify moiré patterns. Furthermore, if an object or background in an image has a texture pattern similar to moiré artifacts, it is difficult for our method to distinguish whether it is a moiré pattern or not. Therefore, further improvements to the existing architecture are still needed. In the future, we will focus on improving the network architecture to better recognize for the above cases.

Author Contributions

C.-H.Y., C.L. and C.-H.H. conceived and designed the experiments; C.L. conducted an investigation; C.-H.Y. provided resources; C.L. and C.-H.H. performed the experiments; C.L. and C.-H.H. validated the experimental results; C.L. visualized the experimental data; C.-H.Y. supervised the research; C.L. and C.-H.Y. wrote the original draft; C.-H.Y., C.L. and C.-H.H. reviewed and edited the articles. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology, Taiwan, grant number MOST 110-2221-E-003-005-MY3 and MOST 111-2221-E-003-007-MY3.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Nishioka, K.; Hasegawa, N.; Ono, K.; Tatsuno, Y. Endoscope System Provided with Low-Pass Filter for Moiré Removal. U.S. Patent 6,025,873, 15 February 2000. [Google Scholar]
  2. Liu, F.; Yang, J.; Yue, H. Moiré pattern removal from texture images via low-rank and sparse matrix decomposition. In Proceedings of the IEEE Visual Communications and Image Processing (VCIP), Singapore, 13–16 December 2015; pp. 1–4. [Google Scholar]
  3. Wei, Z.; Wang, J.; Nichol, H.; Wiebe, S.; Chapmane, D. A median-Gaussian filtering framework for Moiré pattern noise removal from X-ray microscopy image. Micron 2012, 43, 170–176. [Google Scholar] [CrossRef] [PubMed]
  4. Yang, J.; Liu, F.; Yue, H.; Fu, X.; Hou, C.; Wu, F. Textured Image Demoiréing via Signal Decomposition and Guided Filtering. IEEE Trans. Image Process. 2017, 26, 3528–3541. [Google Scholar] [CrossRef] [PubMed]
  5. Yang, J.; Zhang, X.; Cai, C.; Li, K. Demoiréing for screen-shot images with multi-channel layer decomposition. In Proceedings of the IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  6. Sun, Y.; Yu, Y.; Wang, W. Moiré Photo Restoration Using Multiresolution Convolutional Neural Networks. IEEE Trans. Image Process. 2018, 27, 4160–4172. [Google Scholar] [CrossRef] [PubMed]
  7. Liu, B.; Shu, X.; Wu, X. Demoiréing of Camera-Captured Screen Images Using Deep Convolutional Neural Network. arXiv 2018, arXiv:1804.03809. [Google Scholar]
  8. He, B.; Wang, C.; Shi, B.; Duan, L. Mop Moiré Patterns Using MopNet. In Proceedings of the International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2424–2432. [Google Scholar]
  9. Yu, X.; Dai, P.; Li, W.; Ma, L.; Shen, J.; Li, J.; Qi, X. Towards efficient and scale-robust ultra-high-definition image demoiréing. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 646–662. [Google Scholar]
  10. Yang, S.; Lei, Y.; Xiong, S.; Wang, W. High resolution demoire network. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 888–892. [Google Scholar]
  11. Yang, C.; Yang, Z.; Ke, K.; Chen, T.; Grzegorzek, M.; See, J. Doing More with Moiré Pattern Detection in Digital Photos. IEEE Trans. Image Process. 2023, 32, 694–708. [Google Scholar] [CrossRef] [PubMed]
  12. Niu, Y.; Lin, Z.; Liu, W.; Guo, W. Progressive Moire Removal and Texture Complementation for Image Demoireing. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3608–3621. [Google Scholar] [CrossRef]
  13. Nguyen, D.H.; Lee, S.; Lee, C. Multiscale Coarse-to-Fine Guided Screenshot Demoiréing. IEEE Signal Process. Lett. 2023, 30, 898–902. [Google Scholar] [CrossRef]
  14. Zheng, B.; Yuan, S.; Slabaugh, G.; Leonardis, A. Image Demoireing with Learnable Bandpass Filters. In Proceedings of the IEEE conference on computer vision and pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3636–3645. [Google Scholar]
  15. Liu, L.; Liu, J.; Yuan, S.; Slabaugh, G.; Leonardis, A.; Zhou, W.; Tian, Q. Wavelet-Based Dual-Branch Network for Image Demoiréing. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 86–102. [Google Scholar]
  16. Sun, C.; Lai, H.; Wang, L.; Jia, Z. Efficient Attention Fusion Network in Wavelet Domain for Demoireing. IEEE Access 2021, 9, 53392–53400. [Google Scholar] [CrossRef]
  17. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  18. Park, D.; Kang, D.; Kim, J.; Chun, S. Multi-Temporal Recurrent Neural Networks for Progressive Non-Uniform Single Image Deblurring with Incremental Temporal Training. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 327–343. [Google Scholar]
  19. Zamir, S.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.; Yang, M.; Shao, L. Multi-Stage Progressive Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14821–14831. [Google Scholar]
  20. Cho, S.; Ji, S.; Hong, J.; Jung, S.; Ko, S. Rethinking Coarse-To-Fine Approach in Single Image Deblurring. In Proceedings of the International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 4641–4650. [Google Scholar]
  21. Zamir, S.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.; Yang, M. Restormer: Efficient Transformer for High-Resolution Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5728–5739. [Google Scholar]
  22. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed]
  23. Yue, Z.; Yong, H.; Zhao, Q.; Meng, D.; Zhang, L. Variational denoising network: Toward blind noise modeling and removal. In Proceedings of the Advances in Neural Information Processing System (BeurlPS), Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  24. Chen, Z.; Zhang, Y.; Ding, L.; Xia, B.; Gu, J.; Kong, L.; Yuan, X. Hierarchical Integration Diffusion Model for Realistic Image Deblurring. In Proceedings of the Advances in Neural Information Processing System, Vancouver, BC, Canada, 8–12 December 2024. [Google Scholar]
  25. Li, Y.; Fan, Y.; Xiang, X.; Demandolx, D.; Ranjan, R.; Timofte, R.; Van, L. Efficient and Explicit Modelling of Image Hierarchies for Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 18278–18289. [Google Scholar]
  26. Zhang, J.; Zhang, Y.; Jin, Y.; Xu, J.; Xu, X. MDU-Net: Multi-scale densely connected U-Net for biomedical image segmentation. Health Inf. Sci. Syst. 2023, 11, 13. [Google Scholar] [CrossRef] [PubMed]
  27. Chen, S.; Ye, T.; Liu, Y.; Chen, E. Dual-former: Hybrid self-attention transformer for efficient image restoration. Digit. Signal Process. 2024, 149, 104485. [Google Scholar] [CrossRef]
  28. Zamir, S.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.; Yang, M.; Shao, L. Learning enriched features for real image restoration and enhancement. In Proceedings of the European Conference on Computer Vision (ECCV), Online, 23–28 August 2020; pp. 492–511. [Google Scholar]
  29. Wang, Z.; Cun, X.; Bao, J.; Zhou, W.; Liu, J.; Li, H. Uformer: A General U-shaped transformer for image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 17683–17693. [Google Scholar]
  30. Deng, J.; Dong, W.; Socher, R.; Li, J.-L.; Li, K.; Li, F.-F. ImageNet: A large-scale hierarchical image database. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Online, 19–25 June 2009; pp. 248–255. [Google Scholar]
  31. Huang, H.; He, R.; Sun, Z.; Ta, T. Wavelet-SRNet: A Wavelet Based CNN for Multi-Scale Face Super Resolution. In Proceedings of the International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1689–1697. [Google Scholar]
  32. Gueguen, L.; Sergeev, A.; Kadlec, B.; Liu, R.; Yosinski, J. Faster Neural Networks Straight from JPEG. In Proceedings of the Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  33. Liu, Z.; Xu, J.; Peng, X.; Xiong, R. Frequency-Domain Dynamic Pruning for Convolutional Neural Networks. In Proceedings of the Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  34. Xu, Z.; Zhang, Y.; Xiao, Y. Training Behavior of Deep Neural Network in Frequency Domain. In Proceedings of the Neural Information Processing: 26th International Conference (ICONIP), Sydney, NSW, Australia, 12–15 December 2019; pp. 264–274. [Google Scholar]
  35. Xu, K.; Qin, M.; Sun, F.; Wang, Y.; Chen, Y.; Ren, F. Learning in the frequency domain. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1740–1749. [Google Scholar]
  36. Liu, Y.; Li, Q.; Sun, Z. Attribute-Aware Face Aging with Wavelet-Based Generative Adversarial Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 11877–11886. [Google Scholar]
  37. Li, Q.; Shen, L.; Lai, Z. WaveCNet: Wavelet Integrated CNNs to Suppress Aliasing Effect for Noise-Robust Image Classification. IEEE Trans. Image Process. 2021, 30, 7074–7089. [Google Scholar] [CrossRef] [PubMed]
  38. Mallat, G.S. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674–693. [Google Scholar] [CrossRef]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, Nevada, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  40. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  41. Huang, G.; Liu, Z.; Maaten, L.; Weinberger, K. Densely Connected Convolutional Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  42. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  43. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 694–711. [Google Scholar]
  44. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An imperative style, high-performance deep learning library. In Proceedings of the Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 8024–8035. [Google Scholar]
  45. Loshchilov, L.; Hutter, F. SGDR: Stochastic Gradient Descent with Warm Restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar]
  46. Huynh-Thu, Q.; Ghanbari, M. Scope of validity of PSNR in image/video quality assessment. Electron. Lett. 2008, 44, 800–801. [Google Scholar] [CrossRef]
  47. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  48. Zhang, R.; Isola, P.; Efros, A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 586–595. [Google Scholar]
Figure 1. Moiré textures of different scales, frequencies, and colors.
Figure 1. Moiré textures of different scales, frequencies, and colors.
Sensors 24 02762 g001
Figure 2. Illustration of the network architecture of our proposed method.
Figure 2. Illustration of the network architecture of our proposed method.
Sensors 24 02762 g002
Figure 3. Illustration of the network architecture of the proposed MRN and DMRN.
Figure 3. Illustration of the network architecture of the proposed MRN and DMRN.
Sensors 24 02762 g003
Figure 4. Comparisons of visual quality among DMCNN, WDNet, MBCNN, and UHDM.
Figure 4. Comparisons of visual quality among DMCNN, WDNet, MBCNN, and UHDM.
Sensors 24 02762 g004
Figure 5. Comparisons of visual quality evaluation of complex patterns areas among DMCNN, WDNet, MBCNN, and UHDM.
Figure 5. Comparisons of visual quality evaluation of complex patterns areas among DMCNN, WDNet, MBCNN, and UHDM.
Sensors 24 02762 g005
Figure 6. Real moiré images and the corresponding outputs of our proposed method.
Figure 6. Real moiré images and the corresponding outputs of our proposed method.
Sensors 24 02762 g006
Table 1. Comparisons of quantitative results evaluated in terms of average PSNR (dB), SSIM, and LPIPS. (↑) indicates greater is preferable, while (↓) indicates lesser is preferable.
Table 1. Comparisons of quantitative results evaluated in terms of average PSNR (dB), SSIM, and LPIPS. (↑) indicates greater is preferable, while (↓) indicates lesser is preferable.
DatasetMetricsInputDMCNNWDNetMBCNNEAFNUHDMOur Method
TIP datasetPSNR ↑20.3026.7728.0830.0329.7030.1131.22
SSIM ↑0.7380.8710.9040.8930.8930.9200.951
LPIPS ↓0.1910.1420.1640.096-0.0820.077
LCDMoiréPSNR ↑10.4435.4829.6644.04-45.3445.82
SSIM ↑0.5720.9790.9670.995-0.9970.998
Params (M)-1.4263.36014.19238.00010.6238.071
Table 2. Results of the ablation study.
Table 2. Results of the ablation study.
Only
One
Branch
Only
MRN Network
Only
L D
Loss
w/o
FFT
Loss
Full
Method
PSNR29.9629.8530.7430.9231.22
SSIM0.9360.9350.9430.9470.951
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yeh, C.-H.; Lo, C.; He, C.-H. Multibranch Wavelet-Based Network for Image Demoiréing. Sensors 2024, 24, 2762. https://doi.org/10.3390/s24092762

AMA Style

Yeh C-H, Lo C, He C-H. Multibranch Wavelet-Based Network for Image Demoiréing. Sensors. 2024; 24(9):2762. https://doi.org/10.3390/s24092762

Chicago/Turabian Style

Yeh, Chia-Hung, Chen Lo, and Cheng-Han He. 2024. "Multibranch Wavelet-Based Network for Image Demoiréing" Sensors 24, no. 9: 2762. https://doi.org/10.3390/s24092762

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop