Next Article in Journal
Combination of VMD Mapping MFCC and LSTM: A New Acoustic Fault Diagnosis Method of Diesel Engine
Next Article in Special Issue
A Novel Improved YOLOv3-SC Model for Individual Pig Detection
Previous Article in Journal
Wavelet LSTM for Fault Forecasting in Electrical Power Grids
Previous Article in Special Issue
A Routing Optimization Method for Software-Defined Optical Transport Networks Based on Ensembles and Reinforcement Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Moiré Pattern Elimination in Both Frequency and Spatial Domains for Image Demoiréing

1
State Key Laboratory of Media Convergence and Communication, Communication University of China, Beijing 100024, China
2
Academy of Broadcasting Science, National Radio of Television Administration, Beijing 100866, China
3
Key Laboratory of Convergent Media and Intelligent Technology, Ministry of Education, Communication University of China, Beijing 100024, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(21), 8322; https://doi.org/10.3390/s22218322
Submission received: 6 September 2022 / Revised: 26 October 2022 / Accepted: 26 October 2022 / Published: 30 October 2022
(This article belongs to the Special Issue Smart Mobile and Sensing Applications)

Abstract

:
Recently, with the rapid development of mobile sensing technology, capturing scene information by mobile sensing devices in the form of images or videos has become a prevalent recording method. However, the moiré pattern phenomenon may occur when the scene contains digital screens or regular strips, which greatly degrade the visual performance and image quality. In this paper, considering the complexity and diversity of moiré patterns, we propose a novel end-to-end image demoiré method, which can learn moiré pattern elimination in both the frequency and spatial domains. To be specific, in the frequency domain, considering the signal energy of moiré pattern is widely distributed in the frequency, we introduce a wavelet transform to decompose the multi-scale image features, which can help the model identify the moiré features more precisely to suppress them effectively. On the other hand, we also design a spatial domain demoiré block (SDDB). The SDDB module can extract moiré features from the mixed features, then subtract them to obtain clean image features. The combination of the frequency domain and the spatial domain enhances the model’s ability in terms of moiré feature recognition and elimination. Finally, extensive experiments demonstrate the superior performance of our proposed method to other state-of-the-art methods. The Grad-CAM results in our ablation study fully indicate the effectiveness of the two proposed blocks in our method.

1. Introduction

In recent years, mobile phones have greatly changed our lives and are widely used in many scenes [1,2]. They can help us to record important information in time by shooting images or videos. However, when we use mobile phones to record scenes with LED screens or display screens, the images or videos tend to have wavy interference patterns, known as moiré patterns. The reason for this is that when the photosensitive elements in the scanning instrument and digital camera are disturbed by high frequency, two equal amplitude sine waves with similar frequencies are superimposed, and the amplitude of the synthesized signal will change according to the difference of the two frequencies; then, colored and irregular-shaped stripes will appear on the images [3]. One way to avoid moiré patterns is to improve the camera resolution, making it larger than the resolution of the captured screen, but this is very expensive. Another way is to add a low-pass filter in front of the sensing device of the camera to reduce the moiré patterns, but the disadvantage of this is that the addition of the low-pass filter will impair the image details, making the image content blur.
With the rapid development of deep learning, using algorithms based on deep learning for image demoiréing has received considerable attention. For example, Liu et al. [4] considered that the energy distribution of moiré patterns is relatively concentrated in the frequency domain and used discrete cosine transform (DCT) to transfer the image from the spatial domain to the frequency domain. They decomposed the image into texture components and moiré components in the frequency domain so as to better eliminate the high-frequency moiré patterns in the images. However, it is difficult to completely eliminate moiré patterns only through the frequency domain. Yang et al. [5] used layer decomposition on polyphase components (LDPC) to decompose the image into a background layer and a moiré layer. In addition, to better remove the moiré patterns, the method was applied on Y and RGB channels. Sun et al. [6] proposed a multi-resolution network (DMCNN) for the multi-frequency characteristics of moiré patterns. However, this network cannot deal with low frequencies and large color blocks. Some images have poor visual performance. In 2020, He et al. [7] proposed a full high-definition demoiréing network (FHDe 2 Net) to solve the problem of high-resolution image demoiréing. They used a global-to-local pattern removal strategy for fine detail preservation in high-resolution images and adopted DCT to transform the image into the frequency domain so as to better address the problems of the wider moiré pattern scale range. In 2022, Yu et al. [8] proposed ESDNet, a method for demoiréing in 4K ultra-high definition. In order to eliminate the multi-scale moiré patterns, they built a semantic-aligned scale-aware module. However, most existing methods attempted to eliminate the moiré patterns only in the frequency domain or the spatial domain. Moiré patterns have a certain diversity and complexity. In the frequency domain, moiré patterns span low frequencies and high frequencies. In the spatial domain, moiré patterns mix with the image texture seriously and also cause color distortions. It is difficult to completely remove moiré patterns while keeping the original texture just from one domain.
To address the above-mentioned problems, we propose a novel and effective end-to-end demoiré network, which can eliminate the moiré patterns both in frequency and spatial domains, named FSD-Net. Specifically, in the frequency domain, we introduce the wavelet transform to decompose the multi-scale image features, which can help the network better identify the moiré features so as to suppress the moiré features in the image generation. In the spatial domain, we design a spatial domain demoiré block (SDDB), which can extract the moiré features from the mixed image features. After the extraction, the moiré features will be subtracted from the mixed image feature, which can obtain clean features to generate a clean image without moiré patterns. By demoiréing both in the frequency domain and in spatial domain, our proposed network can obtain superior performance regarding eliminating moiré patterns. Experimental results demonstrate that our proposed method outperforms several state-of-the-art demoiré methods.
The contributions of our research can be summarized as follows:
  • We propose a novel method to eliminate moiré patterns both in the frequency domain and spatial domain. Experimental results indicate that our method achieves state-of-the-art performance compared with other methods.
  • We introduce wavelet transform to decompose the multi-scale image features, Which may help the network to better identify the moiré features so as to suppress the moiré features during the image generation.
  • We design a spatial-domain demoiré block, which can effectively extract moiré features from mixed image features. Then we can subtract the moiré features from the mixed features to obtain clean features, which are used during the image generation.

2. Related Work

2.1. Traditional Methods

Moiré patterns are often caused by the aliasing of two equal-amplitude sine waves with similar frequencies. The direct solution is to change the frequency of one sine wave. In 2000, Nishioka et al. [9] proposed that adding a low-pass filter in front of the digital camera lens could effectively remove the moiré patterns. This method can avoid the moiré patterns during the shooting, though it also loses some details and texture information. Another method is to upgrade the camera with a higher resolution photosensitive element, but such a camera is expensive and cannot fundamentally solve the moiré problem. Sidorov et al. [10] proposed a spectral model, which leverages the magnitude of the Fourier spectrum of the image to identify the moiré patterns. In addition, signal decomposition is also an important method for removing moiré patterns. Yang et al. [11] proposed a novel image demoiréing method by signal decomposition and guided filtering. Firstly, they adopted a low-rank and sparse matrix decomposition model to remove moiré patterns in the green (G) channel. Then, they removed moiré patterns in red (R) and blue (B) channels via guided filtering by the obtained texture layer of the G channel.

2.2. Deep Learning Methods

Unlike with other image restoration tasks, such as image denoising [12,13,14], image dehazing [15,16], and image demosaicing [17,18], the difficulty of image demoiréing is how to remove moiré patterns with various frequencies and color distortion. With the widespread popularity of deep learning, deep convolutional neural networks have also been applied to image demoiréing. In 2018, Sun et al. [6] first proposed a multi-resolution fully convolutional neural network DMCNN for multi-frequency moiré feature elimination. In 2019, Gao et al. [19] considered that the relationship among multi-scale features is significantly ignored and designed a feature-enhancing branch to fuse high-level features with low-level ones, which can restore the image details during image demoiréing. In 2019, Cheng et al. [20] also dealt with multi-scale features and proposed a dynamic feature-encoding module, which can encode the variations of moiré patterns. He et al. [21] proposed a Moiré pattern removal neural network (MopNet) based on DenseNet. The model integrated the moiré frequency distribution, edge intensities, and appearance categories into design learning modules. However, when irregular and unstable backgrounds such as sand and stone ground are encountered, the edge features of the pattern are difficult to determine and need to be improved. In 2020, Zheng et al. [22] proposed a novel multi-scale bandpass convolutional neural network (MBCNN), which splits image demoiréing into two steps: moiré texture removal and tone mapping. In 2020, Liu et al. [23] proposed a WDNet network, including a direction perception module. The module can carry out the convolution operation in 8 different directions so as to better capture the spatial distributions of moiré patterns. He et al. [7] proposed a full high-definition demoiréing network (FHDe 2 Net) to solve the high-resolution image demoiréing by a cascade of two networks focused on global and local level moiré removal, respectively. Moreover, they also proposed a lower high-resolution content separation branch, which can preserve the fine details against the distortions in demoiré processing. In 2021, Park et al. [24] proposed an unsupervised end-to-end moiré pattern removal method based on cyclic moiré learning. Compared with other methods, this method used an unpaired set of clean and moiré images. In 2022, Yu et al. [8] proposed a baseline model, ESDNet, a method for removing moiré patterns in 4K ultra-high-definition images, and constructed a semantic-aligned scale-aware module to solve the moiré elimination effectively. In the same year, Dai et al. [25] proposed the first moiré pattern removal method with implicit feature space alignment and selective feature aggregation for hand-held video.

2.3. Wavelet-Based Methods

The wavelet transform can decompose complex and composite information into elementary simple forms at different positions and scales, which can benefit us in understanding the information. Wavelet-based methods have been explored in several computer vision tasks. For example, Lotfi et al. [26] adopted the Daubechies 4 wavelet transform and first-order color moments to represent the image information, then a neural network was proposed to identify the category of aircraft images. Nayak et al. [27] utilized a two-dimensional discrete wavelet transform for extracting brain magnetic resonance image features. Liu et al. [28] used wavelet-based methods to capture age-related texture details at multiple scales in the frequency domain. Huang et al. [29] used wavelets for face super-resolution, where neural networks were used to predict the wavelet coefficients.

3. Methodology

In this section, we first describe the structure of our proposed FSD-Net for image demoiréing and the overall pipeline. Next, we present the details of the frequency domain demoiré block and the spatial domain demoiré block, which are the crucial components of FSD-Net. After that, we present the loss function.

3.1. Overall Network

The overall structure of our proposed FSD-Net is shown in Figure 1. As can be seen, our network consists of a generator and a discriminator. The generator is a U-shaped network with skip connections between the encoder and the decoder. In the encoder, we adopt four down encoder Blocks and a res block to encode the multi-scale moiré image features. Using the decoder, we gradually remove the moiré patterns from the multi-scale feature maps both in the frequency domain and spatial domain by our proposed demoiré block. The discriminator is designed following the structure of [30], which can help further enhance the quality of the generated images.
To be specific, the input of FSD-Net is a moiré image with the size of I R 3 × H × W . Firstly, we adopt an init block to obtain the initial image feature maps, denoted as f e 0 R C × H × W , where C, H, and W are the channel, height and weight size of the feature maps, respectively. Then, the initial feature maps will be sent into the encoder, which consists of four down encoder blocks and a res block. Each down encoder block consists of a 3 × 3 convolution with stride 2, an instance normalization, and a ReLU function. The res block is used to further encode the image feature maps; the output of the encoder is f e 4 R 16 C × ( H / 16 ) × ( W / 16 ) .
Next, following the encoder of the generator, the feature maps f e 4 are passed through four stages of decoding. Each stage contains a stack of the proposed demoiré block and an up dec block. The motivation of the demoiré block is related to the two major challenges during moiré pattern elimination. One is that the moiré patterns exhibit considerable variation in frequency. The other is that in the spatial domain, the moiré features are seriously mixed with image texture. It is difficult to filter out the moiré patterns under these problems. To address the above-mentioned issues, we built the demoiré block with two core designs: the frequency domain demoiré block (FDDB) and the spatial domain demoiré block (SDDB). In the demoiré block, we first put the decoder feature maps f d i into the FDDB, which can recognize the moiré feature maps in the frequency domain by wavelet transform and then weight the moiré feature maps to suppress them during the image generation. Then, the SDDB is added at the end of FDDB to further remove the moiré features, which are mixed with image texture features by subtracting operations, to obtain cleaner features. After the demoiré block, we adopt a transposed convolution operation in the up-decoder block to decode the image features. Finally, we adopt a gen block to restore the image to its original image size.

3.2. Frequency Domain Demoiré Block

The key innovation of our proposed FDDB is to disentangle image decoding feature maps into multiple frequency sub-bands and recognize the moiré features in the frequency domain, then suppress the moiré features by weighting coefficients. We introduce wavelet transform (WT) to transform image features from the spatial domain to the frequency domain. Wavelet transform consists of a low-pass filter and a high-pass filter; it applies the low-pass filter and high-pass filter alternately along feature columns and rows to produce four sub-band frequency features, denoted as LL, LH, HL, and HH, where L indicates the low frequencies and H represents the high frequencies. By using the wavelet transform, the model can better distinguish the moiré features and image features.
The structure of the frequency domain demoiré block is shown in Figure 2. As shown in the figure, we first adopt wavelet transform to decompose features into multiple wavelet sub-bands with different frequency contents. The operation is defined as:
f d L L , f d L H , f d H L , f d H H = W T ( f d ) ,
where f d R C × H × W is the input image features, f d L L , f d L H , f d H L , and f d H H represent the four sub-bands, the size of the sub-band feature map is C × H / 2 × W / 2 , and W T ( ) denotes the wavelet transform function.
Once the four sub-bands are obtained, we adopt a pooling layer to obtain the feature vector. In this step, it is important to align the low frequency and high frequency pooling parameters of the same input feature map. So after the pooling layer, the four sub frequency domain pooling parameters of each feature map is arranged with the corresponding positions.
Finally, the flattened vector will be sent to a FC layer and a sigmoid function to obtain the weighted parameters, which will be used to weight the input image features so as to suppress the moiré features and obtain clean image features.

3.3. Spatial Domain Demoiré Block

The above FDDB based on wavelet transform focuses on the recognition of frequency differences between the feature maps. However, some moiré patterns are mixed with the original image texture, and it is difficult to discriminate one from the other; the FDDB suffers from a limited capability to remove the moiré patterns in this case. To address this limitation, we design a structure that removes moiré patterns directly in the spatial domain. We achieve this by designing a spatial domain demoiré block, which is illustrated in Figure 3. In the SDDB, we leverage the depth-wise convolution operation to extract the moiré features from the mixed features. Then, we use a channel-wise convolution operation to obtain refined moiré features. Finally, we obtain the clean image features by subtracting the moiré features from the original input features. These operations can be defined as follows:
f m = C o n v ( f ^ d ) ,
f ^ d * = f ^ d f m ,
where C o n v ( ) denotes both the depth-wise convolution and the channel-wise convolution, and ⊖ denotes the feature-wise subtraction operation.

3.4. Loss Function

In our research, to obtain better demoiréing performance, we adopt a combined loss function, including content loss L c t , perceptual loss L p e r , and adversarial loss L a d v . The overall loss function of the generator can be formulated as:
L = λ c t × L c t + λ p e r × L p e r + λ a d v × L a d v G ,
where λ c t , λ p e r , and λ a d v are the balancing parameters of content loss, perceptual loss, and adversarial loss, respectively.
Content Loss. We use the L 1 loss to measure the content loss between the GT image and the demoiréing image generated by our proposed network.
L c t = I g t I d e 1 ,
where I g t denotes the GT image, and I d e denotes the demoiréing image.
Perceptual Loss. To penalize the perceptual and semantic discrepancy, we adopt the pre-trained VGG-19 network to extract the features of the GT image I g t and the features of the demoiréing image I d e . Then, we use the L 1 loss to measure the perceptual loss. The formula is defined as follows:
L p e r = i ( ϕ i ( I g t ) ϕ i ( I d e ) 1 ) ,
where ϕ ( ) denotes the VGG-19 network and i denotes the i-th layer of VGG-19 network. In our experiments, we employ the feature maps of the four layers conv2_2, conv3_4, conv4_4, and conv5_4 to calculate the perceptual loss.
Adversarial Loss. To effectively synthesize a realistic image, we introduce an adversarial loss, which can promote the generator to create a realistic demoiré image. We use the binary cross-entropy criterion with a softmax function to calculate the loss value. The loss function for the generator is defined as:
L a d v G = l o g ( D ( I d e ) ) ;
and the loss function for training the discriminator is defined as:
L a d v D = l o g ( D ( I g t ) ) l o g ( 1 D ( I d e ) ) ;
where D ( ) represents the discriminator.

4. Experiments

In this section, we first describe the dataset used in our experiment and the implementation details. Then, we report the subjective and objective evaluation results in comparison with other state-of-the-art methods to demonstrate the effectiveness of our proposed demoiré method. Finally, we conduct some ablation studies to verify the performance benefit brought by each functional component in our method.

4.1. Dataset and Implementation Details

Dataset. Our experiments are conducted on the dataset provided by the Document Image Demoiré Contest, which is a sub-competition of the Baidu NetDisk AI competition. All images in the dataset are collected from real-world scenes. The dataset consists of 1000 training samples and 200 test samples; each sample contains an image with moiré patterns and a ground-truth image without moiré patterns.
Implementation Details. The discriminator in our proposed method has a similar architecture to [30], which has an input size of 256 × 256 . To achieve a good discriminative performance and accommodate the input size of 512 × 512 in our paper, we increased the depth of the discriminator network. We set the initial learning rate to 2 × 10 4 and 1 × 10 4 for the generator and the discriminator, respectively. Then, the two learning rates were both decayed by 0.1 in the 20th epoch and 40th epoch, respectively. The total training epoch was set to 60. The batch size was set to 10. We adopted the Adam optimizer [31] with β 1 = 0.5 , β 2 = 0.99 to optimize our generator and discriminator. All input images were resized into 512 × 512 pixels, and random horizontal flipping was adopted for data enhancement. The balancing parameters λ c t , λ p e r , λ a d v in the loss function were set to 2.0 , 1.0 , and 1.0 , respectively. Our proposed network was implemented with a PyTorch framework and trained with two NVIDIA RTX3090 GPUs.
Evaluation Metrics. Following the previous works, we used common metrics to evaluate the demoiré performance: PSNR (peak signal-to-noise ratio) and SSIM (structural similarity). In addition, we also showed the visual comparison with other state-of-the-art methods to evaluate the effectiveness of our proposed method.

4.2. Comparison to Other Methods

We compared our method with several state-of-the-art methods, including U-Net [32], WDNet [23], MBCNN [22], FHDe 2 Net [7], and HRDN [33].
U-Net is a very excellent network and is widely used in many image generation and restoration tasks. We chose this model as one of the comparative methods and trained it from scratch.
WDNet is a demoiré method based on wavelet transform. In contrast to our proposed method, WDNet first employs 2D fast wavelet transform to decompose the input RGB image into a sequence of wavelet subbands; these subbands will then be sent to the network to remove moiré patterns. Finally, an inverse wavelet transform is adopted to obtain the final demoiré RGB image. Thus, WDNet works mainly in the wavelet domain.
MBCNN is a demoiré method based on multi-scale features; it removes the moiré patterns in the frequency domain by the discrete cosine transform (DCT).
FHDe 2 Net is also a method that removes the moiré patterns in the frequency domain by the DCT. This method adopts a cascaded global-to-local moiré pattern removal strategy, which can handle higher-resolution images.
V is a novel high-resolution demoiré network. It also removes the moiré patterns based on frequency domain and multi-scale features. It also fully takes advantage of the relationship among feature maps with different resolutions to exchange information and enhance details.
The qualitative results. In order to verify the effectiveness of our proposed method, we report the qualitative results compared with other methods. Table 1 illustrates the qualitative results. As can be seen, our method outperforms them both in PSNR and SSIM, achieving state-of-the-art performance. The results fully demonstrate the effectiveness of our proposed method.
The visual results. Figure 4 and Figure 5 show visual comparisons with different degrees of moiré patterns. Figure 4 shows the visual results with relatively slight moiré patterns. We note that WDNet and HRDN are unable to totally remove the moiré patterns compared to other methods. U-Net and MBCNN cannot generate clear image content during the demoiréing process. FHDe 2 Net cannot competently handle the color distortion. In contrast, our method can remove moiré patterns while preserving image details.
Figure 5 shows the visual results with serious moiré patterns. It can be seen that MBCNN cannot reconstruct text details, resulting in image blur. FHDe 2 Net and HRDN cannot perfectly remove moiré patterns. U-Net and WDNet also cannot obtain an ideal demoiréing performance. Our proposed FSD-Net greatly outperforms the above-mentioned methods and generates a superior visual performance. The visual comparisons further confirm the effectiveness of our method in terms of preserving high-quality details while eliminating moiré patterns.

4.3. Ablation Study

To validate the effectiveness of the proposed frequency domain demoiré block and spatial domain demoiré block, we conducted an ablation study, which contains the following variants. (1) Baseline: we adopted only the generator without the frequency domain demoiré block and the spatial domain demoiré block. (2) Baseline+FDB: we added the frequency domain demoiré block on the baseline network. (3) Baseline+SDB: we added the spatial domain demoiré block on the baseline network. (4) Baseline+FDB+SDB: we added both the frequency domain demoiré block and the spatial domain demoiré block to the baseline network. We report the quantitative results in Table 2.
From Table 2, it can be seen that: (1) compared to the baseline, the Baseline+FDB and Baseline+SDB both clearly improve the PSNR and SSIM scores, demonstrating that the two functional blocks both contribute to a performance gain. (2) The combination of the two functional blocks remarkably improves the metrics, further proving the demoiréing solution in both the frequency domain and spatial domain is correct and feasible.
In addition, we further provide the visualization results to better illustrate the effectiveness of our proposed frequency domain demoiré block and spatial domain demoiré block.
Figure 6 shows the Grad-CAM results generated by the 3 × 3 convolutional operation within the spatial domain demoiré block. It can be seen that the active areas are the salient moiré regions, not the image content areas. It indicates that our proposed spatial domain demoiré block can focus on capturing the moiré features and suppress the moiré features during the image generation to obtain a clean image.
To verify the impact of our proposed frequency domain demoiré block, we show the comparison results between the original feature maps and the weighted feature maps by the frequency domain demoiré block. The visualization results are shown in Figure 7. We can see that: (1) in the first row, the left side of (a), (b), and (c) mainly activate the image content, while the corresponding weighted suppression effects in the right side of (a), (b), and (c) are not obvious; (2) In the second row, the features of (d) and (e) mainly represent the moiré regions, so that the weighted suppression is relatively significant. The feature of (f) contains the image content and the moiré content. Our proposed frequency domain demoiré block can effectively identify and suppress the moiré feature. Such visual results prove that our proposed frequency domain demoiré block can assist the model in recognizing and suppressing moiré features.

5. Discussion and Conclusions

In this paper, we study how to eliminate moiré patterns more effectively. We summarize and find that previous methods mainly remove moiré patterns in one domain. However, moiré patterns cover a wide range in frequency and will appear in any area with different colors and shapes, making it difficult to eliminate moiré patterns from one domain. To this end, we explore to eliminate moiré patterns both in frequency and spatial domains. In the frequency domain, we introduce wavelet transform to help identify moiré pattern features so as to suppress them. In the spatial domain, we design an SDDB module to remove the moiré features that are mixed with image features. The comparable experiments show that our proposed method is better than other state-of-the-art methods in moiré pattern elimination.
However, there are still several limitations of our demoiré method. First, the reconstruction quality of the demoiré image still has a gap with the ground-truth image, especially in the regions where moiré patterns and image content are seriously mixed. Therefore, the research of content enhancement is our future work, including image deblurring and super-resolution. In addition, we note that our method cannot well preserve the background color if it is similar with some moiré colors. Thus, how to further improve the ability of our method to distinguish moiré features from image features still needs exploration.

Author Contributions

Conceptualization, C.L., Y.W. and N.Z.; methodology, C.L., N.Z. and R.G.; software, R.G. and S.M.; validation, C.L., N.Z. and R.G.; formal analysis, Y.W. and S.M.; investigation, C.L. and N.Z.; resources, C.L. and R.G.; data curation, R.G. and S.M.; writing—original draft preparation, C.L., N.Z. and R.G.; writing—review and editing, C.L. and S.M.; visualization, R.G. and S.M.; supervision, Y.W.; project administration, Y.W.; funding acquisition, C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by two projects of the Academy of Broadcasting Science, National Radio and Television Administration of China under projects: “Video Super-Resolution Algorithm Design and Software Development for Face Blur Problem” (JBKY20220210) and “Research on 8K UHD + Multi-Camera + Interactive Viewing Cloud Broadcasting System Technologies and Solutions for Performing Arts Scenes” (JBKY20220250).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yan, M.; Li, S.; Chan, C.A.; Shen, Y.; Yu, Y. Mobility Prediction Using a Weighted Markov Model Based on Mobile User Classification. Sensors 2021, 21, 1740. [Google Scholar] [CrossRef] [PubMed]
  2. Jiang, Y.; Song, L.; Zhang, J.; Song, Y.; Yan, M. Multi-Category Gesture Recognition Modeling Based on sEMG and IMU Signals. Sensors 2022, 22, 5855. [Google Scholar] [CrossRef]
  3. Oster, G.; Wasserman, M.; Zwerling, C. Theoretical interpretation of moiré patterns. Josa 1964, 54, 169–175. [Google Scholar] [CrossRef]
  4. Liu, F.; Yang, J.; Yue, H. Moiré pattern removal from texture images via low-rank and sparse matrix decomposition. In Proceedings of the 2015 Visual Communications and Image Processing (VCIP), Singapore, 13–16 December 2015; pp. 1–4. [Google Scholar]
  5. Yang, J.; Zhang, X.; Cai, C.; Li, K. Demoiréing for screen-shot images with multi-channel layer decomposition. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  6. Sun, Y.; Yu, Y.; Wang, W. Moiré photo restoration using multiresolution convolutional neural networks. IEEE Trans. Image Process. 2018, 27, 4160–4172. [Google Scholar] [CrossRef] [Green Version]
  7. He, B.; Wang, C.; Shi, B.; Duan, L.Y. FHDe2Net: Full high definition demoireing network. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 713–729. [Google Scholar]
  8. Yu, X.; Dai, P.; Li, W.; Ma, L.; Shen, J.; Li, J.; Qi, X. Towards Efficient and Scale-Robust Ultra-High-Definition Image Demoireing. arXiv 2022, arXiv:2207.09935. [Google Scholar]
  9. Nishioka, K.; Hasegawa, N.; Ono, K.; Tatsuno, Y. Endoscope System Provided with Low-Pass Filter for Moire Removal. U.S. Patent 19970917429, 26 August 1997. [Google Scholar]
  10. Sidorov, D.N.; Kokaram, A.C. Suppression of moiré patterns via spectral analysis. In Proceedings of the Visual Communications and Image Processing 2002, San Jose, CA, USA, 19 January 2002; SPIE: Bellingham, WA, USA, 2002; Volume 4671, pp. 895–906. [Google Scholar]
  11. Yang, J.; Liu, F.; Yue, H.; Fu, X.; Hou, C.; Wu, F. Textured image demoiréing via signal decomposition and guided filtering. IEEE Trans. Image Process. 2017, 26, 3528–3541. [Google Scholar] [CrossRef] [PubMed]
  12. Elad, M.; Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Tian, C.; Xu, Y.; Li, Z.; Zuo, W.; Fei, L.; Liu, H. Attention-guided CNN for image denoising. Neural Netw. 2020, 124, 117–129. [Google Scholar] [CrossRef] [PubMed]
  15. Ren, W.; Liu, S.; Zhang, H.; Pan, J.; Cao, X.; Yang, M.H. Single image dehazing via multi-scale convolutional neural networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 154–169. [Google Scholar]
  16. Shao, Y.; Li, L.; Ren, W.; Gao, C.; Sang, N. Domain adaptation for image dehazing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2808–2817. [Google Scholar]
  17. Mei, K.; Li, J.; Zhang, J.; Wu, H.; Li, J.; Huang, R. Higher-resolution network for image demosaicing and enhancing. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019; pp. 3441–3448. [Google Scholar]
  18. Shao, L.; Rehman, A.U. Image demosaicing using content and colour-correlation analysis. Signal Process. 2014, 103, 84–91. [Google Scholar] [CrossRef]
  19. Gao, T.; Guo, Y.; Zheng, X.; Wang, Q.; Luo, X. Moiré pattern removal with multi-scale feature enhancing network. In Proceedings of the 2019 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Shanghai, China, 8–12 July 2019; pp. 240–245. [Google Scholar]
  20. Cheng, X.; Fu, Z.; Yang, J. Multi-scale dynamic feature encoding network for image demoiréing. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019; pp. 3486–3493. [Google Scholar]
  21. He, B.; Wang, C.; Shi, B.; Duan, L.Y. Mop moire patterns using mopnet. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 2424–2432. [Google Scholar]
  22. Zheng, B.; Yuan, S.; Slabaugh, G.; Leonardis, A. Image demoireing with learnable bandpass filters. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 3636–3645. [Google Scholar]
  23. Liu, L.; Liu, J.; Yuan, S.; Slabaugh, G.; Leonardis, A.; Zhou, W.; Tian, Q. Wavelet-based dual-branch network for image demoiréing. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 86–102. [Google Scholar]
  24. Park, H.; Vien, A.G.; Koh, Y.J.; Lee, C. Unpaired image demoiréing based on cyclic moiré learning. In Proceedings of the 2021 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Tokyo, Japan, 14–17 December 2021; pp. 146–150. [Google Scholar]
  25. Dai, P.; Yu, X.; Ma, L.; Zhang, B.; Li, J.; Li, W.; Shen, J.; Qi, X. Video Demoireing With Relation-Based Temporal Consistency. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 17622–17631. [Google Scholar]
  26. Lotfi, M.; Solimani, A.; Dargazany, A.; Afzal, H.; Bandarabadi, M. Combining wavelet transforms and neural networks for image classification. In Proceedings of the 2009 41st Southeastern Symposium on System Theory, Washington, DC, USA, 15–17 March 2009; pp. 44–48. [Google Scholar]
  27. Nayak, D.R.; Dash, R.; Majhi, B. Brain MR image classification using two-dimensional discrete wavelet transform and AdaBoost with random forests. Neurocomputing 2016, 177, 188–197. [Google Scholar] [CrossRef]
  28. Liu, Y.; Li, Q.; Sun, Z. Attribute-aware face aging with wavelet-based generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Tullahoma, TN, USA, 15–17 March 2019; pp. 11877–11886. [Google Scholar]
  29. Huang, H.; He, R.; Sun, Z.; Tan, T. Wavelet-srnet: A wavelet-based cnn for multi-scale face super resolution. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1689–1697. [Google Scholar]
  30. Wang, X.; Xie, L.; Dong, C.; Shan, Y. Real-esrgan: Training real-world blind super-resolution with pure synthetic data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 1905–1914. [Google Scholar]
  31. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  32. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  33. Yang, S.; Lei, Y.; Xiong, S.; Wang, W. High resolution demoire network. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 888–892. [Google Scholar]
Figure 1. The overall structure of our proposed FSD-Net, which consists of a generator and a discriminator. The generator is designed following the image encode-decode structure.
Figure 1. The overall structure of our proposed FSD-Net, which consists of a generator and a discriminator. The generator is designed following the image encode-decode structure.
Sensors 22 08322 g001
Figure 2. The architecture of the frequency domain demoiré block (FDDB).
Figure 2. The architecture of the frequency domain demoiré block (FDDB).
Sensors 22 08322 g002
Figure 3. The structure of the spatial domain demoiré block (SDDB).
Figure 3. The structure of the spatial domain demoiré block (SDDB).
Sensors 22 08322 g003
Figure 4. Visual comparisons between our proposed FSD-Net and other methods.
Figure 4. Visual comparisons between our proposed FSD-Net and other methods.
Sensors 22 08322 g004
Figure 5. Visual comparisons between our proposed FSD-Net and other methods. The example image contains serious moiré patterns.
Figure 5. Visual comparisons between our proposed FSD-Net and other methods. The example image contains serious moiré patterns.
Sensors 22 08322 g005
Figure 6. The total visualization results of the four spatial domain demoiré blocks. The first row is the original moiré images. The second row is the Grad-CAM results.
Figure 6. The total visualization results of the four spatial domain demoiré blocks. The first row is the original moiré images. The second row is the Grad-CAM results.
Sensors 22 08322 g006
Figure 7. The visualization results of the last frequency domain demoiré block. In (af), the left side is the original feature maps, and the right side is the weighted feature maps by the last frequency domain demoiré block.
Figure 7. The visualization results of the last frequency domain demoiré block. In (af), the left side is the original feature maps, and the right side is the weighted feature maps by the last frequency domain demoiré block.
Sensors 22 08322 g007
Table 1. The comparison of qualitative results between our proposed FSD-Net and other methods.
Table 1. The comparison of qualitative results between our proposed FSD-Net and other methods.
NetworkPSNR (dB)SSIM
U-Net27.620.8289
WDNet28.660.8745
MBCNN26.830.8113
FHDe 2 Net26.390.8786
HRDN27.380.8626
ours33.240.8970
Table 2. Quantitative results of different variants of FSD-Net.
Table 2. Quantitative results of different variants of FSD-Net.
NetworkPSNR (dB)SSIM
Baseline32.010.8717
Baseline+FDB32.620.8841
Baseline+SDB32.880.8896
Baseline+FDB+SDB33.240.8970
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, C.; Wang, Y.; Zhang, N.; Gang, R.; Ma, S. Learning Moiré Pattern Elimination in Both Frequency and Spatial Domains for Image Demoiréing. Sensors 2022, 22, 8322. https://doi.org/10.3390/s22218322

AMA Style

Liu C, Wang Y, Zhang N, Gang R, Ma S. Learning Moiré Pattern Elimination in Both Frequency and Spatial Domains for Image Demoiréing. Sensors. 2022; 22(21):8322. https://doi.org/10.3390/s22218322

Chicago/Turabian Style

Liu, Chenming, Yongbin Wang, Nenghuan Zhang, Ruipeng Gang, and Sai Ma. 2022. "Learning Moiré Pattern Elimination in Both Frequency and Spatial Domains for Image Demoiréing" Sensors 22, no. 21: 8322. https://doi.org/10.3390/s22218322

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop