You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

2 November 2023

R-PreNet: Deraining Network Based on Image Background Prior

,
,
and
Institute of Intelligent Control and Image Engineering, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Recommender Systems and Their Advanced Application

Abstract

Single image deraining (SID) has shown its importance in many advanced computer vision tasks. Although many CNN-based image deraining methods have been proposed, how to effectively remove raindrops while maintaining background structure remains a challenge that needs to be overcome. Most of the deraining work focuses on removing rain streaks, but in heavy rain images, the dense accumulation of rainwater or the rain curtain effect significantly interferes with the effective removal of rain streaks, and often introduces some artifacts that make the scene more blurry. In this paper, a novel network architecture, R-PReNet, is introduced for single image denoising with an emphasis on preserving the background structure. The framework effectively exploits the cyclic recursive structure inherent in PReNet. Additionally, the residual channel prior (RCP) and feature fusion modules have been incorporated, enhancing denoising performance by emphasizing background feature information. Compared with the previous methods, this approach offers notable improvement in rainstorm images by reducing artifacts and restoring visual details.

1. Introduction

Rainfall is a prevalent meteorological phenomenon [1] that adversely affects the visual quality of images and hampers the performance of subsequent image processing tasks such as object recognition [2], object detection [3], autonomous driving, and video surveillance [4,5,6]. Consequently, the removal of rain streaks from rainy images has emerged as a significant and meaningful research topic, gaining attention in recent years. Single-image deraining refers to the restoration of a clean, rain-free image scene from a rainy single image. However, given the intricate amalgamation of background information and raindrop details, simultaneously eliminating the raindrops and preserving the background remains a challenging issue. We found in an experiment that the PReNet deraining network model [7] can reconstruct a relatively clear rain-free image, but in the test of a rainstorm dataset, the background structure of the reconstructed image corresponding to the rainstorm image has also been damaged to some extent, that is, the introduction of artifacts, and the destruction of the image background can sometimes lead to serious problems, such as blurry or missing traffic signs, which may result in serious accidents in autonomous driving. In order to address this problem, this paper introduces an additional image background prior to protect the background structure, so that a clearer and correct reconstruction of rainless images can be obtained in the case of processing rainstorm images, as shown in Figure 1.
Figure 1. Image deraining in the real world. PReNet [7] and R-PReNet were trained on RainTrainH. (a) is a real rain image, (b) is the result image after using PReNet to remove rain, and (c) is the result image after using this algorithm to remove rain. This images show that R-PReNet can effectively remove rain streaks while retaining better background textures and maintaining the basic tone of the original image.
In this article, we explore the effective reconstruction problem of complex combinations of background and raindrops, and propose a new algorithm called R-PReNet that can effectively remove raindrops and protect background information. This algorithm fully utilizes the cyclic recursive structure of PReNet and its capability to remove rain streaks. On this basis, this article introduces residual channel prior (RCP) [8,9,10,11] in the model to achieve background structure protection. In addition, this article also proposes the use of the ‘Squeeze Excitation’ residual module (SE ResBlock) [12] to extract deep features of RCP, and the interactive fusion feature module (IFM) [11] to fully utilize RCP information, achieving high-quality rainless image reconstruction.
Our contributions are summarized as follows:
This article replicates and tests the PReNet deraining network on three popular image deraining datasets (Rain100H [13], Rain100L [13], Rain14000 [14]) and real rainy image datasets (Practical_by_Yang [13]), and studies the results of deraining.
This article explores the effectiveness of residual channel prior (RCP) for background protection and proposes an image deraining network structure based on RCP. Numerous experiments have shown that our method outperforms the original method on commonly used rainfall datasets, restoring visually clean images and good detail.
An RCP extraction module and an interactive fusion module (IFM) are introduced, designated for RCP extraction and guidance, respectively. These aim to attain deep features of the RCP and guide the network to recover more background details.
The remainder of this paper is structured as follows. Section 2 briefly reviews relevant studies on image denoising methods. Section 3 presents the comprehensive R-PReNet denoising network based on image background prior and delves into the RCP residual channel prior and IFM fusion techniques. Experimental results and comparisons are detailed in Section 4. The conclusion is given in Section 5.

3. Proposed Work

In this section, the overall network architecture of the proposed algorithm is presented. The implementation details of the introduced residual channel prior (RCP) are first described. Subsequently, the structure of the progressive recursive network (PReNet), serving as the backbone network, is showcased. Finally, a method for fusing high-dimensional features of the RCP is proposed.

3.1. Residue-Progressive Recurrent Network

As shown in Figure 2, R-PReNet consists of two main parts: (i) the RCP feature extraction and fusion module, and (ii) the progressive recurrent network. Features from rainy images are first extracted and then merged with the RCP characteristics. Subsequently, the combined features are concatenated with the image attributes. The components of this approach will be detailed in the following sections.
Figure 2. The overall structure of residue-progressive recurrent network (R-PreNet), where (a) shows the overall network framework of R-PreNet; (b) shows progressive recurrent network composition in R-PreNet, where f i n is a convolutional layer with ReLU, f r e s is a recursive ResBlocks, f o u t is a convolutional layer, f r e c u r r e n t is a convolutional LSTM, and © is a connectivity layer; (c) is the RCP fusion feature module.

3.2. Residue Channel Prior (RCP)

The appearance of rain streaks is commonly modeled as a linear combination of the background and rain streak layers [14,20,22,34]. Based on this model, Li et al. [8] demonstrated that subtracting the minimum color channel from the maximum color channel produces a rain-free image. Rain streaks are colorless (white or grey) and appear at the same location in different RGB color channels. As such, subtracting the minimum color channel from the maximum one nullifies the presence of rain streaks, as in Figure 3.
Figure 3. RCP extraction module.
The colored-image intensity of a rainy image is defined [8] as:
I ~ x = τ   ρ r s x L σ + T τ B π
where L = ( L r ,   L g ,   L b ) T is the color vector of luminance and B = ( B r ,   B g ,   B b ) T is the color vector of background reflection.
L = L r + L g + L b ,           B = B r + B g + B b
In the model (Equation (2)), the first term represents the rain streak component, while the second term denotes the background component. σ = L /   L and π = B /   B define the chromaticities of L and B. T represents the exposure time, while τ denotes the time taken by a raindrop to pass through pixel x . ρ r s consists of the refraction coefficients of the raindrop, the specular reflection coefficients, and the internal reflection coefficients. The assumption is made that ρ r s is wavelength-independent, implying that raindrops are colorless.
As a consequence, it becomes necessary to cancel the light chromaticity σ in the rain-streak term in Equation (2) to generate a residual channel without rain streaks. To achieve this, any existing color constancy algorithm [35] is employed to estimate σ , and then apply the following normalization step to the input image:
I x = I ~ ( x ) σ = I r s x i + I b g ( x )
where i = ( 1   , 1 ,   1 ) T , I r s = τ ρ r s L , I b g = ( T τ ) B / σ .
Vector division is done element-wise. It should be noted that upon normalizing the image, not only is the luminance of light eliminated, but the color effects of spectral sensitivity are also removed. Hence, according to the previous equation and a rainy image I , the residual channel is defined as:
I r e s x = I M x I m ( x )
where:
I M x = max { I r x , I g x , I b x }
I m x = min { I r x , I g x , I b x }
I r e s is the residual channel of the image I , which has no rain streaks.

3.3. RCP High-Dimensional Feature Extraction

Although the operation of subtracting a color channel from another in the image space is beneficial and the structural information of the RCP is clearer than the rainy image, it can be destructive to the background image because of information loss. Therefore, the operations utilizing the structural information of RCP are shifted to the feature domain. An RCP feature extraction module is introduced to extract the high-dimensional features of the RCP.
Based on the squeeze-and-excitation (SE) block proposed by Hu et al. [36], which focuses on channel relationships to construct informative features, this residual block adaptively recalibrates the channel feature responses by explicitly modeling the interdependencies between channels. Given that the RCP module interacts through color channels, the SE ResBlock structure, as illustrated in Figure 4, is employed to extract the high-dimensional features Fp of the RCP, aiming to reduce noise in the initial features and enrich the semantic information of the features.
Figure 4. SE-ResBlock module.

3.4. Interactive Fusion Features

While high-dimensional features of the RCP have been extracted, effectively leveraging these RCP features to guide the model remains a challenging task.
A simple solution is directly concatenating RCP features with image features, but this is ineffective for guiding model deraining and may cause feature interference. To address this problem, an interactive fusion module (IFM) [37] is introduced, consisting of two branches (rainy image features and prior features) to progressively combine features. As shown in Figure 5, two 3 × 3 kernel-sized convolutions are performed to map the rainy image features F o   and RCP features F p to F o ^   and F p ^ .
Figure 5. Interactive fusion feature module.
Next, the similarity map S between F o ^ and F p ^ is computed using element multiplication:
S = S i g m o i d ( F o ^ F p ^ )
The similarity map S is utilized to enhance the background information of rainy images compromised by rain streaks. Furthermore, given that the background of RCP resembles that of the rainy image, the similarity map S also emphasizes feature information in the prior, further bolstering its structural integrity.

3.5. Progressive Recurrent Network

The progressive recurrent network consists of the following four parts: (i) a convolutional layer f i n receives network inputs, (ii) a recurrent layer f r e c u r r e n t   propagates cross-stage feature dependencies, (iii) several residual blocks f r e s extract the deep representation, and (ii) a convolutional layer f o u t outputs deconvolutional results. Where f i n takes as input the current estimation x t 1 , the rainy image y, and the concatenation of the background fusion prior features G. A convolutional long short-term memory (LSTM) is employed for the recurrent layers, given its empirical advantage in image deraining, through which cross-stage feature dependencies can be propagated to facilitate rain streaks removal:
x t 0.5 = f i n ( x t 1 , y , G )
s t = f r e c u r r e n t ( s t 1 , x t 0.5 )
x t = f o u t ( f r e s ( s t ) )
where f i n , f r e s , and f o u t are stage-invariant, the network parameters are reused across different stages. The recurrent layer f r e c u r r e n t takes x t 0.5 and the recurrent state s t 1 as inputs to stage t − 1. By unfolding PreNet [7] with T recurrent stages, the deep representation of rain streak removal is favored by recurrent state propagation. The deraining results from the intermediate stages of the network structure indicate that the accumulation of storm streaks can be gradually eliminated.

3.6. Loss Function

Given a clean single channel image I and a noisy image K of size m × n, the mean square error (MSE) is defined as:
M S E = 1 m n   i = 0 m 1 j = 0 n 1   [ I i ,   j K i ,   j ] 2
On this basis, PSNR (dB) is defined as:
P S N R = 10 log 10 ( M A X I 2 M S E )
where M A X I 2 is the maximum possible pixel value of the image. If each pixel is represented by 8 bits of binary, then it is 255. In general, if the pixel value is represented by B-bit binary, M A X I 2 = 2 B 1 .
If it is a color image, there are usually three ways to calculate it:
1. Calculate the PSNR of the RGB image’s three channels separately and then take the average value.
2. Calculate the MSE of the RGB image’s three channels, then divide by 3.
3. Convert the image to YCbCr format, and then only calculate the PSNR of the Y component, which is the brightness component.
Among them, the second and third methods are more common. This algorithm uses the second method.
The peak signal-to-noise ratio (PSNR) is an objective measure of image distortion or noise level. The larger the PSNR value between two images, the more similar it is. The general benchmark is 30 dB, and the deterioration of images below 30 dB is more obvious.
SSIM also describes the similarity of two images, and the formula is measured based on three comparisons between samples x and y: luminance, contrast and structure.
l x , y = 2 μ x μ y + c 1 μ x 2 + μ y 2 + c 1
c x , y = 2 σ x σ y + c 2 σ x 2 + σ y 2 + c 2
s x , y = σ x y + c 3 σ x σ y + c 3
Generally, c 3 = c 2 2 . where, μ x is the mean value of x and μ y is the mean value of y .   σ x 2 is the variance of x and σ y 2 is the variance of y ; σ x y 2 is the variance of x y .
c 1 = ( k 1 L ) 2 and c 2 = ( k 2 L ) 2 are two constants to avoid division by zero, and L is the range of pixel values.
k 1 = 0.01 and k 2 = 0.03 are the default values.
Then:
S S I M x , y = [ l x , y α c x , y β s ( x , y ) γ ]
During each calculation, an N × M window is taken from the image, and then the window is constantly sliding for calculation. Finally, the average value is taken as the global SSIM.
SSIM specifies the MSSIM of the returned image. This is also a floating-point number between zero and one (the higher the better).
A negative SSIM loss [38] is adopted as the objective function. For a model with T stages, there are T outputs, x 1 , x 2 , …, x T , with supervision applied only to the final output x T . The negative SSIM loss is:
L = S S I M ( x T ,   x g t )
where   x g t is the corresponding ground-truth clean image.

4. Experiments

The model was trained on Ubuntu OS, NVIDIA GeForce GTX 3080Ti GPU using Pytorch framework in Python environment with 12GB of RAM. To validate the effectiveness of the model, evaluations were conducted on three popular image-deraining synthetic datasets (Rain100H, Rain100L, Rain14000) and a real rainy images dataset (Practical_by_Yang) to evaluate our approach:
Combined with the visual effect in the Figure 6 and recognition effect in the Figure 7 of the real rain image, it can be seen that the R-PReNet algorithm has a significant background protection effect. Because the results of task-type evaluation of multi-purpose image deraining (MPID) [39] algorithm on a real dataset show that in most cases, the processing of the rain removal algorithm reduces the recognition accuracy. This paper points out that the rain removal algorithm is not optimized to improve the recognition accuracy in the training process, but some important real semantic information is lost in the rain removal process, which reduces the recognition accuracy. Therefore, the background protection module is added in this algorithm to improve the recognition accuracy.
Figure 6. Image deraining results tested in both synthetic and real datasets. The first column presents the rainy image, the second column shows the actual no-rain images from the synthetic dataset (no example images on the real dataset), the third column is the deraining result of the PReNet algorithm, and the fourth column is the deraining result of the R-PReNet algorithm of this paper. The two or three block images below each image enlarge the details of the images above. It can be seen that R-PReNet can reconstruct the rain-free image with clearer background structure and reduce the introduction of artifacts.
Figure 7. The identification result of the image deraining results tested in both synthetic and real datasets. The first column is the target confidence degree of target recognition after using PReNet algorithm to remove rain, and the second column is the target confidence degree of target recognition after using R-PReNet algorithm to remove rain. The recognition algorithm uses YOLOv4 algorithm for target detection and recognition, which is pre-trained on the MS COCO dataset.

4.1. Experimental Setup

4.1.1. Datasets

In this paper, evaluations were primarily conducted on synthetic datasets and real datasets. The synthetic image datasets included (1) Rain100L, where 200 pairs of images were used for training and 100 pairs of images were used for testing; (2) Rain100H which had 200 synthetic images used for training and 100 images used for testing; and (3) Rain14000, which was composed of training and test images with a ratio of 12,600:1400 split. The real dataset consists of (1) the Practical_by_Yang dataset with 34 images without ground-truth; and (2) 25 real rainy images from certain movie and television productions.

4.1.2. Evaluation Indicators

In these experiments, for images with ground-truths, evaluations for each method were made using two commonly adopted quantitative metrics: peak signal-to-noise ratio (PSNR) [40] and structural similarity index (SSIM) [38]. For the images without ground-truth (i.e., real dataset), visual results were provided.

4.2. Ablation Study

4.2.1. Effectiveness on RCP Module

The first ablation study evaluates the performance of R-PReNet with experimental results with and without the RCP module, networks with and without the RCP mode, as well as baseline algorithms JORDER [13] and RESCAN [28] were trained and tested on the Rain100L, Rain100H, and Rain14000 datasets. Table 1 shows the performance of the above algorithms on the quantitative results in PSNR and SSIM. Both quantitative and visual results show that the recurrent network with RCP module outperforms the network without RCP module and the baseline algorithm.
Table 1. Performance comparison of synthetic datasets on network structure with and without RCP module.

4.2.2. Effectiveness on IFM Module

To investigate the effectiveness of the feature fusion module, two different network architectures were compared: (a) with the RCP module, but the RCP high-dimensional features were directly connected with the rainy image features into the network, and (b) with the RCP module and the IFM module, which used interactive fusion to combine the RCP high-dimensional features and the rainy image features together into the network. Networks with and without the FIM module, as well as baseline algorithms JORDER [13], RESCAN [28], and PReNet [7], were trained and tested on the datasets Rain100L, Rain100H, and Rain14000, respectively. Table 2 shows the quantitative results of the above algorithms in PSNR and SSIM. Both quantitative and visual results showed that the recurrent network with IFM module outperforms the network without IFM module and the baseline network.
Table 2. Performance comparison of synthetic datasets with and without IFM module network structure.
The data in the tables are, respectively, PSNR and SSIM, where PSNR is expressed as the peak signal-to-noise ratio between the image after rain removal and the original rain image. The larger the PSNR value between the two images, the more similar it is. The general benchmark of PSNR is 30 dB, and the degradation of images below 30 dB is more obvious. The SSIM is represented here as the structural similarity between the image after the rain and the ground-truth, and the value is a floating-point number between zero and one. According to the experimental data, R-PReNet, by protecting background information, has improved image background and details in visual effects and PSNR and SSIM quality evaluation data compared with PReNet. However, since this algorithm protects the background of the image and optimizes the rain removal effect from the details, the improvement of PSNR/SSIM will not be very large.

5. Conclusions

In this paper, a progressive recursive denoising network based on background preservation is proposed. The experiments show that this algorithm can remove rain streaks and protect background information at the same time. In the preprocessing stage of rainy images, a residual channel is initially extracted from the rainy image. The extracted residual channel, devoid of rain streaks, is utilized to extract high-dimensional features. Subsequently, these extracted features are interactively fused with rainy image features and then fed into the progressive recursive network. The input for each stage of the network consists of the fused features, the reconstructed image from the previous stage, and the original rainy image. After generations of progressive recursion, the final rain-free image is produced. Comprehensive experimental evaluations show that our method outperforms the original algorithm on both synthetic and real rainy images.

Author Contributions

Methodology, C.J.; Software, C.J.; Investigation, F.M.; Resources, F.M.; Data curation, T.L. and Y.C.; Writing—original draft, C.J.; Writing—review & editing, F.M.; Supervision, F.M.; Funding acquisition, F.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant 61305040 and the Scientific Research Program Serving Local Special Projects of Shaanxi Provincial Education Department of China under Grant 23JM018.

Data Availability Statement

The datasets collation link used in this algorithm is updated as follows: https://pan.baidu.com/s/1o_lFQclEstiKEdCQOlaVeg?pwd=mbcz (accessed on 30 July 2023). Among them, the synthetic datasets Rain100H and Rain100L are from the public dataset, which are provided in the paper “Joint Rain Detection and Removal from a Single Image”. Synthetic dataset Rain14000 is from the synthetic dataset and is provided in the paper “Removing Rain from Single Images via a Deep Detail Network”; The real dataset is provided in the paper “Joint Rain Detection and Removal from a Single Image”; The real dataset of “real_rain” is from the public images on the network, the specific link is in the “download_link” txt file.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wijesinghe, D.C.; Mishra, P.K.; Withanage, N.C.; Abdelrahman, K.; Mishra, V.; Tripathi, S.; Fnais, M.S. Application of GIS, Multi-Criteria Decision-Making Techniques for Mapping Groundwater Potential Zones: A Case Study of Thalawa Division, Sri Lanka. Water 2023, 15, 3462. [Google Scholar] [CrossRef]
  2. Josi, A.; Alehdaghi, M.; Cruz, R.M.O.; Granger, E. Multimodal Data Augmentation for Visual-Infrared Person ReID with Corrupted Data. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW), Waikoloa, HI, USA, 3–7 January 2023; pp. 1–10. [Google Scholar] [CrossRef]
  3. Chaturvedi, S.S.; Zhang, L.; Yuan, X.; Weather, A. Pay “Attention” to Adverse Weather: Weather-aware Attention-based Object Detection. In Proceedings of the 2022 26th International Conference on Streaks Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 4573–4579. [Google Scholar] [CrossRef]
  4. Xiao, J.; Long, H.; Li, R.; Li, F. Research on Methods of Improving Robustness of Deep Learning Algorithms in Autonomous Driving. In Proceedings of the 2022 IEEE International Conference on Advances in Electrical Engineering and Computer Applications (AEECA), Dalian, China, 20–21 August 2022; pp. 644–647. [Google Scholar] [CrossRef]
  5. Tyagi, H.; Kumar, V.; Kumar, G. A Review Paper on Real-Time Video Analysis in Dense Environment for Surveillance System. In Proceedings of the 2022 International Conference on Fourth Industrial Revolution Based Technology and Practices (ICFIRTP), Uttarakhand, India, 26–27 March 2022; pp. 171–183. [Google Scholar] [CrossRef]
  6. Zhang, Z.; Lu, W.; Sun, W.; Min, X.; Wang, T.; Zhai, G. Surveillance Video Quality Assessment Based on Quality Related Retraining. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 4278–4282. [Google Scholar] [CrossRef]
  7. Ren, D.; Zuo, W.; Hu, Q.; Zhu, P.; Meng, D. Progressive Image Deraining Networks: A Better and Simpler Baseline. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Streaks Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 3932–3941. [Google Scholar] [CrossRef]
  8. Li, R.; Tan, R.T.; Cheong, L.F. Robust optical flow in rainy scenes. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 288–304. [Google Scholar]
  9. Li, R.; Tan, R.T.; Cheong, L.F. All in one bad weather removal using architectural search. In Proceedings of the IEEE/CVF Conference on Computer Vision and Streaks Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3175–3185. [Google Scholar]
  10. Li, R.; Tan, R.T.; Cheong, L.F.; Aviles-Rivero, A.I.; Fan, Q.; Schonlieb, C.B. Rainflow: Optical flow under rain streaks and rain veiling effect. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7304–7313. [Google Scholar]
  11. Yi, Q.; Li, J.; Dai, Q.; Fang, F.; Zhang, G.; Zeng, T. Structure-Preserving Deraining with Residue Channel Prior Guidance. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 4218–4227. [Google Scholar] [CrossRef]
  12. Zhong, X.; Gong, O.; Huang, W.; Li, L.; Xia, H. Squeeze-and-Excitation Wide Residual Networks in Image Classification. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 395–399. [Google Scholar] [CrossRef]
  13. Yang, W.; Tan, R.T.; Feng, J.; Liu, J.; Guo, Z.; Yan, S. Deep joint rain detection and removal from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1357–1366. [Google Scholar]
  14. Fu, X.; Huang, J.; Zeng, D.; Huang, Y.; Ding, X.; Paisley, J. Removing Rain from Single Images via a Deep Detail Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Washington, DC, USA, 2017. [Google Scholar]
  15. He, K.; Sun, J.; Tang, X. Guided image filtering. In Proceedings of the 11th European conference on Computer Vision, Heraklion Crete, Greece, 5–11 September 2010; pp. 1–14. [Google Scholar]
  16. Xu, J.; Zhao, W.; Liu, P.; Tang, X. Removing rain and snow in a single image using guided filter. In Proceedings of the 2012 IEEE International Conference on Computer Science and Automation Engineering, Zhangjiajie, China, 25–27 May 2012; pp. 304–307. [Google Scholar]
  17. Zheng, X.; Liao, Y.; Guo, W.; Fu, X.; Ding, X. Single-image-based rain and snow removal using multi-guided filter. In Neural Information Processing: 20th International Conference, ICONIP 2013, Daegu, Republic of Korea, 3–7 November 2013; Springer: Berlin/Heidelberg, Germany, 2013; pp. 258–265. [Google Scholar]
  18. Ding, X.; Chen, L.; Zheng, X.; Huang, Y.; Zeng, D. Single image rain and snow removal via guided L0 smoothing filter. Multimed. Tools Appl. 2016, 75, 2697–2712. [Google Scholar] [CrossRef]
  19. Kim, J.H.; Lee, C.; Sim, J.Y.; Kim, C.S. Single-image deraining using an adaptive nonlocal means filter. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; pp. 914–917. [Google Scholar]
  20. Kang, L.W.; Lin, C.W.; Fu, Y.H. Automatic single-image-based rain streaks removal via image decomposition. IEEE Trans. Image Process. 2012, 21, 1742–1755. [Google Scholar] [CrossRef] [PubMed]
  21. Kang, L.W.; Lin, C.W.; Lin, C.T.; Lin, Y.C. Self-learning-based rain streak removal for image/video. In Proceedings of the 2012 IEEE International Symposium on Circuits and Systems (ISCAS), Seoul, Republic of Korea, 20–23 May 2012; Volume 57, pp. 1871–1874. [Google Scholar]
  22. Luo, Y.; Xu, Y.; Ji, H. Removing rain from a single image via discriminative sparse coding. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3397–3405. [Google Scholar]
  23. Li, Y.; Tan, R.T.; Guo, X.; Lu, J.; Brown, M.S. Rain streak removal using layer priors. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  24. Wei, W.; Meng, D.; Zhao, Q.; Xu, Z.; Wu, Y. Semi-supervised transfer learning for image rain removal. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3877–3886. [Google Scholar]
  25. Gu, S.; Meng, D.; Zuo, W.; Zhang, L. Joint convolutional analysis and synthesis sparse representation for single image layer separation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1717–1725. [Google Scholar]
  26. Mu, P.; Chen, J.; Liu, R.; Fan, X.; Luo, Z. Learning bilevel layer priors for single image rain streaks removal. IEEE Signal Process. Lett. 2018, 26, 307–331. [Google Scholar] [CrossRef]
  27. Eigen, D.; Krishnan, D.; Fergus, R. Restoring an image taken through a window covered with dirt or rain. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 633–640. [Google Scholar]
  28. Li, X.; Wu, J.; Lin, Z.; Liu, H.; Zha, H. Recurrent squeeze-and-excitation context aggregation net for single image deraining. In Proceedings of the15th European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; Springer: Cham, Switzerland, 2018; pp. 262–277. [Google Scholar]
  29. Yasarla, R.; Sindagi, V.A.; Patel, V.M. Syn2real transfer learning for image deraining using gaussian processes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Streaks Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2726–2736. [Google Scholar]
  30. Huang, H.; Yu, A.; He, R. Memory Oriented Transfer Learning for Semi-Supervised Image Deraining. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Streaks Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 7728–7737. [Google Scholar] [CrossRef]
  31. Zhu, H.; Peng, X.; Zhou, J.T.; Yang, S. Singe Image Rain Removal with Unpaired Information: A Differentiable Programming Perspective. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 9332–9339. [Google Scholar]
  32. Jin, X.; Chen, Z.; Lin, J.; Chen, Z.; Zhou, W. Unsupervised single image deraining with self-supervised constraints. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2761–2765. [Google Scholar]
  33. Wei, Y.; Zhang, Z.; Wang, Y.; Xu, M.; Yang, Y.; Yan, S.; Wang, M. DerainCycleGAN: Rain Attentive CycleGAN for Single Image Deraining and Rainmaking. IEEE Trans. Image Process. 2021, 30, 4788–4801. [Google Scholar] [CrossRef] [PubMed]
  34. Yang, W.; Tan, R.T.; Feng, J.; Liu, J.; Yan, S.; Guo, Z. Joint rain detection and removal from a single image with contextualized deep networks. IEEE Trans. Streaks Anal. Mach. Intell. 2019, 46, 1377–1393. [Google Scholar] [CrossRef] [PubMed]
  35. Cheng, D.; Prasad, D.K.; Brown, M.S. Illuminant estimation for color constancy: Why spatial-domain methods work and the role of the color distribution. JOSA A 2014, 31, 1049–1058. [Google Scholar] [CrossRef] [PubMed]
  36. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Streaks Recognition, Salt Lake City, UT, USA, 18–23 June 2017; pp. 7132–7141. [Google Scholar]
  37. Hu, Y.; Hou, N.; Chen, C.; Chng, E.S. Interactive Feature Fusion for End-to-End Noise-Robust Speech Recognition. In Proceedings of the ICASSP 2022–2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 6292–6296. [Google Scholar]
  38. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  39. Li, S.; Araujo, I.B.; Ren, W.; Wang, Z.; Tokuda, E.K.; Junior, R.H.; Cesar-Junior, R.; Zhang, J.; Guo, X.; Cao, X. Single Image Deraining: A Comprehensive Benchmark Analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2019, Long Beach, CA, USA, 15–20 June2019. [Google Scholar]
  40. Huynh-Thu, Q.; Ghanbari, M. Scope of validity of psnr in image/video quality assessment. Electron. Lett. 2008, 44, 800–801. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.