Next Article in Journal
Recent Advances in Underwater Signal Processing
Previous Article in Journal
IRDC-Net: An Inception Network with a Residual Module and Dilated Convolution for Sign Language Recognition Based on Surface Electromyography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Local Adaptive Image Filtering Based on Recursive Dilation Segmentation

1
School of Computer Science, Nanjing University of Posts and Telecommunications, Nanjing 210046, China
2
School of Bell Honors, Nanjing University of Posts and Telecommunications, Nanjing 210046, China
3
School of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing 210046, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(13), 5776; https://doi.org/10.3390/s23135776
Submission received: 16 May 2023 / Revised: 11 June 2023 / Accepted: 16 June 2023 / Published: 21 June 2023
(This article belongs to the Section Physical Sensors)

Abstract

:
This paper introduces a simple but effective image filtering method, namely, local adaptive image filtering (LAIF), based on an image segmentation method, i.e., recursive dilation segmentation (RDS). The algorithm is motivated by the observation that for the pixel to be smoothed, only the similar pixels nearby are utilized to obtain the filtering result. Relying on this observation, similar pixels are partitioned by RDS before applying a locally adaptive filter to smooth the image. More specifically, by directly taking the spatial information between adjacent pixels into consideration in a recursive dilation way, RDS is firstly proposed to partition the guided image into several regions, so that the pixels belonging to the same segmentation region share a similar property. Then, guided by the iterative segmented results, the input image can be easily filtered via a local adaptive filtering technique, which smooths each pixel by selectively averaging its local similar pixels. It is worth mentioning that RDS makes full use of multiple integrated information including pixel intensity, hue information, and especially spatial adjacent information, leading to more robust filtering results. In addition, the application of LAIF in the remote sensing field has achieved outstanding results, specifically in areas such as image dehazing, denoising, enhancement, and edge preservation, among others. Experimental results show that the proposed LAIF can be successfully applied to various filtering-based tasks with favorable performance against state-of-the-art methods.

1. Introduction

Image filtering is a practical technique for suppressing incongruous noise while preserving the intrinsic structure information in images, which has been widely applied in many computer-vision applications, including image denoising [1,2,3], exposure image fusion [4], single-image dehazing [5], etc. Generally, image filtering methods accomplish filtering tasks by designing a translation-variant from the filtered input images to the output images under the guidance of guided images. More and more efforts on image filtering have been made with many meaningful achievements.
In existing remote sensing image dehazing methods [6,7,8], block artifacts tend to occur quite easily. However, RDS makes full use of multiple integrated information sources including pixel intensity, hue information, and especially spatial adjacent information, leading to more robust filtering results which effectively address the occurrence of block artifacts and provide a better solution to the issue. During the process of handling remote sensing images, it is essential to use our proposed method to smooth or enhance [9] the images while maintaining the critically important edge preservation and enhancement effects.
The existing image filtering methods can be mainly divided into two categories: local methods and global methods.
Local methods: The core idea of local methods [10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28] is to achieve image smoothing solely within local regions in the image. The most representative is the bilateral filter [10,11], which filters the image by regulating the value of each pixel by taking a weighted average of its neighboring pixels. In [12], a fast approximation of the bilateral filter was proposed by imposing the combined color sparseness prior and local statistics prior to the filtering input images. In [13], an accelerated bilateral filter was developed to efficiently compute filtering results via applying a KD-tree sampling algorithm. To achieve satisfactory filtering results, Gavaskar et al. [14] designed an improved bilateral filter with an adaptive Gaussian range kernel. Apart from the bilateral filter, the guided image filter proposed by He et al. in [15] also has a remarkable filtering capability; it derives the filtering result from the guided image using a local linear model. By considering the structural differences between guided and input images, Ham et al. presented an image filtering method in [16] which uses static and dynamic guidance. Sun et al. developed a weighted guided image filter (WGIF) [17] to achieve better-smoothing performance by incorporating edge-aware weighting operators into the filtering process. Inspired by unsharp masking, Shi et al. presented a simplified guided filter in [18] based on filtering prior from a low-pass filter.
Global methods: Different from the local methods, global methods [29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44] formulate the filtering tasks as a global optimization problem consisting of a fidelity term and a regulation term. Typically, the image filtering task is accomplished via building a global weighted least square model to suppress the noise in the image, as in [29]. Min et al. proposed an edge-preserving image smoothing filter [30], which solves the optimization problem by designing a series of global linear sub-systems. In [31], the side window filtering technique was proposed to effectively preserve the edge information. Xu et al. present an effective approach [45] for structure–texture image decomposition, leveraging the discriminative patch recurrence to develop a nonlocal transform that can better distinguish and sparsify texture components. Motivated by the rapid development of machine learning theory, many global filtering methods [3,46,47,48,49,50,51,52,53,54,55,56,57] have been designed under the deep learning architecture. For instance, a CNN-based joint filter [46] was proposed by selectively taking advantage of the structure information consistent with both guided and input images. The Self2self [3] introduces a self-supervised learning method for image denoising, which trains a denoising network solely on the input noisy image, using dropout on pairs of Bernoulli-sampled instances of the image. A fully convolutional neural (FCN)-based image filter method was developed in [58], which determined the learnable parameter through end-to-end training. However, these methods suffer from poor deployability because trained parameters for one specific application can hardly be applied to other different categories of smoothing tasks, making it difficult to put them into practical use at the current stage.
In this paper, we observe that the pixel in the filtered result tends to be obtained by averaging the similar pixels nearby. Motivated by this observation, we first propose a simple image segmentation method called recursive dilation segmentation (RDS) to partition the image into several regions with similar pixels inside, which directly respects the intrinsic spatial information between two adjacent pixels. Then, constrained by the segmented results obtained from RDS, a local adaptive filtering technique named local adaptive image filtering (LAIF) is developed to filter the input image by selectively utilizing the local pixels. Benefiting from the effective RDS, the filtering process, i.e., to smooth the flat regions and preserve the structure information, can be easily realized under the guidance of segmented results. Different from previous works, the proposed LAIF makes full use of the integrated information with respect to pixel intensity, hue component, and especially spatial adjacent information, which makes the filtered results more accurate and reasonable.
The paper is organized as follows. The motivation of this work is introduced in Section 2 and LAIF is proposed in Section 3. This is followed by the experimental results of the LAIF and state-of-the-art methods in Section 4. Finally, the conclusions are provided in Section 5.

2. Motivation

Generally, the task of image filtering is to design a linear translation-variant from a filtering input image p to an output image q under the guidance of a guided image I, which can be defined as:
q x = y ω W xy I p y
where x and y are pixel indexes, W xy I is a designed kernel weight to cope with the linear relationship between p and q, and ω is an N-by-N local patch. This linear translation-variant model is employed in the proposed LAIF and there is no doubt that the design of the kernel weight is extremely important in the translation-variant model.
The inspiration of the design of the kernel weight in this work comes from two two classic edge-preserving filters, i.e., bilateral filtering (BF) [10] and guided image filtering (GIF) [15]. Before introducing the proposed LAIF, it is necessary to review these two classic filtering algorithms.
In bilateral filtering, W xy depends on the distance between pixels and the pixel intensity differences. Mathematically, it can be formulated as
W xy B F I = 1 K x exp x y 2 σ d 2 I x I y 2 σ r 2
where K x is a normalization coefficient, and σ d and σ r are the standard deviations in the spatial domain and range domain, respectively.
Figure 1a shows the filtering result of BF. Notably, the filtering ability of BF is remarkable, which can be explained as follows: according to Equation (2), the output pixel is more likely to be influenced by the nearby pixels with similar distance information and the pixel intensity information (see the weight kernel shown in the figure), thus, the pixels in flat regions are smoothed by the similar pixels nearby and the edges are preserved.
Different from BF, GIF is developed based on a common assumption: the filtering output q is a linear transform of the guided image I in a local window, which can be expressed as
q x = a k I x + b k , x ω k
where ω k is a local window centered at pixel k , and a k and b k are two constant coefficients in ω k .
By formulating the filtering task into an optimization problem, the solution is given by
a k = 1 ω x ω k I x p x μ k p ¯ k σ k 2 + ϵ
b k = p k a k μ k
where μ k and σ k 2 are the mean and variance of the guided image in ω k , p k is the mean of the input filtering image in ω k , · is an operator returning the number of pixels inside, and ϵ is a regularization parameter used in optimization problem. The kernel weight can be obtained by averaging a k and b k to a ¯ i and b ¯ i .
Figure 1b gives the filtering result of GIF. As analyzed in [15], GIF has a strong filtering capability: for pixels in the “low-variance” area, where the pixels have similar intensity information, their value becomes the average of the neighboring pixels, while for pixels in the “high-variance” area, their pixel values are nearly unchanged.
As mentioned above, both BF and GIF smooth the image based on the local pixels with similar properties with respect to distance information or intensity information. Therefore, for the pixel to be smoothed, only the similar pixels nearby need to be utilized to obtain the filtering result. This motivates us to segment the image into several regions, where the pixels in one partition region share a similar property, before locally smoothing the image based on the segmented results.

3. Proposed Method

In this section, a simple but effective image filter technique, i.e., LAIF is proposed. Only two modules are included in LAIF, i.e., the recursive dilation segmentation module and the local adaptive image smoothing module.

3.1. Recursive Dilation Segmentation Module

Before introducing recursive dilation segmentation (RDS), we first define the dissimilarity between two pixels in the guided image, which is an RGB-channel image I in this work, by considering both pixel intensity distance and hue distance, and mathematically describe it as:
δ x , y = I x I y 2 + λ · H x H y 2
where δ x , y is the quantifying dissimilarity between x and y , H is the hue component of RGB image I , and λ is a balance weight parameter that makes a trade-off between two different categories of distance. In this way, two pixels can be defined as ‘similar’ if
δ x , y < ε
where ε is a threshold parameter that provides an estimation of the maximum possible δ between two similar pixels.
Given the criteria for judging whether two pixels are similar, the image can be easily segmented via RDS.
The fundamental idea of RDS is to partition the image into several segmentation regions in which each pixel belonging to the same segmentation part has similar spatial information and pixel intensity, in a recursive dilation way. More specifically, the image will be gradually partitioned into several non-overlapping regions via several segmentation operations. For the ith segmentation operation, we firstly select one pixel that has not been traversed in the image as the dilating center pixel, then, all its similar neighboring pixels will be judged according to Equation (7) before recursively dilating to them and treating them as the new dilating center pixels. The dilation procedure in each segmentation operation will continue recursively until all the adjacent similar pixels have been traversed. All the pixels traversed within the same ith segmentation operation will be stored into the same partition pixel set S i . This operation will be executed multiple times until all the image’s pixels are traversed. The segmentation process is described in Figure 2 and Algorithm 1.
Algorithm 1: Proposed LAIF
Sensors 23 05776 i001
It is worth mentioning that the proposed RDS algorithm is different from the clustering-based image segmentation methods presented in [59,60] because RDS directly applies spatial adjacent information in a gradual extension way. Benefiting from this property, the partitioned pixels in the same pixel set are more analogous with respect to both spatial information and color information.

3.2. Local Adaptive Image Smoothing Module

Different from the linear translation-variant model expressed in Equation (1), in this work we make an assumption that the pixels in the filtering output are only related to the pixels with similar properties within a local window in the filtering input image, which can be formulated as
q x = y S x ω x W xy ( I ) · p y
where ω x is a local patch with the size s of N × N centered at pixel x , and S x is a partition pixel set that the centering pixel x belongs to. In GIF, the pixels will be smoothed by a local average operation if they are surrounded by similar pixels. Accordingly, a local average operation is employed in LAIF as well, to smooth the target pixels by the similar pixels nearby, thus, the filtering kernel weight can be obtained as
W xy x = 1 S x ω x , y S x ω x
The essence of this weight parameter is to selectively utilize local information for local adaptive smoothing. The benefit of doing so is that the pixel information with respect to spatial distance, pixel intensity, and the hue component can be effectively integrated and utilized. Thus, by substituting Equation (9) into Equation (8), the filtering result can be obtained as
q x = y S x ω x 1 S x ω x · p y
and the whole algorithm is illustrated in Figure 3.
As shown in Figure 4, LAIF has a remarkable capability to smooth the flat regions and preserve edge information in contour regions. The filtering capability of LAIF can be explained intuitively as follows. Considering the pixels in the flat regions surrounded by many similar pixels, the local average operation will be employed on them, thus, the corresponding regions can be smoothed as shown in Figure 4a. As for the edge regions, two pixels on different sides are bound to belong to different pixel sets because of the distinct color and intensity information. This means pixels across the edge are not averaged together, making conspicuous edges emerge between them, as shown in Figure 4b, so that the edge information is well preserved. However, if we directly apply filtering to the overall segmentation results, the final outcome is quite likely to be influenced by poor segmentation results. Therefore, we have employed the method of iteration between segmentation and filtering. In the joint iterative process of segmentation and filtering, each step of optimization can provide a more accurate input for the next step. In this way, segmentation and filtering can influence and optimize each other in every iteration, resulting in an overall output that comes increasingly close to the optimal solution. This process is illustrated through Algorithm 1.

4. Experiments

In this section, a series of experiments in various computer vision tasks are conducted to evaluate the performance of LAIF compared with state-of-the-art methods, i.e., GIF [15], JBF [19], EBF [20], and SDF [16]. The parameters used in the compared methods were adjusted according to the corresponding references. All the experiments were conducted in MATLAB R2020a, on a PC with Intel(R) Core(TM) i7-8750H CPU @ 2.20 GHz (12 CPUs), 2.2 GHz.

4.1. Parameter Study

In this section, we conduct an experiment to study two key parameters in this method, i.e., pixel distance threshold ε and local filtering patch size s. According to Equation (7), ε determines the criteria for judging whether two pixels are similar, the image will be partitioned into fewer but larger regions with the increase in ε . As for the local filtering patch size, it determines the number of selected pixels when implementing local adaptive filtering. A larger local filtering patch size makes more pixels with similar properties be averaged together according to Equation (10). To investigate the effect of ε and the size of the filtering patch s, we filter two images under the guidance of themselves using LAIF with different values of ε and different sizes of the filtering patch. The different filtered results are given in Figure 5. As demonstrated from the figure, LAIF with a larger ε tends to show a stronger smoothing ability, effectively removing the unwanted noise in the texture region, while the results via a larger filtering patch size seem more uniform and smooth. Therefore, the filtering ability of LAIF can be changed flexibly by adjusting the values of ε and s, enabling LAIF to be applied to different filtering tasks.

4.2. Dehazing

The task of image dehazing is to remove haze from a single image. In traditional methods [61], the coarse transmission can be estimated via the patch-wise prior and optimized by an edge-preserving filter to obtain the accurate transmission map. In this section, we consider the hazy image as the guided image and optimize the coarse transmission via different filtering methods. Figure 6 shows the dehazed images restored by the corresponding optimized transmission map. Notably, there exist some halo effects near the edge regions for GIF, JBF, EBF, and SDF (see the highlighted zoomed-in panels), which means that the block effect is not eliminated effectively by the filtering method. On the contrary, LAIF can effectively optimize the transmission map under the guidance of input hazy images. We selected two publicly available datasets, I-HAZE [62] and O-HAZE [63], and used the DCP [61] method to perform haze removal. The results were evaluated using the SSIM and PSNR metrics. The results are shown in Table 1.
For the I-HAZE dataset, our proposed method is among the top performers in both SSIM and PSNR. With an SSIM score of 0.7274, it slightly trails EBF’s 0.7297 but surpasses the performance of GIF, GBF, and SDF. In terms of PSNR, our method outperforms all others, with a score of 28.1024, slightly edging out EBF’s 28.0868. When it comes to the O-HAZE dataset, our method maintains its strong performance. It leads in SSIM scoring, achieving 0.7124, which is marginally better than EBF’s 0.7002. With PSNR, our method again secures one of the top two positions. It scores 28.2481, slightly higher than GBF’s 28.2235. In summary, across both datasets and two distinct metrics our proposed method demonstrates consistently strong performance, often achieving or being very close to the best score among all of the tested methods.

4.3. Denoising

In this section, we investigate the denoising performance of LAIF via denoising a no-flash noisy image under the guidance of the corresponding flash image. A comparison between the proposed LAIF and state-of-the-art methods, i.e., GIF, JBF, EBF, and SDF is provided in Figure 7. Visually, the edge texture information of the recovered results via EBF and SDF are prone to blurriness. For the denoising result by JBF, some texture details are removed, which does not correspond to the guided image. In comparison, our recovered results are clean and the structure information is well preserved. We conducted experiments related to image denoising, including the evaluation of flash/no-flash denoising and standard denoising. We selected the FAID [64] dataset for our experiments. These images from the FAID dataset are a subset of the Flash and Ambient Illuminations Dataset. In the evaluation process of flash/no-flash denoising, we added Gaussian noise to the ambient light images and used the flash images as guidance. Additionally, we also employed the standard denoising methods (Gaussian filter denoising [65], median filter denoising, total variation denoising [66]) for direct denoising operations.
Table 2 evaluates various denoising methods using SSIM and PSNR metrics. The proposed LAIF method excels in flash/no-flash denoising, achieving the highest scores (SSIM: 0.7588, PSNR: 28.5654), indicating superior structure preservation and noise reduction. In standard denoising, LAIF remains competitive, surpassing Gaussian filter and total variation in SSIM and nearly matching median filter in PSNR.

4.4. Detail Enhancement

The fundamental idea of detail enhancement is to magnify the high-frequency detail layer, i.e., the difference between the input image and the filtering output image. Figure 8 gives the enhanced images obtained by the combination of the boosted detail layer (three times magnification) and the input image. As shown in the figure, JBF and SDF tend to introduce a gradient reversal effect into the results. In comparison, the enhanced results via the proposed LAIF can highlight the detail information in the images without producing negative visual effects, e.g., halo artifacts or gradient reversal. In the evaluation of detail enhancement, we selected the publicly available LIVE Release2 [67] dataset. This dataset contains various types of distorted images. We enhanced the details of images distorted by JPEG, JPEG2000, and Gaussian blur, and then evaluated them using the SSIM and PNSR metrics. The results are shown in Table 3. Our proposed method generally exhibits superior performance in the SSIM metric across all datasets. This suggests that our method is highly effective in enhancing structural details in images while preserving their natural appearance. In conclusion, these results overall suggest that our method consistently delivers strong performance in detail enhancement.

4.5. Edge Preserving

Edge preserving is a necessary ability for ‘good’ image filtering. We first obtain segmented guidance images and segmentation information through recursive dilation. These are then fed into the filter to perform filtering operations on the image. The filtered result represents the edge-preserved outcome. For other methods, we used the official smoothing techniques provided to assess their edge-preservation effects. To evaluate the edge-preserving ability of the proposed LAIF, we provide the filtering results of state-of-the-art filtering methods, and the corresponding one-dimensional illustrations, in Figure 9. It can be observed from the figure that GIF, JBF, and EBF are prone to streak artifacts, deviating the edges from their original shapes. In contrast, the filtered results via the proposed LAIF are free of halo effects and the edge information is well preserved. Moreover, LAIF can smooth the noise in flat regions, as shown in the result. Because the effect of edge preservation cannot be directly quantified and the effect of edge preservation is often reflected in its dehazing ability, we can evaluate the edge preservation and dehazing ability together. This is because in traditional dehazing methods, the better the smoothness and edge preservation when estimating the transmittance, the better the dehazing effect. Therefore, the effect of edge preservation is also reflected in its dehazing ability. The results are shown in Table 1.

5. Conclusions

In this paper, we observe that the pixel in a filtered result can be obtained using similar pixels nearby. By applying the recursive dilation segmentation method, an image can be partitioned into several segmentation regions, the pixels within each of which share a similar property. The main advantage of this method is that it can directly make use of the intrinsic spatial adjacent information between two neighboring pixels in a recursive dilation manner. Moreover, the pixel intensity information and hue component information are also considered in the dilation process. Benefiting from this image segmentation method, the unwanted texture details in the flat regions can be effectively smoothed while the structure information in the edge regions can be well preserved, via employing a local adaptive image filter derived from RDS. In addition, the application of LAIF in the remote sensing field has achieved outstanding results, specifically in areas such as image dehazing, denoising, enhancement, and edge preserving, among others. We investigate the filtering ability of LAIF by conducting a series of experiments between LAIF and the state-of-the-art methods, demonstrating that it performs favorably against the others in terms of both filtering quality and generic ability.

Author Contributions

Methodology, D.Z. and J.Z.; Formal analysis, C.C. and K.C.; Writing—original draft, J.Z.; Project administration, M.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 61902198; in part by the Natural Science Foundation of Jiangsu Province under Grant BK20190730; in part by the Research Foundation of Nanjing University of Posts and Telecommunications under Grant NY219135; and in part by the Key Laboratory of Radar Imaging and Microwave Photonics, Ministry of Education, Nanjing University of Aeronautics and Astronautics.

Institutional Review Board Statement

Did not require ethical approval.

Informed Consent Statement

The study did not involve humans.

Data Availability Statement

Due to privacy concerns, it is inconvenient to provide.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ma, R.; Hu, H.; Xing, S.; Li, Z. Efficient and Fast Real-World Noisy Image Denoising by Combining Pyramid Neural Network and Two-Pathway Unscented Kalman Filter. IEEE Trans. Image Process. 2020, 29, 3927–3940. [Google Scholar] [CrossRef] [PubMed]
  2. Cao, Y.; Yang, M.Y.; Tisse, C.L. Effective Strip Noise Removal for Low-Textured Infrared Images Based on 1-D Guided Filtering. IEEE Trans. Circuits Syst. Video Technol. 2016, 26, 2176–2188. [Google Scholar] [CrossRef]
  3. Quan, Y.; Chen, M.; Pang, T.; Ji, H. Self2Self with Dropout: Learning Self-Supervised Denoising From Single Image. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1887–1895. [Google Scholar] [CrossRef]
  4. Li, S.; Kang, X.; Hu, J. Image Fusion with Guided Filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar] [CrossRef] [PubMed]
  5. Li, Z.; Zheng, J. Single Image De-Hazing Using Globally Guided Image Filtering. IEEE Trans. Image Process. 2018, 27, 442–450. [Google Scholar] [CrossRef]
  6. Liu, J.; Wang, S.; Wang, X.; Ju, M.; Zhang, D. A Review of Remote Sensing Image Dehazing. Sensors 2021, 21, 3926. [Google Scholar] [CrossRef]
  7. He, Z.; Gong, C.; Hu, Y.; Li, L. Remote Sensing Image Dehazing Based on an Attention Convolutional Neural Network. IEEE Access 2022, 10, 68731–68739. [Google Scholar] [CrossRef]
  8. Jiang, B.; Chen, G.; Wang, J.; Ma, H.; Wang, L.; Wang, Y.; Chen, X. Deep Dehazing Network for Remote Sensing Image with Non-Uniform Haze. Remote Sens. 2021, 13, 4443. [Google Scholar] [CrossRef]
  9. Kaplan, N.H.; Erer, I.; Kumlu, D. Image Enhancement Methods for Remote Sensing: A Survey. In Recent Remote Sensing Sensor Applications; Marghany, M., Ed.; IntechOpen: Rijeka, Croatia, 2021; Chapter 5. [Google Scholar] [CrossRef]
  10. Tomasi, C.; Manduchi, R. Bilateral filtering for gray and color images. In Proceedings of the Sixth International Conference on Computer Vision (IEEE Cat. No.98CH36271), Bombay, India, 7 January 1998; pp. 839–846. [Google Scholar] [CrossRef]
  11. Petschnigg, G.; Szeliski, R.; Agrawala, M.; Cohen, M.; Hoppe, H.; Toyama, K. Digital photography with flash and no-flash image pairs. ACM Trans. Graph. (TOG) 2004, 23, 664–672. [Google Scholar] [CrossRef] [Green Version]
  12. Mozerov, M.G.; van de Weijer, J. Global Color Sparseness and a Local Statistics Prior for Fast Bilateral Filtering. IEEE Trans. Image Process. 2015, 24, 5842–5853. [Google Scholar] [CrossRef]
  13. Adams, A.; Gelfand, N.; Dolson, J.; Levoy, M. Gaussian KD-Trees for Fast High-Dimensional Filtering. In Proceedings of the ACM SIGGRAPH 2009 Papers, Association for Computing Machinery, SIGGRAPH ’09, New York, NY, USA, 3–7 August 2009. [Google Scholar] [CrossRef] [Green Version]
  14. Gavaskar, R.G.; Chaudhury, K.N. Fast Adaptive Bilateral Filtering. IEEE Trans. Image Process. 2019, 28, 779–790. [Google Scholar] [CrossRef] [Green Version]
  15. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef]
  16. Ham, B.; Cho, M.; Ponce, J. Robust Guided Image Filtering Using Nonconvex Potentials. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 192–207. [Google Scholar] [CrossRef] [Green Version]
  17. Sun, Z.; Han, B.; Li, J.; Zhang, J.; Gao, X. Weighted Guided Image Filtering With Steering Kernel. IEEE Trans. Image Process. 2020, 29, 500–508. [Google Scholar] [CrossRef]
  18. Shi, Z.; Chen, Y.; Gavves, E.; Mettes, P.; Snoek, C.G.M. Unsharp Mask Guided Filtering. IEEE Trans. Image Process. 2021, 30, 7472–7485. [Google Scholar] [CrossRef]
  19. Kopf, J.; Cohen, M.F.; Lischinski, D.; Uyttendaele, M. Joint Bilateral Upsampling. ACM Trans. Graph. 2007, 26, 96. [Google Scholar] [CrossRef]
  20. Liu, W.; Zhang, P.; Chen, X.; Shen, C.; Huang, X.; Yang, J. Embedding Bilateral Filter in Least Squares for Efficient Edge-Preserving Image Smoothing. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 23–35. [Google Scholar] [CrossRef] [Green Version]
  21. Guo, X.; Li, Y.; Ma, J.; Ling, H. Mutually Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 694–707. [Google Scholar] [CrossRef]
  22. Li, Z.; Zheng, J.; Zhu, Z.; Yao, W.; Wu, S. Weighted Guided Image Filtering. IEEE Trans. Image Process. 2015, 24, 120–129. [Google Scholar] [CrossRef]
  23. Deng, X.; Hasan, M.; Carr, N.; Xu, Z.; Marschner, S. Path graphs: Iterative path space filtering. ACM Trans. Graph. 2021, 40, 276:1–276:15. [Google Scholar] [CrossRef]
  24. Ochotorena, C.N.; Yamashita, Y. Anisotropic Guided Filtering. IEEE Trans. Image Process. 2020, 29, 1397–1412. [Google Scholar] [CrossRef]
  25. Ju, M.; Ding, C.; Ren, W.; Yang, Y.; Zhang, D.; Guo, Y.J. IDE: Image Dehazing and Exposure Using an Enhanced Atmospheric Scattering Model. IEEE Trans. Image Process. 2021, 30, 2180–2192. [Google Scholar] [CrossRef]
  26. Belyaev, A.; Fayolle, P.A. Adaptive Curvature-Guided Image Filtering for Structure + Texture Image Decomposition. IEEE Trans. Image Process. 2018, 27, 5192–5203. [Google Scholar] [CrossRef] [PubMed]
  27. Ju, M.; Ding, C.; Guo, C.A.; Ren, W.; Tao, D. IDRLP: Image Dehazing Using Region Line Prior. IEEE Trans. Image Process. 2021, 30, 9043–9057. [Google Scholar] [CrossRef]
  28. Wu, L.; Jong, C.C. A High-Throughput VLSI Architecture for Real-Time Full-HD Gradient Guided Image Filter. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 1868–1877. [Google Scholar] [CrossRef]
  29. Farbman, Z.; Fattal, R.; Lischinski, D.; Szeliski, R. Edge-preserving decompositions for multi-scale tone and detail manipulation. ACM Trans. Graph. (TOG) 2008, 27, 1–10. [Google Scholar] [CrossRef]
  30. Min, D.; Choi, S.; Lu, J.; Ham, B.; Sohn, K.; Do, M.N. Fast Global Image Smoothing Based on Weighted Least Squares. IEEE Trans. Image Process. 2014, 23, 5638–5653. [Google Scholar] [CrossRef] [PubMed]
  31. Yin, H.; Gong, Y.; Qiu, G. Side Window Filtering. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–17 June 2019; pp. 8750–8758. [Google Scholar] [CrossRef]
  32. Berger, T.; Stromberg, J.; Eltoft, T. Adaptive regularized constrained least squares image restoration. IEEE Trans. Image Process. 1999, 8, 1191–1203. [Google Scholar] [CrossRef]
  33. Hirakawa, K.; Parks, T. Image denoising using total least squares. IEEE Trans. Image Process. 2006, 15, 2730–2742. [Google Scholar] [CrossRef]
  34. Mesarovic, V.; Galatsanos, N.; Katsaggelos, A. Regularized constrained total least squares image restoration. IEEE Trans. Image Process. 1995, 4, 1096–1108. [Google Scholar] [CrossRef] [Green Version]
  35. Gendreau, T. Process management issues in message-based multiprocessors. In Proceedings of the Conference Proceedings ’88, IEEE Southeastcon, Knoxville, TN, USA, 10–13 April 1988; pp. 103–107. [Google Scholar] [CrossRef]
  36. Zhou, Z.; Wang, B.; Ma, J. Scale-Aware Edge-Preserving Image Filtering via Iterative Global Optimization. IEEE Trans. Multimed. 2018, 20, 1392–1405. [Google Scholar] [CrossRef]
  37. Caraffa, L.; Tarel, J.P.; Charbonnier, P. The Guided Bilateral Filter: When the Joint/Cross Bilateral Filter Becomes Robust. IEEE Trans. Image Process. 2015, 24, 1199–1208. [Google Scholar] [CrossRef] [Green Version]
  38. Gamboa, L.E.; Guertin, J.; Nowrouzezahrai, D. Scalable appearance filtering for complex lighting effects. ACM Trans. Graph. 2018, 37, 277:1–277:13. [Google Scholar] [CrossRef]
  39. Sugimoto, K.; Kamata, S.I. Compressive Bilateral Filtering. IEEE Trans. Image Process. 2015, 24, 3357–3369. [Google Scholar] [CrossRef]
  40. Papari, G.; Idowu, N.; Varslot, T. Fast Bilateral Filtering for Denoising Large 3D Images. IEEE Trans. Image Process. 2017, 26, 251–261. [Google Scholar] [CrossRef]
  41. He, S.; Yang, Q.; Lau, R.W.H.; Yang, M.H. Fast Weighted Histograms for Bilateral Filtering and Nearest Neighbor Searching. IEEE Trans. Circuits Syst. Video Technol. 2016, 26, 891–902. [Google Scholar] [CrossRef]
  42. Ong, F.; Milanfar, P.; Getreuer, P. Local Kernels That Approximate Bayesian Regularization and Proximal Operators. IEEE Trans. Image Process. 2019, 28, 3007–3019. [Google Scholar] [CrossRef] [Green Version]
  43. Nehab, D.F.; Maximo, A. Parallel recursive filtering of infinite input extensions. ACM Trans. Graph. 2016, 35, 204:1–204:13. [Google Scholar] [CrossRef]
  44. Hung, K.W.; Siu, W.C. Robust Soft-Decision Interpolation Using Weighted Least Squares. IEEE Trans. Image Process. 2012, 21, 1061–1069. [Google Scholar] [CrossRef]
  45. Xu, R.; Xu, Y.; Quan, Y. Structure-Texture Image Decomposition Using Discriminative Patch Recurrence. IEEE Trans. Image Process. 2021, 30, 1542–1555. [Google Scholar] [CrossRef]
  46. Li, Y.; Huang, J.B.; Ahuja, N.; Yang, M.H. Joint Image Filtering with Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 1909–1923. [Google Scholar] [CrossRef] [Green Version]
  47. Han, R.; Feng, W.; Wang, S. Fast Learning of Spatially Regularized and Content Aware Correlation Filter for Visual Tracking. IEEE Trans. Image Process. 2020, 29, 7128–7140. [Google Scholar] [CrossRef]
  48. Chen, S.; Duan, C.; Yang, Y.; Li, D.; Feng, C.; Tian, D. Deep Unsupervised Learning of 3D Point Clouds via Graph Topology Inference and Filtering. IEEE Trans. Image Process. 2020, 29, 3183–3198. [Google Scholar] [CrossRef]
  49. Tian, L.; Tu, Z.; Zhang, D.; Liu, J.; Li, B.; Yuan, J. Unsupervised Learning of Optical Flow With CNN-Based Non-Local Filtering. IEEE Trans. Image Process. 2020, 29, 8429–8442. [Google Scholar] [CrossRef] [PubMed]
  50. Li, T.; Xu, M.; Zhu, C.; Yang, R.; Wang, Z.; Guan, Z. A Deep Learning Approach for Multi-Frame In-Loop Filter of HEVC. IEEE Trans. Image Process. 2019, 28, 5663–5678. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Chen, J.; Adams, A.; Wadhwa, N.; Hasinoff, S.W. Bilateral guided upsampling. ACM Trans. Graph. 2016, 35, 203:1–203:8. [Google Scholar] [CrossRef] [Green Version]
  52. Huang, Z.; Sun, J.; Guo, X.; Shang, M. Adaptive Deep Reinforcement Learning-Based In-Loop Filter for VVC. IEEE Trans. Image Process. 2021, 30, 5439–5451. [Google Scholar] [CrossRef]
  53. Bianchi, F.M.; Grattarola, D.; Livi, L.; Alippi, C. Graph Neural Networks with Convolutional ARMA Filters. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3496–3507. [Google Scholar] [CrossRef] [PubMed]
  54. Wang, D.; Xia, S.; Yang, W.; Liu, J. Combining Progressive Rethinking and Collaborative Learning: A Deep Framework for In-Loop Filtering. IEEE Trans. Image Process. 2021, 30, 4198–4211. [Google Scholar] [CrossRef]
  55. Liu, J.; Guo, G. Vehicle Localization During GPS Outages With Extended Kalman Filter and Deep Learning. IEEE Trans. Instrum. Meas. 2021, 70, 1–10. [Google Scholar] [CrossRef]
  56. Adams, A. Fast median filters using separable sorting networks. ACM Trans. Graph. 2021, 40, 70:1–70:11. [Google Scholar] [CrossRef]
  57. Feng, L.; Cai, Y.; Wei, E.; Li, J. Graph neural networks with global noise filtering for session-based recommendation. Neurocomputing 2022, 472, 113–123. [Google Scholar] [CrossRef]
  58. Zhang, W.; Jiao, L.; Li, Y.; Liu, J. Sparse Learning-Based Correlation Filter for Robust Tracking. IEEE Trans. Image Process. 2021, 30, 878–891. [Google Scholar] [CrossRef]
  59. Chen, C.; Luo, J.; Parker, K. Image segmentation via adaptive K-mean clustering and knowledge-based morphological operations with biomedical applications. IEEE Trans. Image Process. 1998, 7, 1673–1683. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  60. Rezaee, M.; van der Zwet, P.; Lelieveldt, B.; van der Geest, R.; Reiber, J. A multiresolution image segmentation technique based on pyramidal segmentation and fuzzy clustering. IEEE Trans. Image Process. 2000, 9, 1238–1248. [Google Scholar] [CrossRef] [PubMed]
  61. He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar] [CrossRef]
  62. Ancuti, C.O.; Ancuti, C.; Timofte, R.; Vleeschouwer, C.D. I-HAZE: A dehazing benchmark with real hazy and haze-free indoor images. arXiv 2018, arXiv:1804.05091v1. [Google Scholar]
  63. Ancuti, C.O.; Ancuti, C.; Timofte, R.; Vleeschouwer, C.D. O-HAZE: A dehazing benchmark with real hazy and haze-free outdoor images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, NTIRE Workshop, 2018, NTIRE CVPR’18, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  64. Aksoy, Y.; Kim, C.; Kellnhofer, P.; Paris, S.; Elgharib, M.; Pollefeys, M.; Matusik, W. A Dataset of Flash and Ambient Illumination Pairs from the Crowd. In Proceedings of the 15th European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
  65. Yang, P.; Lu, W.; Liu, K.; Zhang, G. Image enhancement using smoothing with a Gaussian filter. Pattern Recognit. Lett. 2010, 31, 406–412. [Google Scholar]
  66. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Phys. Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  67. Sheikh, H.R.; Sabir, M.F.; Bovik, A.C. A Statistical Evaluation of Recent Full Reference Image Quality Assessment Algorithms. IEEE Trans. Image Process. 2006, 15, 3440–3451. [Google Scholar] [CrossRef]
Figure 1. Demonstration of filtering processes of BF and GIF. (a) The filtering process of BF and an enlarged section of the result. The red frames show the filtering weight kernel of two pixels in the image (red means a larger weight and blue means a smaller weight). (b) The filtering process of GIF and an enlarged section of the result. The white words in the zoomed-in panel are the description of the panel.
Figure 1. Demonstration of filtering processes of BF and GIF. (a) The filtering process of BF and an enlarged section of the result. The red frames show the filtering weight kernel of two pixels in the image (red means a larger weight and blue means a smaller weight). (b) The filtering process of GIF and an enlarged section of the result. The white words in the zoomed-in panel are the description of the panel.
Sensors 23 05776 g001
Figure 2. Illustration of one segmentation operation. The squares with the same color denote pixels with similar properties based on Equation (7), the red frame is the traversed region, and the red arrows stand for the dilation directions.
Figure 2. Illustration of one segmentation operation. The squares with the same color denote pixels with similar properties based on Equation (7), the red frame is the traversed region, and the red arrows stand for the dilation directions.
Sensors 23 05776 g002
Figure 3. Illustration of the proposed LAIF.
Figure 3. Illustration of the proposed LAIF.
Sensors 23 05776 g003
Figure 4. Illustration of the segmented result and the corresponding filtering result of LAIF. The (a) is the interior of the segmentation result and The (b) is The boundary of the segmentation result.
Figure 4. Illustration of the segmented result and the corresponding filtering result of LAIF. The (a) is the interior of the segmentation result and The (b) is The boundary of the segmentation result.
Sensors 23 05776 g004
Figure 5. The results of LAIF using different values of ε and sizes of filtering patch s.
Figure 5. The results of LAIF using different values of ε and sizes of filtering patch s.
Sensors 23 05776 g005
Figure 6. Dehazing results obtained by the optimized transmission map via LAIF and other state-of-the-art methods, and the corresponding zoomed-in patches in the blue frames. (a) Hazy images as the guided images. (b) Coarse transmission map as the input images. (c) GIF. (d) JBF. (e) EBF. (f) SDF. (g) LAIF.
Figure 6. Dehazing results obtained by the optimized transmission map via LAIF and other state-of-the-art methods, and the corresponding zoomed-in patches in the blue frames. (a) Hazy images as the guided images. (b) Coarse transmission map as the input images. (c) GIF. (d) JBF. (e) EBF. (f) SDF. (g) LAIF.
Sensors 23 05776 g006
Figure 7. Filtering results via LAIF and other state-of-the-art methods, and the corresponding zoomed-in patches in the blue frames. (a) Flash images as guided images. (b) No-flash images as input images. (c) GIF. (d) JBF. (e) EBF. (f) SDF. (g) LAIF.
Figure 7. Filtering results via LAIF and other state-of-the-art methods, and the corresponding zoomed-in patches in the blue frames. (a) Flash images as guided images. (b) No-flash images as input images. (c) GIF. (d) JBF. (e) EBF. (f) SDF. (g) LAIF.
Sensors 23 05776 g007
Figure 8. Detail enhancement results via LAIF and other state-of-the-art methods, and the corresponding zoomed-in patches in the blue frames. (a) Guided images. (b) Input images. (c) GIF. (d) JBF. (e) EBF. (f) SDF. (g) LAIF.
Figure 8. Detail enhancement results via LAIF and other state-of-the-art methods, and the corresponding zoomed-in patches in the blue frames. (a) Guided images. (b) Input images. (c) GIF. (d) JBF. (e) EBF. (f) SDF. (g) LAIF.
Sensors 23 05776 g008
Figure 9. Filtering results via LAIF and other state-of-the-art methods, and the 1D detail illustration on the blue scanline (black corresponds to the input image, red corresponds to the output image). (a) Guided images. (b) Input images. (c) GIF. (d) JBF. (e) EBF. (f) SDF. (g) LAIF.
Figure 9. Filtering results via LAIF and other state-of-the-art methods, and the 1D detail illustration on the blue scanline (black corresponds to the input image, red corresponds to the output image). (a) Guided images. (b) Input images. (c) GIF. (d) JBF. (e) EBF. (f) SDF. (g) LAIF.
Sensors 23 05776 g009
Table 1. Comparative analysis of methods (EBF, GIF, GBF, SDF, ours) on two public haze datasets (I-HAZE and O-HAZE) using the SSIM and PSNR metrics. The top two performing methods are underlined.
Table 1. Comparative analysis of methods (EBF, GIF, GBF, SDF, ours) on two public haze datasets (I-HAZE and O-HAZE) using the SSIM and PSNR metrics. The top two performing methods are underlined.
DatasetMetricMethods
EBFGIFGBFSDFOurs
I-HAZESSIM0.72970.72050.72480.62170.7274
PSNR28.086828.038928.083427.802728.1024
O-HAZESSIM0.70020.68910.68870.61170.7124
PSNR28.127128.169128.223527.854828.2481
Table 2. Flash/no-flash denoising and standard denoising capability evaluation based on SSIM and PSNR metrics. ‘w/o’ means without.
Table 2. Flash/no-flash denoising and standard denoising capability evaluation based on SSIM and PSNR metrics. ‘w/o’ means without.
Metricw/o DenoisingFlash/No-Flash DenoisingStandard Denoising
EBFGIFGBFSDFOursGFMFTVOurs
SSIM0.40220.72410.71050.73460.73120.75880.69780.63250.55680.7018
PSNR28.048427.733228.223227.484727.568928.565427.652328.033527.840227.9041
Table 3. Comparative analysis of methods (EBF, GIF, GBF, SDF, ours) on three datasets (gblur, jp2k, jpeg) of LIVE Release2 using the SSIM and PSNR metrics. The top two performing methods are underlined.
Table 3. Comparative analysis of methods (EBF, GIF, GBF, SDF, ours) on three datasets (gblur, jp2k, jpeg) of LIVE Release2 using the SSIM and PSNR metrics. The top two performing methods are underlined.
DatasetMetricMethods
EBFGIFGBFSDFOurs
gblurSSIM0.76540.76090.67240.73940.7945
PSNR33.767829.113429.368630.015131.1163
jp2kSSIM0.69120.70480.57960.62830.7267
PSNR29.422528.966829.057529.427829.541
jpegSSIM0.65340.67180.52380.56940.6912
PSNR28.997628.760328.754229.028828.8364
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Chen, C.; Chen, K.; Ju, M.; Zhang, D. Local Adaptive Image Filtering Based on Recursive Dilation Segmentation. Sensors 2023, 23, 5776. https://doi.org/10.3390/s23135776

AMA Style

Zhang J, Chen C, Chen K, Ju M, Zhang D. Local Adaptive Image Filtering Based on Recursive Dilation Segmentation. Sensors. 2023; 23(13):5776. https://doi.org/10.3390/s23135776

Chicago/Turabian Style

Zhang, Jialiang, Chuheng Chen, Kai Chen, Mingye Ju, and Dengyin Zhang. 2023. "Local Adaptive Image Filtering Based on Recursive Dilation Segmentation" Sensors 23, no. 13: 5776. https://doi.org/10.3390/s23135776

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop