Next Article in Journal
Proactive Handover Decision for UAVs with Deep Reinforcement Learning
Previous Article in Journal
Concept of Fluorescent Transport Activity Biosensor for the Characterization of the Arabidopsis NPF1.3 Activity of Nitrate
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Dehazing Using LiDAR Generated Grayscale Depth Prior

1
Department of Aerospace Engineering, Automation and System Research Institute, Seoul National University, Seoul 08826, Korea
2
School of Mechanical Convergence System Engineering, Kunsan National University, Gunsan 54150, Korea
3
Department of Mechanical System Engineering (Department of Aeronautics, Mechanical and Electronic Convergence Engineering), Kumoh National Institute of Technology, Gumi 39177, Korea
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(3), 1199; https://doi.org/10.3390/s22031199
Submission received: 28 December 2021 / Revised: 28 January 2022 / Accepted: 2 February 2022 / Published: 5 February 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
In this paper, the dehazing algorithm is proposed using a one-channel grayscale depth image generated from a LiDAR point cloud 2D projection image. In depth image-based dehazing, the estimation of the scattering coefficient is the most important. Since scattering coefficients are used to estimate the transmission image for dehazing, the optimal coefficients for effective dehazing must be obtained depending on the level of haze generation. Thus, we estimated the optimal scattering coefficient for 100 synthetic haze images and represented the distribution between the optimal scattering coefficient and dark channels. Moreover, through linear regression of the aforementioned distribution, the equation between scattering coefficients and dark channels was estimated, enabling the estimation of appropriate scattering coefficient. Transmission image for dehazing is defined with a scattering coefficient and a grayscale depth image, obtained from LiDAR 2D projection. Finally, dehazing is performed based on the atmospheric scattering model through the defined atmospheric light and transmission image. The proposed method was quantitatively and qualitatively analyzed through simulation and image quality parameters. Qualitative analysis was conducted through YOLO v3 and quantitative analysis was conducted through MSE, PSNR, SSIM, etc. In quantitative analysis, SSIM showed an average performance improvement of 24%.

1. Introduction

Haze is a phenomenon in which the visible distance is reduced due to dust, smoke particles, and polluting particles in the atmosphere. Particles in the atmosphere scatter light; thus, images obtained in these environments decrease contrast and, eventually, deteriorate visibility. Recently developed automatic navigation systems rely heavily on vision sensors [1]. If the input image is in poor condition, the overall system will suffer. Therefore, dehazing technology, which can obtain clear images, can benefit systems such as image classification [2,3,4,5], image recognition [6,7,8,9,10], visual odometry [11,12], and remote sensing [13,14,15].
Currently, the most commonly used sensors for robots and vehicles include light detection and ranging (LiDAR) and camera. The sensors allow the performance of visual odometry, LiDAR odometry, SLAM, autonomous navigation, etc. For these purposes, they can be used as visual-only [11,12,16,17], LiDAR-only [18,19], or fused [20]. When the LiDAR and visual are fused, the two sensors are used complementarily to increase robustness of the system [20]. Even with the increase in robustness, damage to the resulting values can occur if the input data obtained from the sensors are inherently in poor condition. Therefore, it is necessary to make quality input data to prevent this degradation.
We can attach cameras to numerous platforms such as vehicles, drones, and robots to get image data. If the weather allows, we can get clear images such as Figure 1a. In this case, dehazing is unnecessary for effective vision-based processes. However, in the event of haze due to smoke or fine dust in the atmosphere, such as in Figure 1b, utilizing such processes becomes challenging [21]. Therefore, the contrasts from images obtained in a hazed environment should be enhanced.
Each image pixel value in hazy images can be expressed with atmospheric scattering model, a linear combination of the pixel values from actual image, transmission image, and airlight [22,23,24]. Airlight and transmission images are required to perform scene radiance recovery through the scattering coefficient model.
Traditional algorithms use color attenuation prior [25] and dark channel prior [26] to create a transmission image, whereas recent research uses deep learning to perform dehazing. The proposed method succeeds traditional methods, with its contribution in utilizing depth image and scattering coefficient to perform dehazing.
Existing methods for obtaining depth images include using stereo camera or depth camera. More recent research adopts deep learning in obtaining depth images from monocular images through training of existing depth images [27]. In this paper, the depth image is obtained by 2D projection of LiDAR point cloud. In this way, by obtaining a depth image through LiDAR and performing dehazing on vision data, we would like to propose a more complementary and robust LiDAR–vision fusion system.
Our contribution is as follows: (1) Proposal of a depth image-based dehazing technique available in LiDAR–vision fusion systems; (2) Proposal of a scattering coefficient estimation technique through the DCM-scattering coefficient model.
An outline of the paper is as follows. Section 2 outlines the theoretical background of dehazing and the related works applied to the proposed method. Section 3 outlines the overall description of the proposed method, and Section 4 summarizes the analysis of the simulation results obtained through the proposed method. Finally, Section 5 briefly describes the conclusions, the limitations of the proposed method, and the future works for improving the limit.

2. Image Dehazing

2.1. Related Works

Image dehazing has always been a popular method to obtain clear images for image processing. Due to its popularity, numerous methods of dehazing have been proposed. Assumption-based and prior-based methods are typically used. Tan et al. [28] found that the contrast was higher for images without haze than those with haze. Thus, Tan et al. [28] performed single image dehazing by maximizing the local contrast for single images. Fattal et al. [29] discovered that pixels of image patches typically exhibit a one-dimensional distribution, and used it to recover the scene transmission. Huo et al. [30] performed dehazing with the white balance algorithm and the atmospheric illuminance prior. Zheng et al. [31] performed dehazing with patch adaptive structure decomposition and multi-exposure image fusion. He et al. [26] uses the assumption that pixels without haze tend to have a meager intensity value for at least one channel out of three RGB channels.
He defined this channel as the dark channel and used it to create a transmission image. The method from He is the most widely used and considered a standard in dehazing. Zhu et al. [25] proposed a color attenuation prior-based method to generate a depth image from a hazy image. Here, the transmission image used to obtain the dehazed image was obtained through the relationship between the depth image and the transmission image.
Recently, learning-based methods have also been proposed [32,33,34,35,36,37]. Cai et al. [32] proposed an end-to-end dehazing using a convolutional neural network (CNN) model. This was done by estimating the transmission image using the BReLU and Maxout activation functions. Ren et al. [33] performed dehazing by using multiscale CNN to estimate the transmission image. In addition, Li et al. proposed AOD-NET [34] and Dehze-cGAN [35] using the generative adversarial network.
In this paper, the proposed method uses a depth image to obtain the transmission image. During the process, the required scattering coefficient is obtained by estimating through the relationship between dark channel and scattering coefficient. Then dehazing is performed using the obtained transmission image.

2.2. Atmospheric Scatterming Model

Due to the light scattered by the atmosphere, and the atmospheric light, the hazy image looks blurry, as shown in Figure 2. This phenomenon can be explained by Equation (1) [22,23,24].
I ( x ) = J ( x ) t ( x ) + A ( x ) ( 1 t ( x ) )
where x represents a two-dimensional vector, comprised of the position of each pixel in the image. J ( x ) is an image before being distorted by haze, which is the ultimate result we want to obtain through the above equation. I ( x ) is the hazy image and t ( x ) is the transmission image, representing the proportion of light that reaches the camera through the atmosphere. A is airlight, and it is assumed that all pixels in the image have the same value.
From Equation (1), we can see that the information I ( x ) from the camera is lost as the actual information J ( x ) and the signal reflected from the target pass through the atmosphere, leaving only the J ( x ) t ( x ) level. In addition, A ( x ) ( 1 t ( x ) ) caused by atmospheric light sources is mixed, resulting in haze, shown in Figure 1b.
Through Equations (1) and (2) can be derived to obtain dehazed image J ( x ) :
J ( x ) = I ( x ) A ( x ) ( 1 t ( x ) ) t ( x )
As the distance between the object and the camera increases, the atmosphere between the camera and the object becomes thicker. In other words, when the scattering coefficient is a constant, the further the distance, the worse the haze becomes. Through this relationship, Equation (3) can be obtained.
t ( x ) = e β d ( x )
where β is the scattering coefficient, which is a constant indicating the level at which light is scattered due to fine particles. d ( x ) is the depth image, x is the distance between the target and the observer to the pixel, and t ( x ) is the transmission image.

2.3. Dark Channel Prior

According to He et al. [26], for pixels without haze, most of the three RGB channels tend to have low values for at least one channel. For channels exhibiting this tendency, the author defines it as the dark channel, hence, such prior using them referred to as the dark channel prior. The following Equation (4) defines the dark channel for image J ( x ) :
J d a r k ( x ) = min c { r , g , b } ( min y Ω ( x ) ( J c ( y ) ) )
where J d a r k is the dark channel of image and J c ( y ) is the color channel of the pixel x of image. Ω ( x ) is a set of pixels within a specific range centered on pixel x .
The equation shows that the value of the dark channel, which corresponds to a specific pixel x of the image, is the smallest value of the pixels around x . In He et al. [26], the pixel which has a small dark channel value is primarily one of the following three cases: (1) shadow area caused by object; (2) colorful object or surface; or (3) black or dark object or surface.
Dark channel images generally have small pixel values because natural images without haze are darkened by color or shadow [26]. However, if haze occurs, these objects will become blurry and invisible, resulting in a large dark pixel value and a white dark channel image. Based on these notions, we can identify the haze intensity of the image.

2.4. Guided Filter

The guided filter uses a guide image as an edge-preserving smoothing filter to perform smoothing without distorting key information of the entered image [38]. The filter assumes that the output images can be modeled linearly with guide images and linear coefficients.
q i = a k I i + b k , i ω k
q i = p i n i
where q i is the output image, I i is the guide image, and a k and b k are linear coefficients constant within ω k . Since the linear coefficient ( a k , b k ) in Equation (5) must be determined, Equation (5) is modeled as Equation (6) where p i is the input image and n i is the noise in the image. Then, we define the cost function to obtain linear coefficients through finding a solution that minimizes the cost function. The cost function is defined in Equation (7).
E ( a k , b k ) = i ω k ( ( a k I i + b k p i ) 2 + ε a k 2 )
where ε is the regularization parameter, which prevents a k from growing infinitely. The solution to minimize Equation (7) is Equations (8) and (9).
a k = 1 | ω | i ω k I i p i μ k p ¯ k σ k 2 + ε
b k = p ¯ k a k μ k
where σ k 2 , μ k is the variance and mean of I i within the ω k region. | ω | is the number of pixels in the region ω k . Lastly, p ¯ k is defined in Equation (10).
p ¯ k = 1 | ω | i ω k p i
After obtaining linear coefficients via Equations (8) and (9), the output image q i can be calculated. In this equation, the size of the region ω k and ε affect edge-preserving and smoothing the output image.

3. Image Dehazing Based on LiDAR Generated Grayscale Depth Prior

The structure of the proposed dehazing method is shown in Figure 3. First, hazy image and point cloud are used as input data. Through a relationship in Section 3.1, the point cloud is projected and converted into a depth image. When projecting the point cloud, the point cloud of the LiDAR must be projected within the camera frame through the calibration of the camera and LiDAR. Thereafter, the scattering coefficient is estimated through a relationship in Section 3.2 by using the image with haze as an input image. The dark channel image used in Section 3.2, obtained from the hazy image, is also used to estimate the atmospheric light. Finally, the transmission image is estimated through the depth image and the scattering coefficient, and after refining the transmission image by applying the guided filter in Section 3.3, the dehazing is performed according to Equation (13).

3.1. Point Cloud Projection

In this study, synthetic haze image generation and verification of dehazing algorithm are performed using a KITTI dataset [39]. In order to generate depth images required for the dehazing algorithm, the point cloud of the KITTI dataset was projected into an image [37]. Using the calibration data from the dataset, the projection, rotation, and translation matrices can be obtained, and the point cloud in 3D format projects into 2D through the relationship shown in Figure 4.
First, when a point in the 3D space is represented as [ X , Y , Z , 1 ] T , its position on the 2D image is expressed in [ x , y , 1 ] T where X , Y , and Z refer to the coordinates of a point cloud in the world frame, and x and y are image pixel coordinates in the camera frame. To perform a projection, points in the 3D space belonging to the world frame should be represented within the camera frame. This can be expressed by multiplying the world frame’s rotation with the matrix extrinsic matrix for translation.
Then, [ X , Y , Z , 1 ] T can be projected onto a two-dimensional plane by normalizing the obtained values and multiplying them by the intrinsic matrix containing focal length ( f x , f y ) and principle points ( c x , c y ) . Figure 5 is a 2D depth image obtained through point cloud projection.
The LiDARs used in the KITTI dataset are mechanical spinning LiDARs with 360-degree coverage. These LiDARs have a high point cloud density, but when the point cloud is matched for that image, they are sparse, as shown in Figure 5, and depth image using these sparse data is challenging to use. Therefore, by increasing the size of the projected point cloud, this sparsity should be lowered. Figure 6 is the depth image depending on the different sizes of point cloud.

3.2. Scattering Coefficient Estimation

To obtain a transmission image for dehazing from depth image, a scattering coefficient is required. However, it is not easy to obtain an accurate scattering coefficient with only the image obtained from camera. Therefore, using synthetic haze image and ground truth image, a model that can estimate the scattering coefficient should be obtained.
The synthetic haze images required for this were synthesized based on the atmospheric scattering model using the KITTI dataset [39] and depth images. The depth image used for the synthesis was generated by monodepth2 [40].
Figure 7 shows that haze is generated throughout the images. Thus, estimating the dark channel of hazy image and calculating average brightness is higher than when calculated in a no-haze situation. This can be confirmed in Figure 8. Such a relationship allows us to model equations that obtain the scattering coefficient from the dark channel’s average brightness. In this study, this average brightness is called the dark channel means (DCMs). To model the equation, an optimal scattering coefficient value for the haze image should be obtained. This can be obtained by performing dehazing of each value of the scattering coefficient, gradually increasing the scattering coefficient, and comparing the obtained results with the ground truth. Comparison of images is performed by calculating the mean square error (MSE) for pixels of each image, and when the mean square error becomes the smallest, the value at that time is set as the optimal scattering coefficient.
Algorithm 1 is pseudocode for estimating optimal scattering coefficients. The input data of the algorithm is ground truth image (GT), hazy image (hazy), and depth image (depth) (line 3). The algorithm initializes the scattering coefficient to 0 and 0.01 (line 1 and 2), incrementally increases them (line 16 and 17), and dehazing is performed using Equation (2) (line 10 and 11) and Equation (3) (line 7 and 8).
After dehazing, MSE is obtained through the dehazed image and GT (line 13 and 14). When MSE becomes the smallest (line 5), the scattering coefficient is determined as the optimal scattering coefficient (line 20).
Using the method in Algorithm 1, the optimal scattering coefficient for each hazy image is estimated. Next, we obtain the DCM of each hazy image and create a distribution chart using the DCM and the optimal scattering coefficient. The following Figure 9 refers to a scattering coefficient—DCM distribution chart obtained by the synthetic KITTI haze dataset. In Figure 9, the x-axis represents the DCM, and the y-axis represents the scattering coefficient.
A total of 100 synthetic haze images were used to obtain the scattering coefficient model. This is the result of synthesizing 20 types of images in 5 stages depending on th level of haze generation. Figure 10 shows synthetic haze images with varying scattering coefficients of step 5.
Using the distribution of the DCM-optimal scattering coefficient for 100 hazy synthetic images, the relationship between the two variables can be derived.
β = 0.0174 D C M 0.5919
Equation (11) is a model obtained by linear regression of the DCM-optimal scattering coefficient distribution for 100 synthetic images.
Algorithm 1 Estimate β
1:  β 1 = 0
2:  β 2 = 0.01
3:  Input: GT, hazy, depth
4:  For  MSE 1 < MSE 2  do
5:           trans 1 = transmission ( β 1 , depth ) e q . ( 3 )
6:           trans 2 = transmission ( β 2 , depth ) e q . ( 3 )
7:          dehaze 1 = dehazing ( hazy , trans 1 ) e q . ( 2 )
8:           dehaze 2 = dehazing ( hazy , trans 2 ) e q . ( 2 )
9:           MSE 1 = MSE ( dehaze 1 , GT )
10:         MSE 2 = MSE ( dehaze 2 , GT )
11:          β 1 = β 1 + 0.01
12:          β 2 = β 2 + 0.01
13: end for
14: return β 1 0.01

3.3. Transmission Image Refine

The transmission image of the atmospheric scattering model can be obtained by Equation (3). The raw transmission image is estimated using the depth image and the scattering coefficient, obtained by point cloud 2D projection and Equation (11), respectively.
The raw transmission image is estimated as shown in Figure 11c. Since the transmission image is generated from the depth image via point cloud 2D projection, the raw transmission image shows the block effect in He et al. [26]. Therefore, the raw transmission image obtained through Equation (3) should be refined. To refine the transmission image, the hazy image and raw transmission image are used, and the guided filtering is performed.

3.4. Background Parameter

Background parameters are used to prevent dehazing performance degradation due to differences in detection range between camera and LiDAR. If an object can be seen from the camera, but is outside the detection range of LiDAR, the pixel value of the transmission image is 1 for the absence of point cloud, so dehazing is not effective. Therefore, for places where point cloud does not exist, it should be set to a value between the maximum pixel value that point cloud can have and the original maximum pixel value of 255. The compensation process for an empty space in which the point cloud does not exist is performed through a background parameter. Figure 12 is the result of the transmission image after applying the background parameter.
The background parameter was set to 195 because it was the most effective after several times of dehazing through real-world haze photographs. It is impossible to set the parameter through quantitative analysis because there is no ground truth image for the actual haze occurrence image. Therefore, the background parameter was set by a heuristic approach. Figure 13 is the result of dehazing through several background parameters.

3.5. Scene Radiance Recovery

3.5.1. Estimation of Atmospheric Light

Generally, the appropriate value for atmospheric light in the hazy image would be the strongest pixel value within the image. In this case, however, it has the disadvantage of not being able to distinguish white objects. To compensate for these shortcomings, the dark channel prior is used. After obtaining the dark channel prior from the hazy image, the top 0.1% of the brightest pixels are drawn from the dark channel. We can consider these pixels as the most hazy pixels. So, among these pixels, the brightest pixel in the input image I ( x ) , is selected as the atmospheric light A ( x ) .

3.5.2. Dehazing Process

The 2D depth image is obtained through projection of the point cloud, and the scattering coefficient is obtained through the DCM—scattering coefficient equation. Transmission images can then be obtained through the acquired depth image and the scattering coefficient. Thus, scene radiance recovery can be performed through the atmospheric scattering model.
J ( x ) = I ( x ) A ( x ) t ( x ) + A ( x )
To avoid noise generation due to the transmission image, it is necessary to set the lower bound of the transmission image. The equation in which the lower bound is added can be expressed as follows.
J ( x ) = I ( x ) A ( x ) max { t ( x ) , 0.1 } + A ( x )

4. Simulation

The dehazing algorithm was written in Python, and simulations performed on the Intel [email protected], 8GB RAM. In simulation, the improvement of the image was determined by comparing the mean square error (MSE), peak signal-to-noise ratio (PSNR), image enhancement factor (IEF), and structural similarity index measure (SSIM) [41] of the hazy image and dehazed image. In addition, our proposed algorithm was quantitatively and qualitatively compared to existing algorithms, such as He et al. [26], Tan et al. [28], and Fattal et al. [29].

4.1. Quantitative Analysis of Dehazing Improvement Quality

Dehazing improvement performance of the proposed algorithm is quantitatively analyzed using performance improvement parameters. When the synthetic haze image is composed, β is set to 0.003 and atmospheric light is set to 210. Figure 14b is the synthetic haze image and Figure 14c is the result of dehazing.
Performance analysis of three pairs of hazy images and dehazed images was performed via PSNR, SSIM, and MSE. First, performance analysis for hazy image and ground truth image is shown in GT-Hazed in Table 1, Table 2 and Table 3 and performance analysis for ground truth and the dehazed image is shown in GT-Dehazed in Table 1, Table 2 and Table 3. The PSNR and SSIM of GT-Dehazed were higher than those of GT-Hazed, and the MSE of GT-Dehazed were lower than that of GT-Hazed. Therefore, the analysis results from the image analysis parameters show that all three images have been improved.

4.2. Quantitative Comparison of Different Dehazing Algorithm

We performed a quantitative performance analysis between the existing dehazing algorithm and the proposed algorithm. Existing algorithms used for comparison of performance are Tan et al. [28], Fattal et al. [29], and He et al. [26]. Analysis of the resulting images was conducted via MSE, PSNR, IEF, and SSIM as shown in Table 4, Table 5 and Table 6. Figure 15 shows an input image, synthetic haze image, and a dehazing image generated from the proposed and existing algorithms, respectively.
The proposed algorithm is designed to make the MSE smallest. Thus, if the DCM of the hazy image input does not deviate significantly from the model of DCM and scattering coefficient, the proposed method achieves the smallest MSE of the four methods. In addition, as MSE became smaller, other performance parameters were improved.

5. Experiments

Based on the proposed algorithm, we performed dehazing using the DENSE dataset [42] from Ulm University in case of natural fog.

5.1. Comparison of Dehazing Results

Ground truth of object detection is as shown in Figure 16. Figure 17 shows the results of the proposed and existing algorithms and Figure 18 shows the results of qualitative evaluation with YOLO V3 using dehazed results. We confirmed that dehazing was correctly performed under real haze and fog conditions through the dataset. In addition, through YOLO V3 object detection, we also confirmed that the image was improved after dehazing.

5.2. Time Consumption

We compared computing time for 1920 × 1024 pixel images. The algorithms were written in Python and performed on the Intel i5-3470 CPU @ 3.2Ghz, 8GB RAM. The proposed algorithm took approximately 0.47 s per image. He’s algorithm took about 0.4 s, Tan’s algorithm took about 0.2 s, and Fattal’s algorithm took about 75 s. Table 7 represents the progressing time and frame per second (FPS).

6. Conclusions

We present a method for performing dehazing via LiDAR depth image and DCM-scattering coefficient model. The proposed algorithm obtains the scattering coefficient model through the DCM and scattering coefficient relationship. Dehazing is then performed through the scattering coefficient and point cloud projection depth image obtained from LiDAR. Through simulations, we confirmed that the dehazed image is obtained effectively. In the simulation, MSE showed improvement over conventional algorithms, and PSNR and IEF, which are dependent on MSE, have also shown improvements. Furthermore, SSIM, an important parameter used in image recognition, showed an average improvement of about 24% over conventional algorithms.
However, the proposed algorithm has a problem to solve. First, when estimated using DCM, the scattering coefficient was able to perform dehazing effectively on most haze images, but using only pixel value mean may be unreliable. If there are many colorful objects in the near distance, the DCM can still be low, even with much haze. Consequently, it will deviate from the scattering coefficient estimation model, which results in dehazing being ineffective. Such problems of DCM could be addressed by CNN and by supervised learning for image and effective scatter coefficient.
In addition, because the depth image is obtained through LiDAR, dehazing may not work effectively if LiDAR malfunctions. We will improve these existing problems through further research. Moreover, there is a real-time problem. As of now, there are difficulties in operating in real time. This occurs because the imaging operation is performed simply with CPU only. Therefore, it is planned to secure real time by making it possible to operate in parallel through GPU operations through future research.

Author Contributions

Conceptualization, W.Y.C., S.Y.K. and C.H.K.; methodology, W.Y.C. and C.H.K.; validation, W.Y.C. and S.Y.K.; formal analysis, W.Y.C. and C.H.K.; investigation, W.Y.C.; writing—original draft preparation, W.Y.C.; writing—review and editing, S.Y.K.; visualization, W.Y.C.; supervision, S.Y.K. and C.H.K.; funding acquisition, S.Y.K. and C.H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

This research was supported by Unmanned Vehicles Core Technology Research and Development Program through the National Research Foundation of Korea(NRF), Unmanned Vehicle Advanced Research Center(UVARC) funded by the Ministry of Science and ICT, the Republic of Korea (No. 2020M3C1C1A01086408).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Taketomi, T.; Uchiyama, H.; Ikeda, S. Visual SLAM algorithms: A survey from 2010 to 2016. IPSJ Trans. Comput. Vis. Appl. 2017, 9, 16. [Google Scholar] [CrossRef]
  2. Shao, L.; Liu, L.; Li, X. Feature learning for image classification via multiobjective genetic programming. IEEE Trans. Neural Netw. Learn. Syst. 2013, 25, 1359–1371. [Google Scholar] [CrossRef]
  3. Zhu, F.; Shao, L. Weakly-supervised cross-domain dictionary learning for visual recognition. Int. J. Comput. Vis. 2014, 109, 42–59. [Google Scholar] [CrossRef]
  4. Luo, Y.; Liu, T.; Tao, D.; Xu, C. Decomposition-based transfer distance metric learning for image classification. IEEE Trans. Image Process. 2014, 23, 3789–3801. [Google Scholar] [CrossRef] [PubMed]
  5. Tao, D.; Li, X.; Wu, X.; Maybank, S.J. Geometric mean for subspace selection. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 31, 260–274. [Google Scholar]
  6. Liu, L.; Shao, L. Learning discriminative representations from RGB-D video data. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence, Beijing, China, 3–9 August 2013. [Google Scholar]
  7. Tao, D.; Li, X.; Wu, X.; Maybank, S.J. General tensor discriminant analysis and gabor features for gait recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1700–1715. [Google Scholar] [CrossRef]
  8. Zhang, Z.; Tao, D. Slow feature analysis for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 436–450. [Google Scholar] [CrossRef] [Green Version]
  9. Wu, F.; Duan, J.; Chen, S.; Ye, Y.; Ai, P.; Yang, Z. Multi-target recognition of bananas and automatic positioning for the inflorescence axis cutting point. Front. Plant Sci. 2021, 12, 705021. [Google Scholar] [CrossRef]
  10. Gong, L.; Fan, S. A CNN-Based Method for Counting Grains within a Panicle. Machines 2022, 10, 30. [Google Scholar] [CrossRef]
  11. Forster, C.; Pizzoli, M.; Scaramuzza, D. SVO: Fast semi-direct monocular visual odometry. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 15–22. [Google Scholar]
  12. Engel, J.; Koltun, V.; Cremers, D. Direct sparse odometry. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 611–625. [Google Scholar] [CrossRef]
  13. Han, J.; Zhang, D.; Cheng, G.; Guo, L.; Ren, J. Object detection in optical remote sensing images based on weakly supervised learning and high-level feature learning. IEEE Trans. Geosci. Remote Sens. 2014, 53, 3325–3337. [Google Scholar] [CrossRef] [Green Version]
  14. Cheng, G.; Han, J.; Guo, L.; Qian, X.; Zhou, P.; Yao, X.; Hu, X. Object detection in remote sensing imagery using a discriminatively trained mixture model. ISPRS J. Photogramm. Remote Sens. 2013, 85, 32–43. [Google Scholar] [CrossRef]
  15. Han, J.; Zhou, P.; Zhang, D.; Cheng, G.; Guo, L.; Liu, Z.; Bu, S.; Wu, J. Efficient, simultaneous detection of multi-class geospatial targets based on visual saliency modeling and discriminative learning of sparse coding. ISPRS J. Photogramm. Remote Sens. 2014, 89, 37–48. [Google Scholar] [CrossRef]
  16. Mur-Artal, R.; Tardós, J.D. Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef] [Green Version]
  17. Qin, T.; Li, P.; Shen, S. Vins-mono: A robust and versatile monocular visual-inertial state estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef] [Green Version]
  18. Zhang, J.; Singh, S. LOAM: Lidar Odometry and Mapping in Real-time. In Proceedings of the Robotics: Science and Systems, Online Conference, 13–16 July 2014. [Google Scholar]
  19. Shan, T.; Englot, B. Lego-loam: Lightweight and ground-optimized lidar odometry and mapping on variable terrain. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4758–4765. [Google Scholar]
  20. Zhang, J.; Singh, S. Visual-lidar odometry and mapping: Low-drift, robust, and fast. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 2174–2181. [Google Scholar]
  21. Agarwal, A.; Maturana, D.; Scherer, S. Visual Odometry in Smoke Occluded Environments; Robotics Institute, Carnegie Mellon University: Pittsburgh, PA, USA, 2014. [Google Scholar]
  22. Narasimhan, S.G.; Nayar, S.K. Vision and the atmosphere. Int. J. Comput. Vis. 2002, 48, 233–254. [Google Scholar] [CrossRef]
  23. Nayar, S.K.; Narasimhan, S.G. Vision in bad weather. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; pp. 820–827. [Google Scholar]
  24. Narasimhan, S.G.; Nayar, S.K. Removing weather effects from monochrome images. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2001), Kauai, HI, USA, 8–14 December 2001; pp. II-186–II-193. [Google Scholar]
  25. Zhu, Q.; Mai, J.; Shao, L. A fast single image haze removal algorithm using color attenuation prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar] [PubMed] [Green Version]
  26. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar]
  27. Zhao, C.; Sun, Q.; Zhang, C.; Tang, Y.; Qian, F. Monocular depth estimation based on deep learning: An overview. Sci. China Technol. Sci. 2020, 63, 1612–1627. [Google Scholar] [CrossRef]
  28. Tan, R.T. Visibility in bad weather from a single image. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  29. Fattal, R. Dehazing using color-lines. ACM Trans. Graph. (TOG) 2014, 34, 13. [Google Scholar] [CrossRef]
  30. Huo, F.; Zhu, X.; Zeng, H.; Liu, Q.; Qiu, J. Fast Fusion-Based Dehazing with Histogram Modification and Improved Atmospheric Illumination Prior. IEEE Sens. J. 2020, 21, 5259–5270. [Google Scholar] [CrossRef]
  31. Zheng, M.; Qi, G.; Zhu, Z.; Li, Y.; Wei, H.; Liu, Y. Image dehazing by an artificial image fusion method based on adaptive structure decomposition. IEEE Sens. J. 2020, 20, 8062–8072. [Google Scholar] [CrossRef]
  32. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An end-to-end system for single image haze removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Ren, W.; Liu, S.; Zhang, H.; Pan, J.; Cao, X.; Yang, M.-H. Single image dehazing via multi-scale convolutional neural networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 154–169. [Google Scholar]
  34. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. Aod-net: All-in-one dehazing network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4770–4778. [Google Scholar]
  35. Li, R.; Pan, J.; Li, Z.; Tang, J. Single image dehazing via conditional generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8202–8211. [Google Scholar]
  36. Li, R.; Pan, J.; He, M.; Li, Z.; Tang, J. Task-oriented network for image dehazing. IEEE Trans. Image Process. 2020, 29, 6523–6534. [Google Scholar] [CrossRef]
  37. Zhao, S.; Zhang, L.; Shen, Y.; Zhou, Y. RefineDNet: A weakly supervised refinement framework for single image dehazing. IEEE Trans. Image Process. 2021, 30, 3391–3404. [Google Scholar] [CrossRef]
  38. He, K.; Sun, J.; Tang, X. Guided image filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1397–1409. [Google Scholar] [CrossRef]
  39. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef] [Green Version]
  40. Godard, C.; Mac Aodha, O.; Firman, M.; Brostow, G.J. Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 3828–3838. [Google Scholar]
  41. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  42. Bijelic, M.; Gruber, T.; Mannan, F.; Kraus, F.; Ritter, W.; Dietmayer, K.; Heide, F. Seeing through fog without seeing fog: Deep multimodal sensor fusion in unseen adverse weather. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11682–11692. [Google Scholar]
Figure 1. Clear image and hazy image: (a) clear image; (b) hazy image.
Figure 1. Clear image and hazy image: (a) clear image; (b) hazy image.
Sensors 22 01199 g001
Figure 2. Atmospheric scattering model.
Figure 2. Atmospheric scattering model.
Sensors 22 01199 g002
Figure 3. Flowchart of the proposed method.
Figure 3. Flowchart of the proposed method.
Sensors 22 01199 g003
Figure 4. Point cloud 2D projection.
Figure 4. Point cloud 2D projection.
Sensors 22 01199 g004
Figure 5. Result of point cloud projection: (a) projected 2D point cloud; (b) 3D point cloud.
Figure 5. Result of point cloud projection: (a) projected 2D point cloud; (b) 3D point cloud.
Sensors 22 01199 g005
Figure 6. Generated grayscale depth image with different size of point cloud: (a) size = 1; (b) size = 3; (c) size = 5; (d) size = 7.
Figure 6. Generated grayscale depth image with different size of point cloud: (a) size = 1; (b) size = 3; (c) size = 5; (d) size = 7.
Sensors 22 01199 g006
Figure 7. Synthetic haze image with ground truth: (a) ground truth; (b) synthetic haze image.
Figure 7. Synthetic haze image with ground truth: (a) ground truth; (b) synthetic haze image.
Sensors 22 01199 g007
Figure 8. Generated dark channel depending on different intensity of haze: (a) dark channel; (b) hazy image.
Figure 8. Generated dark channel depending on different intensity of haze: (a) dark channel; (b) hazy image.
Sensors 22 01199 g008
Figure 9. Distribution of DCM-optimal scattering coefficient.
Figure 9. Distribution of DCM-optimal scattering coefficient.
Sensors 22 01199 g009
Figure 10. Synthetic image generated from difference of scattering coefficient: (a) β = 0; (b) β = 0.001; (c) β = 0.0015; (d) β = 0.002; (e) β = 0.0025; (f) β = 0.003.
Figure 10. Synthetic image generated from difference of scattering coefficient: (a) β = 0; (b) β = 0.001; (c) β = 0.0015; (d) β = 0.002; (e) β = 0.0025; (f) β = 0.003.
Sensors 22 01199 g010
Figure 11. Refine transmission image: (a) hazy image; (b) grayscale depth image generated from point cloud projection; (c) raw transmission image from grayscale depth image; (d) refined transmission image.
Figure 11. Refine transmission image: (a) hazy image; (b) grayscale depth image generated from point cloud projection; (c) raw transmission image from grayscale depth image; (d) refined transmission image.
Sensors 22 01199 g011
Figure 12. Comparison applying background parameter: (a) before applying; (b) transmission image generated from (a); (c) applying background parameter; (d) transmission image generated from (c).
Figure 12. Comparison applying background parameter: (a) before applying; (b) transmission image generated from (a); (c) applying background parameter; (d) transmission image generated from (c).
Sensors 22 01199 g012
Figure 13. Dehazed results at various background parameters: (a) hazy input; (b) 120; (c) 195; (d) 255.
Figure 13. Dehazed results at various background parameters: (a) hazy input; (b) 120; (c) 195; (d) 255.
Sensors 22 01199 g013
Figure 14. Hazy image with ground truth and dehazed image: (a) ground truth; (b) hazy image (c) dehazed image.
Figure 14. Hazy image with ground truth and dehazed image: (a) ground truth; (b) hazy image (c) dehazed image.
Sensors 22 01199 g014
Figure 15. Dehazing result of various algorithms: (a) hazy image; (b) ground truth; (c) Fattal’s method; (d) He’s method; (e) Tan’s method; (f) proposed method.
Figure 15. Dehazing result of various algorithms: (a) hazy image; (b) ground truth; (c) Fattal’s method; (d) He’s method; (e) Tan’s method; (f) proposed method.
Sensors 22 01199 g015aSensors 22 01199 g015b
Figure 16. Ground truth of object detection.
Figure 16. Ground truth of object detection.
Sensors 22 01199 g016
Figure 17. Results of various algorithms: (a) hazy input; (b) proposed; (c) Fattal; (d) He; (e) Tan.
Figure 17. Results of various algorithms: (a) hazy input; (b) proposed; (c) Fattal; (d) He; (e) Tan.
Sensors 22 01199 g017
Figure 18. Qualitative comparison of various algorithms: (a) hazy input; (b) proposed; (c) Fattal; (d) He; (e) Tan.
Figure 18. Qualitative comparison of various algorithms: (a) hazy input; (b) proposed; (c) Fattal; (d) He; (e) Tan.
Sensors 22 01199 g018
Table 1. Quantitative analysis of improvement using first row image from Figure 14.
Table 1. Quantitative analysis of improvement using first row image from Figure 14.
GT-HazedGT-DehazedImprovement (%)
PSNR (dB)14.6824.4466.43
SSIM0.83490.965815.68
MSE0.03410.0036−89.41
Table 2. Quantitative analysis of improvement using second row image from Figure 14.
Table 2. Quantitative analysis of improvement using second row image from Figure 14.
GT-HazedGT-DehazedImprovement (%)
PSNR (dB)13.1425.1991.63
SSIM0.73730.952525.48
MSE0.04850.0031−93.81
Table 3. Quantitative analysis of improvement using third row image from Figure 14.
Table 3. Quantitative analysis of improvement using third row image from Figure 14.
GT-HazedGT-DehazedImprovement (%)
PSNR (dB)15.4123.86654.92
SSIM0.80660.85025.410
MSE0.02880.0041−85.76
Table 4. Comparison result 1 using upper image from Figure 15.
Table 4. Comparison result 1 using upper image from Figure 15.
PSNR (dB)IEFSSIMMSE
Fattal et al.16.561.5420.80330.0221
He et al.17.601.9550.87110.0174
Tan et al.14.000.85430.60790.0398
Proposed24.449.4510.96580.0036
Table 5. Comparison result 2 using middle image from Figure 15.
Table 5. Comparison result 2 using middle image from Figure 15.
PSNR (dB)IEFSSIMMSE
Fattal et al.15.861.8670.71130.0260
He et al.17.142.5080.90730.0193
Tan et al.14.001.2190.66010.0398
Proposed25.1916.010.92520.0030
Table 6. Comparison result 3 using lower image from Figure 15.
Table 6. Comparison result 3 using lower image from Figure 15.
PSNR (dB)IEFSSIMMSE
Fattal et al.17.051.4610.68030.0197
He et al.18.191.9000.88280.0152
Tan et al.15.100.93260.63140.0309
Proposed23.877.0150.85020.0041
Table 7. Time consumption of each algorithms.
Table 7. Time consumption of each algorithms.
ProposedHe *TanFattal
Time (s)0.470.40.275
FPS2.12.550.013
* Guided filter used (not soft matting).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chung, W.Y.; Kim, S.Y.; Kang, C.H. Image Dehazing Using LiDAR Generated Grayscale Depth Prior. Sensors 2022, 22, 1199. https://doi.org/10.3390/s22031199

AMA Style

Chung WY, Kim SY, Kang CH. Image Dehazing Using LiDAR Generated Grayscale Depth Prior. Sensors. 2022; 22(3):1199. https://doi.org/10.3390/s22031199

Chicago/Turabian Style

Chung, Won Young, Sun Young Kim, and Chang Ho Kang. 2022. "Image Dehazing Using LiDAR Generated Grayscale Depth Prior" Sensors 22, no. 3: 1199. https://doi.org/10.3390/s22031199

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop