Next Article in Journal
Structural Optimization Design of Dual Robot Gripper Unloading Device Based on Intelligent Optimization Algorithms and Generative Design
Next Article in Special Issue
Embedded Processing for Extended Depth of Field Imaging Systems: From Infinite Impulse Response Wiener Filter to Learned Deconvolution
Previous Article in Journal
Drivers’ Comprehensive Emotion Recognition Based on HAM
Previous Article in Special Issue
Object Detection Based on Roadside LiDAR for Cooperative Driving Automation: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DRGAN: Dense Residual Generative Adversarial Network for Image Enhancement in an Underwater Autonomous Driving Device

1
College of Information Engineering, Taizhou University, Taizhou 225300, China
2
School of Automation and Electrical Engineering, Shenyang Ligong University, Shenyang 110159, China
3
School of Computer Science and Engineering, Changshu Institute of Technology, Changshu 215506, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(19), 8297; https://doi.org/10.3390/s23198297
Submission received: 25 August 2023 / Revised: 28 September 2023 / Accepted: 4 October 2023 / Published: 7 October 2023
(This article belongs to the Special Issue Advances in Sensing, Imaging and Computing for Autonomous Driving)

Abstract

:
Underwater autonomous driving devices, such as autonomous underwater vehicles (AUVs), rely on visual sensors, but visual images tend to produce color aberrations and a high turbidity due to the scattering and absorption of underwater light. To address these issues, we propose the Dense Residual Generative Adversarial Network (DRGAN) for underwater image enhancement. Firstly, we adopt a multi-scale feature extraction module to obtain a range of information and increase the receptive field. Secondly, a dense residual block is proposed, to realize the interaction of image features and ensure stable connections in the feature information. Multiple dense residual modules are connected from beginning to end to form a cyclic dense residual network, producing a clear image. Finally, the stability of the network is improved via adjustment to the training with multiple loss functions. Experiments were conducted using the RUIE and Underwater ImageNet datasets. The experimental results show that our proposed DRGAN can remove high turbidity from underwater images and achieve color equalization better than other methods.

1. Introduction

With developments in science and technology, Internet of Things (IoT) technology has been introduced into advanced underwater vision tasks, such as autonomous underwater driving, ocean scene analysis, and fisheries. Underwater data, such as the marine environment, marine density, and seafloor pathways, are monitored to understand the state of the underwater environment and the growth and health of underwater life through the intelligent analysis of real-time photographs taken underwater. However, the accuracy of the intelligent analysis results is greatly influenced by the underwater images’ quality; a complex imaging environment results in color casts and a loss of detail in the images obtained [1,2]. As a result, it is critical to achieve clarity and improve the details in underwater images.
Underwater image enhancement is developing rapidly via both traditional and deep learning. In traditional methods, Drews et al. [3], influenced by the dark channel prior (DCP [4]), proposed transmission estimation in underwater single images (UDCP), which does not take into account the effects of red channels but is prone to overexpression. Ma et al. [5] proposed the restoration of underwater images using a mix of improved dark channel prior and gray world methods; this new model improved the DCP and gray world theory to restore underwater images. Ancuti et al. [6] used multi-scale fusion to generate an image that was clear after white balance and gamma correction had been performed on the damaged image. To improve underwater image quality, Liang et al. [7] combined color correction based on attenuation maps with a detail retention and haze removal method based on multi-scale decomposition. Marques et al. [8] derived an effective atmospheric illumination model from local contrast information based on human observations and, from this model, they generated an enhanced image for highlighting details and an enhanced image for removing darkness. In turn, the underwater image is enhanced via multi-scale fusion. The interpretability of traditional methods is obvious, but the effect needs to be further improved.
In recent years, the application of deep learning methods in underwater image processing has become more and more prominent, especially deep learning methods based on genetic algorithms. For example, Li et al. [9] proposed Water-GAN to enhance underwater images. Synthetic underwater images are utilized as datasets for training the neural network to perform underwater image color correction. Fabbri et al. [10] suggested the enhancement of underwater imagery using generative adversarial networks. They first applied CycleGAN to paired images to create degraded underwater images. The underwater image pairs were then selected as datasets for further network training. Guo et al. [11] designed a multi-level intensive generative adversarial network, containing two multi-scale dense blocks that can correct color differences and enhance image details. Islam et al. [12] suggested fast underwater image enhancement to enhance the perception of images (FUnIE-GAN) based on U-Net, which improves image detail clarity by using residual connections in the generator. GAN-RS, a multi-branch discriminator proposed by Chen et al. [13], was developed to increase the quality of underwater images. However, numerous training parameters require careful tuning. If you train with incorrect parameters, the resulting images will produce artifacts. Huang et al. [14] proposed Semi-UIR, to enhance the model performance with a semi-supervised method and mean-teacher-based underwater image restoration model, by constructing a reliable bank and contrast learning. Compared with the traditional learning methods, a deep learning method can better solve the image color distortion problem and has a superior portability and learning ability in image processing.
The above methods focus on enhancing underwater images, as shown in Figure 1. However, their algorithm is not well-suited to intricate scenarios due to the lack of attention given to the color and data loss caused by the imaging environment. In addition, most methods that are available improve the network by increasing the network depth; however, this will result in problems such as gradients, training difficulties, and unstable parameters [15].
To solve the problems above, we propose the implementation of the Dense Residual Generative Adversarial Network (DRGAN). Here are the primary contributions:
(1)
A multi-scale feature extraction module is proposed to extract image detail information and expand the receptive field.
(2)
A dense residual block is proposed to fuse feature maps into clear images, not only fully utilizing all layers with local dense connections but also adding residual connections to reuse information.
(3)
We combine multiple loss functions to facilitate the learning of the generator regarding the generation of clear images. The experimental results show that DRGAN outperforms the state-of-the-art methods in terms of qualitative and quantitative indicators.
The remainder of this work is structured as follows. Section 2 discusses related work, such as dense residual theory and GANs. Section 3 describes our proposed method in detail, and Section 4 presents and discusses the experimental results and analysis. Finally, Section 5 concludes the paper.

2. Related Work

2.1. Generative Adversarial Network

Generative adversarial networks are composed of two distinct neural networks: a generator and a discriminator [16]. In this paper, we employ a generator to produce a distinct image from the deteriorated image; the discriminator utilizes both the clear image and the generated image as the input, and it outputs the probability that the generated image is true. The generator and discriminator engage in an adversarial relationship during training to encourage the discriminator to accurately distinguish between genuine and counterfeit samples. In the end, we want the generator to produce images of high quality as a result of the network.

2.2. Residual Network

He et al. [17] suggested a residual network as a solution to the issue of numerous parameters decreasing due to the network’s excessive depth.
As shown in Figure 2, the addition of the details of the shallow layers to the subsequent deep layers allows the deep layers to focus on learning, avoids the loss of feature information, and prevents model degradation. Consequently, utilizing this technique on deep networks can address issues such as the escalation of gradients during training.

2.3. Densely Connected Convolutional Network

The distinction between densely connected convolutional networks (DenseNets) [18] and residual networks lies in the fact that DenseNets facilitate the transmission of data between the various layers of the network and boost the number of links in each layer of the network, leading to improved feature reuse and a more powerful gradient propagation. As illustrated in Figure 3, the feature output of the preceding layer can be sent to all the following layers, and the transmission of feature information is improved via the linking of the layers in the network in pairs. Due to each layer being connected to all the previous layers when the gradient is back-propagated, the gradient is transferred to all the preceding layers in turn, a small number of convolutional kernels can still produce a substantial amount of feature information, and the preceding layers can fine-tune their parameters by taking advantage of the gradient data from the subsequent layers. Through this process, the issue of gradient disappearance is lessened, and the network’s training performance is enhanced.
The dense connections in Figure 3 can be represented as:
X n = H n X 1 , X 2 , , X n ,
where [ X 0 , X 1 , , X n 1 ] is the output of the characteristic map.

3. Our Method

3.1. Generative Network

Our generator includes a multi-scale feature extraction module (MSFEM) and dense residual block (DRB), facilitating the generation of crystal-clear underwater images. Using the yellow convolution unit in the MSFEM as a case in point, as demonstrated in Figure 4, Conv1 is a convolution with a stride of 1, the number of convolution kernels is 16, the ReLU is an activation function, and the BN is a representation of batch normalization.
As shown in Figure 5, 1 × 1 and 3 × 3 focus on extracting image detail information, the 5 × 5 and 7 × 7 convolution kernels can better extract the global features of the image, and each parallel convolution unit consists of two identical convolution kernels. By combining the above, MSFEM can extract different spatial features from the input image. We use concatenation to combine four feature maps to realize feature information fusion. To a certain extent, the loss of shallow details is avoided. The output result can be expressed as:
F = Concat ( f 1 , f 2 , f 3 , f 4 ) ,
where F is the feature image processed by the module, and f 1 , f 2 , f 3 , and f 4 are the feature images obtained by the four convolution units, respectively.
As illustrated in Figure 4, we designed a DRB composed of 3 × 3 convolution kernels. The goal of dense connections is to process as much information as possible from all layers, and the use of residual connections not only improves the utilization of information and ensures image integrity but also enables the next dense residual block to preprocess the information.
By combining residual and dense connections, the DRB ensures the correct transmission of feature information and reduces the computational complexity of the module. More importantly, compared with other networks, the network efficiency is improved without additional logistical costs, making feature information fusion more efficient. The result of the residual can be expressed as:
O u t p u t = h ( x 1 ) + F ( x 1 , W l ) ,
where O u t p u t is the output image of the module, h ( x 1 ) is the direct mapping, F ( x 1 , W l ) is the residual part, x 1 is the input, and W l is the convolution operation.
Derived from the SSIM test results shown in Figure 6, the number of dense residual block cycles of the network is set to 7 to achieve the most desired outcome.

3.2. Discrimination Network

The proposed discriminating network, as illustrated in Figure 7, is made up of five-layer convolution units, like the design of PatchGAN [19]. The convolution unit follows the Conv-BN-Leaky ReLU structure, and the step size of multiple units is set to 2 to increase the receptive field of the output characteristics. Due to the problem of neurons not being able to learn after the ReLU function enters the negative interval, we chose the leaky ReLU function in order to limit the appearance of silent neurons. The network uses the generated image and undistorted image as the inputs, and it outputs an image with a size of 30 × 30 . The discrimination network operates on small-sized image blocks, which greatly reduces the number of parameters and amount of computation, while also alleviating the problem of slow convergence that is characteristic of the GAN. The leaky ReLU expression is shown in the formula:
L e a k y R e L U = x , x > 0 α x , x 0 ,
where α is a tiny constant used to maintain some negative axis values so that the information on the negative axis is not completely lost.
In the last layer, the sigmoid function is used to map the output pixel range to the undistorted image, which can clearly distinguish the authenticity of the created image and the undistorted image in a certain area. The function expression is as follows:
f ( x ) = e x ( 1 + e x ) 2 ,

3.3. Loss Function

The adjustment of the network is facilitated by the linear combination of GAN loss and SSIM loss, as outlined below.
(1)
GAN Loss
GAN loss functions are used to make the generated sample distribution as close to the true sample distribution as possible. The following is the definition of countermeasure loss:
L GAN ( G , D ) = E X , Y [ log D ( Y ) ] + E X , Y [ log ( 1 D ( X , G ( X , Z ) ) ) ] ,
where X is the degraded image, Y is the undistorted image, E is the mathematical expectation, and Z denotes the random noise. To ensure that D recognizes the image produced by G as an undistorted image, G generates an image that conforms to the undistorted data distribution as much as possible.
(2)
SSIM Loss
The structural similarity of the two images is measured using SSIM loss. SSIM loss functions similarly to the human visual system. It is sensitive to the perception of local structural changes and is conducive to enhancing the image’s texture details. SSIM loss is defined as:
L SSIM ( P ) = 1 SSIM p ˜ ,
where P is the image block, and p ˜ is the image block’s center pixel.
(3)
Perceptual Loss
The parameters of the feature map in the trained convolutional neural network define the perceptual loss. The image details obtained after the function participates in the training are more realistic. The perceptual loss is defined as:
L Z = 1 W i , j H i , j x = 1 W i , j y = 1 H i , j ϕ i , j I Q x , y ϕ i , j G I S x , y ,
where ϕ i , j is the characteristic diagram of the output of the j convolution layer before the i pooling layer in the pre-training VGG19 network, and W i , j and H i , j are the dimensions of the characteristic diagram. In this paper, i is taken as 4, and j is taken as 3; the VGG 4 , 3 convolution characteristic diagram is selected to define the loss.
(4)
Overall Loss
The function of the overall loss obtained via the linear combination of the three loss functions can effectively improve the robustness of the network and is defined as:
L = λ 1 L GAN + λ 2 L SSIM + λ 3 L Z .
After many experiments, λ 1 is taken as 1, λ 2 as 100, and λ 3 as 10.

4. Experiment

To verify the effectiveness of DRGAN, in this study, we firstly set the experimental details. Then, we compared DRGAN with different representative methods. These methods included Fusion [6], ICCB [20], L^2UWE [8], FUnIE-GAN [12] (replaced with FUnIE below), Semi-UIR [14], and UWCNN [21]. Finally, to validate the components of DRGAN, we performed ablation experiments. Furthermore, we conducted experiments such as feature point matching and edge detection to validate the usefulness of our approach in real-world applications.

4.1. Experimental Details

We conducted experiments on the Underwater ImageNet [10] dataset and RUIE [22] dataset, respectively. The details are as follows. (1) From the Underwater ImageNet dataset, we randomly selected 4000 pairs of images from underwater scenes for training and 2000 pairs for testing. (2) We exploited the trained model of the Underwater ImageNet dataset to test the RUIE dataset, which demonstrated the generalization ability of DRGAN. We trained DRGAN with Adam and set the training and test image size to 256 × 256 × 3, the batch size to 2, and the epoch to 50. TensorFlow was used as the deep learning framework on an Ubuntu 18.04 machine with 32 GB RAM and a GTX1070Ti (8 GB).

4.2. Subjective Evaluation

The color of the undistorted swatch image would be degraded because of the complex underwater imaging environment. Therefore, the color restoration impact of DRGAN could be efficiently tested through color recovery experiments on the color card [23].
As can be seen in Figure 8, the Fusion method reduces the contrast between the yellow and pink color blocks, while deepening the overall hue of the color card picture, and the image processed via the ICCB algorithm suffers from a color distortion problem. Although the Semi-UIR algorithm can achieve color correction, the visual effect is negatively affected by the overall redness of the processed image. The problem of low discrimination is shown in the image that was processed via the L^2UWE algorithm, as evidenced by the dark purple and green color cards that are visually close to black. Overall, the color cards processed via the UWCNN algorithm suffer from poor color correction, as shown through the blueish hue. Overall, the FUnIE algorithm tends to make the color cards appear red during the experiment. On the contrary, our method achieves promising visual results with the color card images, especially when dealing with indistinguishable color patches (specifically black, purple, and dark green), validating the superiority of the color correction capability of our method.
Next, the method was applied to images from a complex underwater environment. The input image was affected by different degrees of color distortion, low brightness, and turbidity, resulting in various degradation phenomena. Figure 9 illustrates the processing results for each method. Images 1–2 are the normal degraded images, Images 3–4 are the atomized images, and Images 5–6 and Images 7–8 are green and blue partial images, respectively.
As can be seen in Figure 9, the Fusion algorithm fails to improve the sharpness and quality of low-brightness and color-distortion images. The ICCB algorithm has some success in improving the brightness and color correction, but the vividness of the image colors is greatly reduced. The L^2UWE algorithm fails to improve green, blue, and normal degraded images. Although the fogging problem can be mitigated, the generated image seems to have insufficient brightness. The FunIE algorithm can solve the problem of low brightness, but the problem of color distortion remains. The fogged image processed via FUnIE shows the problem of an obvious reddish tint, which is not consistent with the real image, and the image processed via the UWCNN algorithm cannot achieve a good visual effect due to the overall bluish color. Image processing using the Semi-UIR method achieves some success in defogging and color correction, but the overall brightness of the final image is low. In addition, as shown in Figure 5, Figure 6, Figure 7 and Figure 8, the ICCB method is unable to perform effective deblurring, as evidenced by the severe color distortion. On the contrary, the results of our method show brighter and clearer images compared to all the tested comparison algorithms. It was found that the algorithm can address degradation in complex underwater environments (off-color, low brightness, high turbidity, etc.) and that it exhibits a strong robustness. It was determined through subjective evaluation that our method produces better-clarity results for images with different degrees of degradation compared to other, similar new methods.

4.3. Objective Evaluation

The image quality when applying our method was further evaluated through five objective evaluation indexes: UCIQE, UIQM, SSIM, PSNR, and CIEDE2000.
(1)
The underwater color image quality evaluation index [24] (UCIQE) is proportional to the underwater picture quality, and the formula for calculating the index is as follows:
U C I Q E = c 1 × σ c + c 2 × c o n l + c 3 × μ s ,
where σ c is the chromaticity standard deviation, c o n l represents the contrast in brightness, μ s represents the average value of saturation, and c 1 , c 2 , and c 3 are weighting coefficients.
(2)
The underwater image quality measurement [25] (UIQM) is a quality-evaluated indicator of non-reference underwater images based on human visual system excitation. The calculation formula is as follows:
U I Q M = c 1 × U I C M + c 2 × U I S M + c 3 × U I C o n M ,
where c 1 is set to 0.0282, c 2 is set to 0.2953, and c 3 is set to 3.5735. The underwater image quality measurement is a linear combination of the underwater image colorfulness measure (UICM), underwater image sharpness measure (UISM), and underwater image contrast measure (UICONM). The higher the UIQM, the better the image’s color balance, sharpness, and contrast.
(3)
The structural similarity index measurement [26] (SSIM) is an index for determining how similar the two images are. When two images, x and y , are given, the calculation formula is:
S S I M ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 ) ,
where μ x and μ y are the average of x and y , respectively; σ x 2 , σ y 2 are the variance of x and y ; and c i = k i L , ( i = 1 , 2 ) is a constant to maintain stability. σ x y is the covariance of x and y ; k 1 = 0.01 , k 2 = 0.03 .
(4)
The peak signal-to-noise ratio (PSNR) is an index to measure image quality. The calculation formula for the mean square error (MSE) is:
M S E = 1 m n i = 0 m 1 j = 0 n 1 I o ( i , j ) I p 2 ,
where two images, I o , I p , are compared. The PSNR is obtained through the MSE, and the calculation formula is:
P S N R = 10 log 10 ( ( 2 n 1 ) 2 M S E )
(5)
The CIEDE2000 evaluation index [27], which has a range of [0, 100], measures the color changes between the standard color card and each processed color block. The color differences are reduced when the index decreases. For the evaluation in Figure 8, we used the CIEDE2000 evaluation index. Table 1 displays the results.
Comparing the data in Table 1, we can see that, like DRGAN, FUnIE, and Semi-UIR both achieve good results, and FUnIE adds residual connections to the generator to enhance the network performance. DRGAN’s CIEDE2000 average result is the lowest, showing that our technique performs better in terms of color recovery.
We used UCIQE to evaluate the images in Figure 9, and the results are shown in Table 2. The results show that the average value of the DRGAN algorithm is higher than that of other algorithms. For Images 1 and Image 2, the DRGAN UCIQE was lower than that of L^2UWE because the original image was less degradable, and Semi-UIR recovery was better than DRGAN enhancement. However, ICCB, with a higher UCIQE, showed significant color aberration in Image 6 and unnatural color restoration in Image 8.
The UIQM results from Figure 9 are shown in Table 3; our average UIQM for DRGAN is higher than for the other algorithms. The light degradation of Image 1 and Image 3 leads to a higher UIQM in the ICCB restoration algorithm than the enhancement effect of DRGAN, and when processed via FUnIE, Image 2 has a yellow color cast. Although Semi-UIR achieves good enhancement results, it is not thorough enough in detail processing, as shown in Image 8.
As the RUIE dataset has no ground truth, we chose the UIQE and UIQM metrics when comparing with other algorithms, and we used the UCIQE, UIQM, SSIM, and PSNR metrics on the Underwater ImageNet dataset. The test results using the Underwater ImageNet dataset and RUIE dataset are shown in Table 4 and Table 5.
We verified the effectiveness of DRGAN on the Underwater ImageNet dataset and applied the model trained on the Underwater ImageNet dataset to the RUIE dataset. In the Underwater ImageNet dataset, DRGAN’s PSRN, UIQM, and UCIQE outperformed the other algorithms, indicating that the DRGAN enhancement results are closer to the real images. On the RUIE dataset, on average, DRGAN also achieved better results. While FUnIE also adds residual connections, it is only in the Green and Atomization environments in the RUIE dataset that the SSIM indicators for the Underwater ImageNet dataset are better than those of DRGAN. From the above, it can be concluded that the addition of dense and residual connections in DRGAN has a better performance-enhancing effect and leads to a better generalization ability.

4.4. Ablation Study

We conducted module ablation experiments using the Underwater ImageNet dataset. Firstly, we evaluated the images processed via different modules using PSNR, SSIM, UCIQE, and UIQM. Table 6 shows the objective metric scores for the ablation experiments, where w/o MSFEM denotes the removal of the multi-scale feature extraction module, w/o DRB denotes the removal of the dense residual block, w/o RES denotes the removal of residual connectivity in the DRB, and w/o DEN denotes the removal of dense connectivity in the DRB.
Table 6 shows the performance w/o MSFEM and w/o DRB on UCIQ, where it can be seen that the performance w/o MSFEM on UIQM is higher than that w/o DRB. Removing both the dense connections and the residual connections influences the model performance. This result fully demonstrates the importance of the two modules we adopted, MSFEM and DRB, for the overall performance of the network.
Then, we randomly selected an image for subjective comparison. Figure 10 shows that the image processed w/o DRB has artifacts and is accompanied by a yellow color cast, while the image processed w/o MSFEM is subjectively better than that w/o DRB, but still has a small amount of color cast. The image enhanced via the full processing model is the best and the most visually natural. Figure 10 also demonstrates that the image color recovery is poor, and there are artifacts, after the removal of the residual connections in the dense residual block. On the contrary, after the removal of the dense connections in the dense residual block, the image is over-enhanced.

4.5. Additional Experiments

Less image feature information makes underwater image detection more challenging. As shown in Figure 11 and Figure 12, several images were selected for surf feature point matching and Canny operator experiments, which verified that our method can enhance edges and other feature information in underwater images.
Figure 11 shows the results from the surf feature point matching; it can be seen that the processed image has significantly more feature points than the original underwater image. These experiments show that the proposed algorithm successfully enriches the characteristics of underwater images, making the subsequent information processing much easier.
Figure 12 shows the results of the Canny operator; after the processing in this method, more details of the image can be added (such as coral patterns, etc.). Compared with the degraded images, DRGAN can clearly show the contour information of the picture. This makes the detection and tracking of features of interest via underwater robots a much less taxing endeavor.

5. Conclusions

In this paper, we propose DRGAN as a means of enhancing underwater images, drawing inspiration from ResNet and DenseNet. Through the utilization of a multi-scale feature extraction module and a dense residual block in the generator, multi-scale feature information is integrated. The incorporation of these multi-stage features broadens the receptive field and safeguards against any decline in network performance due to gradient disappearance. Additionally, DRGAN optimizes network utilization by leveraging the benefits of both residual and dense connections. It is worth mentioning that the generator’s computational efficiency has been enhanced in comparison to networks that solely rely on dense blocks. We employ a discriminator akin to PatchGAN for adversarial training, and this augments the generator’s ability to sharpen images. The findings from the experiments conducted on intricate underwater scenes indicate that DRGAN greatly enhances the quality of images in comparison to various renowned techniques. In the future, we plan to use the proposed method in other areas of marine engineering, such as object recognition and detection within wider underwater scenes.

Author Contributions

The authors confirm contributions to the paper as follows: study conception and design: J.Q. and H.L.; data collection: B.Z. and S.L.; analysis and interpretation of results: J.Q., H.L. and X.X.; draft manuscript preparation: J.Q. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by two funding sources: the Natural Science Foundation of the Jiangsu Higher Education Institutions of China (No.22KJB520036 and No.23KJB510033) and the Natural Science Foundation of Jiangsu Province (No. BK20211357).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The underwater image data that support the findings of this study are openly available at https://irvlab.cs.umn.edu/resources/euvp-dataset (accessed on 30 May 2022) and https://github.com/dlut-dimt/Realworld-Underwater-Image-Enhancement-RUIE-Benchmark (accessed on 19 January 2022).

Acknowledgments

The authors would like to thank the editor and reviewers for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, C.; Guo, J.; Wang, B.; Cong, R.; Zhang, Y.; Wang, J. Single underwater image enhancement based on color cast removal and visibility restoration. J. Electron. Imaging 2016, 25, 033012. [Google Scholar] [CrossRef]
  2. Sun, P.; Sun, C.; Wang, R.; Zhao, X. Object Detection Based on Roadside LiDAR for Cooperative Driving Automation: A Review. Sensors 2022, 22, 9316. [Google Scholar] [CrossRef] [PubMed]
  3. Drews, P.; Nascimento, E.; Moraes, F.; Botelho, S.; Campos, M. Transmission estimation in underwater single images. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Sydney, Australia, 2–8 December 2013; pp. 825–830. [Google Scholar]
  4. He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar] [CrossRef]
  5. Ma, X.; Chen, Z.; Feng, Z. Underwater image restoration through a combination of improved dark channel prior and gray world algorithms. J. Electron. Imaging 2019, 28, 053033. [Google Scholar] [CrossRef]
  6. Ancuti, C.O.; Ancuti, C.; De Vleeschouwer, C.; Bekaert, P. Color balance and fusion for underwater image enhancement. IEEE Trans. Image Process. 2017, 27, 379–393. [Google Scholar] [CrossRef] [PubMed]
  7. Liang, Z.; Wang, Y.; Ding, X.; Mi, Z.; Fu, X. Single underwater image enhancement by attenuation map guided color correction and detail preserved dehazing. Neurocomputing 2021, 425, 160–172. [Google Scholar] [CrossRef]
  8. Marques, T.P.; Albu, A.B. L2UWE: A Framework for the Efficient Enhancement of Low-Light Underwater Images Using Local Contrast and Multi-Scale Fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition workshops, Seattle, WA, USA, 14–19 June 2020; pp. 538–539. [Google Scholar]
  9. Li, J.; Skinner, K.A.; Eustice, R.M.; Johnson-Roberson, M. WaterGAN: Unsupervised Generative Network to Enable Real-time Color Correction of Monocular Underwater Images. IEEE Robot. Autom. Lett. 2017, 3, 387–394. [Google Scholar] [CrossRef]
  10. Fabbri, C.; Islam, M.J.; Sattar, J. Enhancing underwater imagery using generative adversarial networks. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 7159–7165. [Google Scholar]
  11. Guo, Y.; Li, H.; Zhuang, P. Underwater Image Enhancement Using a Multiscale Dense Generative Adversarial Network. IEEE J. Ocean. Eng. 2019, 45, 862–870. [Google Scholar] [CrossRef]
  12. Islam, M.J.; Xia, Y.; Sattar, J. Fast Underwater Image Enhancement for Improved Visual Perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
  13. Chen, X.; Yu, J.; Kong, S.; Wu, Z.; Fang, X.; Wen, L. Towards Real-Time Advancement of Underwater Visual Quality With GAN. IEEE Trans. Ind. Electron. 2019, 66, 9350–9359. [Google Scholar] [CrossRef]
  14. Huang, S.; Wang, K.; Liu, H.; Chen, J.; Li, Y. Contrastive Semi-Supervised Learning for Underwater Image Restoration via Reliable Bank. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18145–18155. [Google Scholar]
  15. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  16. Wei, B.; Wang, D.; Wang, Z.; Zhang, L. Single Image Desnow Based on Vision Transformer and Conditional Generative Adversarial Network for Internet of Vehicles. Comput. Model. Eng. Sci. 2023, 137, 1975–1988. [Google Scholar] [CrossRef]
  17. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  18. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  19. Li, C.; Wand, M. Precomputed Real-Time Texture Synthesis with Markovian Generative Adversarial Networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 17 September 2016; pp. 702–716. [Google Scholar] [CrossRef]
  20. Wang, G.; Tian, J.; Li, P. Image color correction based on double transmission underwater imaging model. Acta Opt. Sin. 2019, 39, 0901002. [Google Scholar] [CrossRef]
  21. Yadav, A.; Upadhyay, M.; Singh, G. Underwater Image Enhancement Using Convolutional Neural Network. arXiv 2021, arXiv:2109.08916. [Google Scholar]
  22. Liu, R.; Fan, X.; Zhu, M.; Hou, M.; Luo, Z. Real-World Underwater Enhancement: Challenges, Benchmarks, and Solutions Under Natural Light. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 4861–4875. [Google Scholar] [CrossRef]
  23. Fernández, P.D.M.; Peña, F.A.G.; Ren, T.I.; Leandro, J.J. Fast and robust multiple ColorChecker detection using deep convolutional neural networks. Image Vis. Comput. 2019, 81, 15–24. [Google Scholar] [CrossRef]
  24. Yang, M.; Sowmya, A. An Underwater Color Image Quality Evaluation Metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef] [PubMed]
  25. Panetta, K.; Gao, C.; Agaian, S. Human-Visual-System-Inspired Underwater Image Quality Measures. IEEE J. Ocean. Eng. 2015, 41, 541–551. [Google Scholar] [CrossRef]
  26. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  27. Sharma, G.; Wu, W.; Dalal, E.N. The CIEDE2000 color-difference formula: Implementation notes, supplementary test data, and mathematical observations. Color Res. Appl. 2005, 30, 21–30. [Google Scholar] [CrossRef]
Figure 1. Image comparison: (a) degraded image, (b) FUnIE-GAN, (c) UDCP, and (d) ours; the image in the second row is the enlarged detail image of the yellow box in the first row.
Figure 1. Image comparison: (a) degraded image, (b) FUnIE-GAN, (c) UDCP, and (d) ours; the image in the second row is the enlarged detail image of the yellow box in the first row.
Sensors 23 08297 g001
Figure 2. The residual network structure.
Figure 2. The residual network structure.
Sensors 23 08297 g002
Figure 3. The diagram of our densely connected convolutional network structure. Different colors are used here to distinguish and emphasize the different nodes of the densely connected convolutional network structure.
Figure 3. The diagram of our densely connected convolutional network structure. Different colors are used here to distinguish and emphasize the different nodes of the densely connected convolutional network structure.
Sensors 23 08297 g003
Figure 4. The diagram of our generative network structure, where different colors represent convolutional layers with different convolutional kernels.
Figure 4. The diagram of our generative network structure, where different colors represent convolutional layers with different convolutional kernels.
Sensors 23 08297 g004
Figure 5. The diagram of our multi-scale feature extraction module.
Figure 5. The diagram of our multi-scale feature extraction module.
Sensors 23 08297 g005
Figure 6. The number of dense residual block cycles.
Figure 6. The number of dense residual block cycles.
Sensors 23 08297 g006
Figure 7. The diagram of our discriminator network structure.
Figure 7. The diagram of our discriminator network structure.
Sensors 23 08297 g007
Figure 8. A qualitative comparison of the results of the color recovery experiment: (a) the degraded image, (b) Fusion, (c) ICCB, (d) L^2UWE, (e) FUnIE, (f) UWCNN, (g) Semi-UIR, (h) ours, and (i) the standard color card.
Figure 8. A qualitative comparison of the results of the color recovery experiment: (a) the degraded image, (b) Fusion, (c) ICCB, (d) L^2UWE, (e) FUnIE, (f) UWCNN, (g) Semi-UIR, (h) ours, and (i) the standard color card.
Sensors 23 08297 g008
Figure 9. A qualitative comparison of the results of the complex underwater environment experiment, including our method and SOTA methods: (a) the degraded images, (b) Fusion, (c) ICCB, (d) L^2UWE, (e) FUnIE, (f) UWCNN, (g) Semi-UIR, and (h) ours.
Figure 9. A qualitative comparison of the results of the complex underwater environment experiment, including our method and SOTA methods: (a) the degraded images, (b) Fusion, (c) ICCB, (d) L^2UWE, (e) FUnIE, (f) UWCNN, (g) Semi-UIR, and (h) ours.
Sensors 23 08297 g009
Figure 10. Image showing the ablation experiment results: (a) the degraded image, (b) w/o MSFEM, (c) w/o RES, (d) w/o DEN, (e) w/o DRB, and (f) DRGAN.
Figure 10. Image showing the ablation experiment results: (a) the degraded image, (b) w/o MSFEM, (c) w/o RES, (d) w/o DEN, (e) w/o DRB, and (f) DRGAN.
Sensors 23 08297 g010
Figure 11. Feature point matching: (left) the degraded image, (right) ours.
Figure 11. Feature point matching: (left) the degraded image, (right) ours.
Sensors 23 08297 g011
Figure 12. The Canny operator results. The images in the first and third rows represent degraded and enhanced images following our algorithm, while the images in the second and fourth rows represent the outcomes of the Canny operator detection output. Additionally, the red box line in the fourth row illustrates the more comprehensive image information that can be acquired through the processing of our algorithm.
Figure 12. The Canny operator results. The images in the first and third rows represent degraded and enhanced images following our algorithm, while the images in the second and fourth rows represent the outcomes of the Canny operator detection output. Additionally, the red box line in the fourth row illustrates the more comprehensive image information that can be acquired through the processing of our algorithm.
Sensors 23 08297 g012
Table 1. The evaluation results for CIEDE2000; the black bold font represents the best data.
Table 1. The evaluation results for CIEDE2000; the black bold font represents the best data.
MethodSensors 23 08297 i001Sensors 23 08297 i002Sensors 23 08297 i003Sensors 23 08297 i004Sensors 23 08297 i005Sensors 23 08297 i006Sensors 23 08297 i007Sensors 23 08297 i008Avg
Fusion26.3433.7433.0232.2735.4422.7029.9319.0527.23
29.3919.3438.3739.6113.5239.2019.6340.03
25.3128.8221.0822.2026.4830.7319.308.02
ICCB26.1829.7140.4935.8427.6532.6011.1548.1924.04
21.5840.9414.3120.0628.2529.168.779.24
16.1031.8411.4317.5227.4021.0414.5712.85
L^2UWE18.2017.1716.0622.5813.1211.5420.9415.9616.35
17.207.3425.4227.6010.0427.677.8625.79
8.2011.5415.3014.8014.2218.5214.8910.67
FUnIE17.9118.1715.0623.5812.5210.6421.2816.2316.82
17.808.3627.4528.209.0729.656.8227.79
8.3011.9415.1515.5015.1920.4715.8810.63
UWCNN23.1140.2331.1132.6337.5138.1033.1117.5428.40
24.5016.2146.9054.8518.0238.7115.2116.21
23.6126.5441.9118.1137.6031.7415.143.80
Semi-UIR21.8010.9013.1315.8112.2410.5124.8213.9114.14
17.2114.9210.6217.947.942.7215.2017.50
12.919.4115.609.227.326.2121.9110.51
Ours19.119.278.3413.909.8513.4122.8011.4013.79
39.3412.9111.216.813.9015.617.213.51
2.717.105.308.1037.6313.7115.1013.81
Table 2. The quantitative comparison using the UCIQE dataset; the black bold font represents the best data.
Table 2. The quantitative comparison using the UCIQE dataset; the black bold font represents the best data.
No.FusionICCBL^2UWEFUnIEUWCNNSemi-UIROurs
10.59700.55790.60570.57910.52720.58910.5982
20.50240.56460.60630.60120.52580.57460.6012
30.51840.51010.53310.53310.51180.63160.6470
40.43610.37770.49710.49710.51890.61980.6200
50.55040.69300.68630.53590.50060.60660.6012
60.54030.55720.68260.67410.59200.59500.6976
70.49140.53120.54040.54040.46050.57360.5821
80.47700.55480.56940.56940.45630.70850.6072
Average0.51170.54330.59010.56740.50830.61240.6218
Table 3. The quantitative comparison using the UIQM dataset; the black bold font represents the best data.
Table 3. The quantitative comparison using the UIQM dataset; the black bold font represents the best data.
No.FusionICCBL^2UWEFUnIEUWCNNSemi-UIROurs
14.86455.05694.55004.24654.47364.82774.8126
24.73725.09875.11555.25984.44044.74554.8373
35.41215.89374.93763.78094.29004.77414.3680
44.02544.09173.37994.04784.14964.62464.6264
54.33094.70023.58944.09044.44314.69954.8586
63.83583.94903.04484.62194.22854.33884.9539
73.59901.68843.20243.83662.42503.67054.4406
83.58392.84502.38272.27341.82413.88424.5264
Average4.29864.16543.77524.01963.78424.47064.8029
Table 4. The quantitative comparison using the Underwater ImageNet dataset; the black bold font represents the best data.
Table 4. The quantitative comparison using the Underwater ImageNet dataset; the black bold font represents the best data.
MethodSSIMPSNRUIQMUCIQE
Fusion0.63625.7944.2660.541
ICCB0.74831.3414.3760.608
L^2UWE0.74626.3874.3380.601
FUnIE0.77434.2574.9950.588
UWCNN0.64828.8684.4630.507
Semi-UIR0.82731.0335.0370.592
Ours0.66137.5925.0870.630
Table 5. The quantitative comparison using RUIE. (M) and (E) represent UIQM and UCIQE, respectively. The black bold font represents the best data.
Table 5. The quantitative comparison using RUIE. (M) and (E) represent UIQM and UCIQE, respectively. The black bold font represents the best data.
MethodBlueGreenAtomizationNormalAverage
(M)(E)(M)(E)(M)(E)(M)(E)(M)(E)
Fusion3.7700.5804.4000.5334.5160.4954.7010.5874.3500.545
ICCB5.4220.5884.2720.5684.7420.5824.4500.6104.6980.585
L^2UWE4.3940.5754.6060.5785.3360.5195.2730.5835.0030.545
FUnIE3.8360.6284.1650.6024.2860.5994.3950.6174.1700.611
UWCNN4.1890.5644.1470.5244.0960.5215.0580.5904.3390.547
Semi-UIR5.0590.5764.9730.5865.0690.5765.470.6295.1010.597
Ours5.6800.6534.8300.6344.7670.6405.8700.6405.3600.641
Table 6. The ablation experiment results for different variants of DRGAN; w/o refers to without. The black bold font represents the best data.
Table 6. The ablation experiment results for different variants of DRGAN; w/o refers to without. The black bold font represents the best data.
ModulesBaselinesPSNRSSIMUCIQEUIQM
MSFEMw/o MSFEM34.3720.6530.6154.734
DRBw/o RES35.1320.6510.6194.833
w/o DEN34.8760.6510.6144.791
w/o DRB32.1950.6490.6094.592
DRGANfull model37.5920.6610.6305.087
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qian, J.; Li, H.; Zhang, B.; Lin, S.; Xing, X. DRGAN: Dense Residual Generative Adversarial Network for Image Enhancement in an Underwater Autonomous Driving Device. Sensors 2023, 23, 8297. https://doi.org/10.3390/s23198297

AMA Style

Qian J, Li H, Zhang B, Lin S, Xing X. DRGAN: Dense Residual Generative Adversarial Network for Image Enhancement in an Underwater Autonomous Driving Device. Sensors. 2023; 23(19):8297. https://doi.org/10.3390/s23198297

Chicago/Turabian Style

Qian, Jin, Hui Li, Bin Zhang, Sen Lin, and Xiaoshuang Xing. 2023. "DRGAN: Dense Residual Generative Adversarial Network for Image Enhancement in an Underwater Autonomous Driving Device" Sensors 23, no. 19: 8297. https://doi.org/10.3390/s23198297

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop