Next Article in Journal
Low-Altitude Aerial Video Surveillance via One-Class SVM Anomaly Detection from Textural Features in UAV Images
Previous Article in Journal
Designing Automated Deployment Strategies of Face Recognition Solutions in Heterogeneous IoT Platforms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unpaired Underwater Image Enhancement Based on CycleGAN

1
Institute of Microelectronics of the Chinese Academy of Sciences, Beijing 100029, China
2
School of Microelectronics, University of Chinese Academy of Sciences, Beijing 100049, China
3
Information Research Center of Military Science, Beijing 100142, China
4
Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Information 2022, 13(1), 1; https://doi.org/10.3390/info13010001
Submission received: 28 November 2021 / Revised: 17 December 2021 / Accepted: 18 December 2021 / Published: 22 December 2021

Abstract

:
Underwater image enhancement recovers degraded underwater images to produce corresponding clear images. Image enhancement methods based on deep learning usually use paired data to train the model, while such paired data, e.g., the degraded images and the corresponding clear images, are difficult to capture simultaneously in the underwater environment. In addition, how to retain the detailed information well in the enhanced image is another critical problem. To solve such issues, we propose a novel unpaired underwater image enhancement method via a cycle generative adversarial network (UW-CycleGAN) to recover the degraded underwater images. Our proposed UW-CycleGAN model includes three main modules: (1) A content loss regularizer is adopted into the generator in CycleGAN, which constrains the detailed information existing in one degraded image to remain in the corresponding generated clear image; (2) A blur-promoting adversarial loss regularizer is introduced into the discriminator to reduce the blur and noise in the generated clear images; (3) We add the DenseNet block to the generator to retain more information of each feature map in the training stage. Finally, experimental results on two unpaired underwater image datasets produced satisfactory performance compared to the state-of-the-art image enhancement methods, which proves the effectiveness of the proposed model.

1. Introduction

With the rapid development of marine resources, underwater robots are necessary to replace humans to work in the complex underwater environment. The underwater robot mainly relies on its visual ability to achieve several tasks, such as object recognition, localization, 3D reconstruction, and route guidance. Due to the light absorption and scattering properties in water, underwater images usually contain color distortion and low contrast. Therefore, how to enhance underwater images becomes an urgent problem for practical underwater applications [1].
Over recent decades, underwater image enhancement has attracted an increasing amount of attention. Wang et al. [2] divided underwater image enhancement methods into three main categories: spatial-domain methods, transform-domain methods, and the popular deep learning-based methods.
The spatial-domain methods usually tune the grayscale range of one image to enhance its contrast and reduce the color distortion [3]. Traditional methods include gray world [4], white balance [4], automatic white balance [5], histogram equalization, adaptive histogram equalization [6], contrast limited adaptive histogram equalization [7], and its variations. Although these methods have had success in enhancing degraded images, they still have significant limitations for severely degraded underwater images, which introduce red artifacts and noise.
The transform-domain methods transfer an underwater image to the frequency domain, then, they enhance the image contrast by amplifying the high-frequency information and suppressing the low-frequency information. Classic transform-domain methods include a low-pass filter [8], high-pass filter [9], homomorphic filter [10], and wavelet transform [11,12,13]. Although these methods decrease the noise and enhance the contrast of an underwater image, the performance of color correction is bad.
The above two categories of methods just enhance each underwater image independently, without learning procedures. Deep learning-based methods can exploit an end-to-end automatic training mechanism to enhance underwater images, which learns the intrinsic underwater features from a set of underwater images. In [14], they replaced the handcrafted features with the nonparametric deep features for the image representation. Other researchers [15,16] introduced the convolutional neural network (CNN) into underwater image enhancement applications. A residual CNN was further proposed in [17]. Furthermore, [18] provided a deep pixel-to-pixel network by designing an encoding–decoding framework. In [19], they utilized domain adversarial learning to enhance underwater images. In [20,21], they improved the quality of visual underwater scenes using Generative Adversarial Networks (GAN), and then [22] proposed a fusion adversarial network. Finally, Hu et al. [23] introduced the natural image quality evaluation to a supervised generative adversarial network.
These methods improved the visual effect and quality of underwater images, but they require a large amount of paired data, i.e., each degraded image has a corresponding clear image. Paired data is difficult to obtain in an underwater environment, which also causes difficulty for underwater image enhancement. Therefore, researchers usually use synthetic data to construct paired data. Figure 1 shows some samples of unpaired underwater images.
To solve the problem of deep learning-based underwater image enhancement methods requiring paired data, we propose a novel underwater cycle generative adversarial network (UW-CycleGAN) for image enhancement, which just needs one set of unpaired underwater degraded images and clear images to train the proposed model. A brief illustration of UW-CycleGAN is shown in Figure 2.
The main contributions of this paper are briefly summarized as follows:
  • We introduce a content loss regularizer into the generator in CycleGAN, which keeps more detailed information in the corresponding generated clear image. This strategy is different from CartoonGAN [24];
  • We add a blur-promoting adversarial loss regularizer into the discriminator in CycleGAN, which reduces the effects of blur and noise and enhances the image clarity;
  • We exploit the improved DenseNet Block in the generator to strengthen the forward transfer of feature maps, so that every feature map can be utilized;
  • We test our proposed UW-CycleGAN on different types of underwater images and obtain a satisfactory performance.
We develop an end-to-end underwater image enhancement system. The structure of this paper is organized as follows: The necessary knowledge about underwater image enhancement is reviewed in Section 2. An improved underwater CycleGAN model for unpaired data, so-called UW-CycleGAN, is proposed in Section 3. The experimental results on two underwater image datasets are illustrated in Section 4. Finally, we conclude this paper in Section 5.

2. Underwater Image Enhancement

As we mentioned above, capturing paired data in the underwater environment is difficult. To study the intrinsic relationship between the degraded image and the corresponding clear image, some researchers designed a simplified physical model according to the refraction, scattering, and attenuation properties of light [25],
I λ ( x ) = J λ ( x ) · t λ ( x ) + ( 1 t λ ( x ) ) · B λ , λ { r , g , b }
where I λ ( x ) denotes the degraded image captured by underwater cameras, J λ ( x ) means the corresponding restored clear image, t λ ( x ) is the medium transmission map, B λ represents the well-proportioned background light, and λ gives the light wavelength.
In order to enhance J λ ( x ) , the key problem of the traditional physical models is to estimate t λ ( x ) and B λ , since only image I λ ( x ) is known. Although this physical model does not need paired data, some assumptions and prior knowledge are required to evaluate t λ ( x ) and B λ , which severely limits the practical applications.
In recent years, many researchers have applied CNN to process underwater images and achieved good results in underwater image enhancement applications. However, CNN based methods need paired data to train their network models, and researchers have to use synthetic data instead. Fortunately, CycleGAN can utilize unpaired data for the conversion of image style, which offers a new direction for underwater image enhancement.

3. Underwater CycleGAN (UW-CycleGAN)

Deep learning-based image enhancement methods usually need paired underwater images to train network models. To solve this problem, we propose a CycleGAN-based underwater image enhancement method (UW-CycleGAN), which can utilize unpaired data to train its model.
Suppose we have the unpaired degraded image set X and clear image set Y. One complete procedure of UW-CycleGAN is shown in Figure 2:
(1)
The mapping function G generates the clear image G ( x ) from x X .
(2)
Another mapping function F reconstructs the degraded image x by G ( x ) F ( G ( x ) ) .
(3)
Discriminator D Y judges whether the generated image G ( x ) and clear image y derive from the same distribution.
In addition, y F ( y ) G ( F ( y ) ) and D X are the similar inverse processes, which ensures the model invertibility. We display some samples of x, G ( x ) , and F ( G ( x ) ) in Figure 3, respectively.

3.1. Loss Function

Zhu et al. [26] proposed a CycleGAN framework to achieve unpaired image-to-image translation, which consisted of adversarial loss and cycle consistency loss.
The adversarial loss restricts the generated image G ( x ) and F ( y ) to derive from the same distribution with the clear image y and degraded image x, respectively:
L a d v ( G , D Y ) = E y P d a t a ( y ) [ ( D Y ( y ) ) 2 ] + E x P d a t a ( x ) [ ( 1 D Y ( G ( x ) ) ) 2 ] L a d v ( F , D X ) = E x P d a t a ( x ) [ ( D X ( x ) ) 2 ] + E y P d a t a ( y ) [ ( 1 D X ( F ( y ) ) ) 2 ]
where P d a t a ( x ) and P d a t a ( y ) represent the distributions of underwater degraded images and clear images, respectively.
The cycle consistency loss ensures the reconstructed images are similar to the input images,
L c y c ( G , F ) = E x P d a t a ( x ) [ F ( G ( x ) ) x 1 ] + E y P d a t a ( y ) [ G ( F ( y ) ) y 1 ] .
where · 1 means the 1 -norm.

3.1.1. Content Loss

The cycle consistency loss (3) only minimizes the difference between the input image x (or y) and its reconstructed image F ( G ( x ) ) (or G ( F ( y ) ) ), which ignores whether the generated image G ( x ) (or F ( y ) ) is visually similar to x (or y). Therefore, we add the content loss regularizer measured by 1 norm,
L c o n ( G , F ) = E x P d a t a ( x ) [ G ( x ) x 1 ] + E y P d a t a ( y ) [ F ( y ) y 1 ]
However, the above function makes the generated image G ( x ) (or F ( y ) ) too similar to the input image x (or y) due to the element-wise subtraction, retaining almost all the information of the input image x (or y). We want to keep the detailed information of the input image x (or y) unchanged, meanwhile, calibrating the image color to enhance the visual quality of the generated image G ( x ) (or F ( y ) ).
In order to achieve this purpose, a VGG19 pretraining network is used to extract the c o n v 4 _ 4 layer feature maps of the input and generated images. We also employ 1 -norm to measure the content loss, since 1 -norm is more robust to noise and outliers, which can recover the underwater image details well. Thus, the new content loss regularizer is rewritten as,
L c o n ( G , F ) = E x P d a t a ( x ) [ V G G ( G ( x ) ) V G G ( x ) 1 ] + E y P d a t a ( y ) [ V G G ( F ( y ) ) V G G ( y ) 1 ]
where V G G ( · ) denotes the VGG19 feature map in this paper.

3.1.2. Blur-Promoting Adversarial Loss

Although the content of the image generated by G is consistent with its corresponding input image, a large amount of noise and blur are also generated at the same time, which effects the visual performance. We need to make the discriminator robust to blur. Therefore, a blur dataset Z is constructed by adding Gaussian blur to the clear image dataset Y. The discriminator D Y should judge z Z as the fake image and y as the real image, so that the images generated by generator G can be clearer. With this idea, the adversarial loss can be rewritten as follows:
L b a d v ( G , D Y ) = E y P d a t a ( y ) [ ( D Y ( y ) ) 2 ] + E z P d a t a ( z ) [ ( 1 D Y ( z ) ) 2 ] + E x P d a t a ( x ) [ ( 1 D Y ( G ( x ) ) ) 2 ] ,
where P d a t a ( z ) represents the distribution of underwater clear images with Gaussian blur.

3.1.3. Full Loss Funtion

Finally, we construct the full loss function of UW-CycleGAN as follows,
L ( G , F , D X , D Y ) = L b a d v ( G , D Y ) + L a d v ( F , D X ) + L c y c ( G , F ) + L c o n ( G , F ) ,
where the generator G, F and the discriminator D X , D Y can be updated by
G * , F * = a r g min G , F max D x , D y L ( G , F , D X , D Y ) .
It should be noted, we equally treat each loss regularizer and never set several hyperparameters to tune experimental results.

3.2. Network Architectures

As illustrated in Figure 2, our UW-CycleGAN network architecture consists of two main generators and two discriminators.
Generators G and F have the same network structure with different parameters, which directly adopts the encoder–decoder structure in this paper. Firstly, one flat convolution stage with convolution kernel size of 7 × 7 and step length of 1 and two down-convolution stages with convolution kernel size of 3 × 3 and step length of 2 are exploited to spatially compress and encode the input images. Then, three DenseNet blocks are used to transfer the feature maps and preserve their high-level features. The detailed structure of the DenseNet block is shown in Figure 4b [27]. In the T layer, with 128 convolution kernels size of 1 × 1 and step length of 1, we reduce the size of feature maps from 64 × 64 × 256 to 64 × 64 × 128 . In the L1 layer, L1 includes 64 convolution kernels size of 1 × 1 and step length 1 and 16 convolution kernels size of 3 × 3 and step length 1, so the size of the output feature maps is 64 × 64 × 16 . Then, we concatenate the output of the T layer and the output of the L1 layer to obtain a feature map size of 64 × 64 × 144 as the inputs for the L2 layer. Similarly, the outputs of the L1 and L2 layers are concatenated as the input for the L3 layer. After several similar operations, we obtain the output size of 64 × 64 × 256 in the L8 layer. Finally, the generated clear images are reconstructed by two up-convolutions, which contain one convolution kernel size of 3 × 3 and step length 1 / 2 and one final convolution kernel size of 7 × 7 and step length 1.
Discriminators D X and D Y also contain the same network structure with different parameters. In the discriminator network, a Markov discriminator, which comprises five full convolution layers outputs a “0–1” indicator matrix size of 70 × 70 and then calculates the mean value of all elements in the matrix as the real/fake output at last.

4. Experiment and Evaluation

In this section, UW-CycleGAN is tested on two real-world unpaired underwater image datasets and is compared with several classic image enhancement methods to evaluate the superiority of UW-CycleGAN. Finally, the ablation experiments verify the importance of each component in our UW-CycleGAN model.

4.1. Datasets and Metrics

URPC2019 (http://www.cnurpc.org/a/js/2019/0805/125.html accessed on 17 December 2021) contains over 4000 underwater images [28], and we took a subset in this paper. We chose 670 underwater images as the training set, in which 335 degraded images belonged to training set X, and the remaining 335 clear images belonged to training set Y. There was no paired relationship between training sets X and Y. The Gaussian blur set E was formed by performing a Gaussian blur operation on the previous training set Y. The testing set consisted of 70 degraded images. We set the color image size of both the training set and the test set as 256 × 256 × 3 .
EUVP (http://irvlab.cs.umn.edu/resources/euvp-dataset accessed on 17 December 2021) contains over 6446 underwater human images. We choose 405 degraded images as the training set X and 405 clear images as the training set Y. The testing set consisted of 200 degraded images. The color image size was also set as 256 × 256 × 3 .
To fairly assess these image enhancement methods from different aspects, we selected three standard metrics, which were average gradient (AG) [29], information entropy (IE) [30], and underwater image quality measure (UIQM) [31]. Lower values of IE reflect better performance, while AG and UIQM are the opposite. The entire network was coded in the pytorch framework and implemented on a workstation with 8 Nvidia Tesla P100 GPUs.

4.2. Experimental Assessment

If only the clear image set Y was used to train the discriminator, the generated image usually had an obvious blur. To solve this problem, we exploited the Gaussian blur set Z to train the discriminator, and the generator could output clear images well.
We compared the proposed model with three traditional underwater image enhancement methods:
  • De-scattering Underwater image (Deunderwater) [32]
  • Diving into Haze-Lines: Color Restoration of Underwater Images (HL) [33]
  • Unsupervised Color Correction Method (UCM) [34]
and three deep learning-based methods:
  • Fast Underwater Image Enhancement for Improved Visual Perception (FUnIE-GAN-UP) [35]
  • Generative Adversarial Networks for Photo Cartoonization (CartoonGAN) [24]
  • Unpaired Image-to-Image Translation using Cycle Consistent Adversarial Networks (CycleGAN) [26].
Figure 5 displays the enhancement results of five underwater scene images on the URPC dataset. Deunderwater recovered the image color to a certain extent, but the contrast and details of the generated image were not good. Although HL restored the image detailed information well, the generated image had the problems of poor contrast and color distortion. UCM and FUnIE-GAN-UP recovered the image color and detail well, while the contrast was relatively bad. CartoonGAN and CycleGAN improved the image color and contrast excellently, but blur existed in the image detail. Our UW-CycleGAN obtained good performance in image contrast, color, and detail.
Figure 6 shows the vision enhancement results of underwater human images on the EUVP dataset. Obviously, Deunderwater, HL and UCM had color distortion problems. FUnIE-GAN-UP, CatroonGAN, and CycleGAN performed reasonably well, and their enhanced images were comparable to our UW-CycleGAN, but UW-CycleGAN was still the best in terms of the image detail and clarity.
Table 1 and Table 2 estimate the underwater image enhancement performance of all methods under three standard metrics. Although the value performance of traditional methods (Deunderwater, HL, and UCM) in both tables are not bad, their visualization performance in Figure 5 and Figure 6 are obviously unsatisfactory. Deep learning-based methods work relatively steadily in both value and visulization performance. Our UW-CycleGAN obtained the best experimental results measured by all objective evaluation metrics.

4.3. Ablation Experiments

We designed a set of ablation experiments to further analyze the importance of each module in our UW-CycleGAN method, and the experimental results are shown in Figure 7 and Table 3. We introduce each ablation experiment as follows:
(i) w/o L C o n t e n t : We removed the content loss from UW-CycleGAN, which led to serious detail loss and image blur in the generated images.
(ii) w/o L B l u r : Without the blur-adversarial loss, the generated images remain intact but had slight blurring.
(iii) G R e s N e t : DenseNet block in UW-CycleGAN was replaced by ResNet-block. Although the subjective difference between G R e s N e t and our UW-CycleGAN is not obvious in Figure 7, the objective evaluation results in Table 3 verify the advantages of DenseNet-block.
In the above, the effect of each component in our UW-CycleGAN was verified.

5. Conclusions

Underwater vehicle vision has important research value in underwater applications. We proposed an end-to-end underwater image enhancement method for unpaired data (UW-CycleGAN). To be specific, we firstly added a content loss regularizer to the generator in traditional CycleGAN through a VGG19 pretraining network. Then, a blur-promoting adversarial loss regularizer was adopted in the discriminator. Finally, we replaced the commonly used ResNet-block in CycleGAN with the DenseNet block in the coding layer. Compared with several image enhancement methods, our proposed methods restored the underwater degraded images with blue-green background and blur into clear images effectively. We also performed ablation experiments to verify the importance of each module in UW-CycleGAN.

Author Contributions

Conceptualization, R.D. and S.C.; data curation, R.D., W.L. and C.L.; formal analysis, R.D.; investigation, R.D.; methodology, R.D. and S.C.; project administration, S.C.; resources, S.C. and Y.Z.; software, R.D. and W.L.; supervision, S.C.; validation, R.D., S.C. and Y.Z.; writing—original draft, R.D.; writing—review and editing, R.D., S.C. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

The research project is supported by the National Natural Science Foundation of China under Grant No. 61876144.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, C.; Guo, J.; Guo, C.; Cong, R.; Gong, J. A hybrid method for underwater image correction. Pattern Recognit. Lett. 2017, 94, 62–67. [Google Scholar] [CrossRef]
  2. Wang, Y.; Song, W.; Fortino, G.; Qi, L.; Zhang, W.; Liotta, A. An Experimental-Based Review of Image Enhancement and Image Restoration Methods for Underwater Imaging. IEEE Access 2019, 7, 140233–140251. [Google Scholar] [CrossRef]
  3. Zhang, W.; Dong, L.; Pan, X.; Zou, P.; Qin, L.; Xu, W. A Survey of Restoration and Enhancement for Underwater Images. IEEE Access 2019, 7, 182259–182279. [Google Scholar] [CrossRef]
  4. Schettini, R.; Corchs, S. Underwater Image Processing: State of the Art of Restoration and Image Enhancement Methods. EURASIP J. Adv. Signal Process. 2010, 2010, 746052. [Google Scholar] [CrossRef] [Green Version]
  5. Weng, C.C.; Chen, H.; Fuh, C.S. A novel automatic white balance method for digital still cameras. In Proceedings of the IEEE International Symposium on Circuits and Systems, Kobe, Japan, 23–26 May 2005. [Google Scholar]
  6. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; Romeny, B.T.H.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  7. Hitam, M.S.; Awalludin, E.A.; Yussof, W.N.J.H.W.; Bachok, Z. Mixture contrast limited adaptive histogram equalization for underwater image enhancement. In Proceedings of the International Conference on Computer Applications Technology, Sousse, Tunisia, 20–22 January 2013. [Google Scholar]
  8. Cheng, C.; Sung, C.; Chang, H. Underwater image restoration by red-dark channel prior and point spread function deconvolution. In Proceedings of the IEEE International Conference on Signal and Image Processing Applications, Kuala Lumpur, Malaysia, 19–21 October 2015. [Google Scholar]
  9. Sun, F.; Zhang, X.; Wang, G. An Approach for Underwater Image Denoising Via Wavelet Decomposition and High-pass Filter. In Proceedings of the International Conference on Intelligent Computation Technology and Automation, Shenzhen, China, 28–29 March 2011. [Google Scholar]
  10. Shahrizan, A.; Ghani, A. Image contrast enhancement using an integration of recursive-overlapped contrast limited adaptive histogram specification and dual-image wavelet fusion for the high visibility of deep underwater image. Ocean Eng. 2018, 162, 224–238. [Google Scholar]
  11. Khan, A.; Ali, S.S.A.; Malik, A.S.; Anwer, A.; Meriaudeau, F. Underwater image enhancement by wavelet based fusion. In Proceedings of the IEEE International Conference on Underwater System Technology: Theory and Applications, Penang, Malaysia, 13–14 December 2016. [Google Scholar]
  12. Sun, J.; Wang, W. Study on Underwater Image Denoising Algorithm Based on Wavelet Transform. J. Phys. Conf. Ser. 2017, 806, 1–10. [Google Scholar]
  13. Vasamsetti, S.; Mittal, N.; Neelapu, B.C.; Sardana, H.K. Wavelet based perspective on variational enhancement technique for underwater imagery. Ocean Eng. 2017, 141, 88–100. [Google Scholar] [CrossRef]
  14. Mukherjee, S.; Valenzise, G.; Cheng, I. Potential of deep features for opinion-unaware, distortion-unaware, no-reference image quality assessment. In Smart Multimedia. ICSM 2019; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2020; pp. 87–95. [Google Scholar]
  15. Wang, Y.; Zhang, J.; Cao, Y.; Wang, Z. A deep CNN method for underwater image enhancement. In Proceedings of the International Conference on Image Processing, Beijing, China, 17–20 September 2017. [Google Scholar]
  16. Anwar, S.; Li, C.; Porikli, F. Deep Underwater Image Enhancement. arXiv 2018, arXiv:1807.03528. [Google Scholar]
  17. Hou, M.; Liu, R.; Fan, X.; Luo, Z. Joint Residual Learning for Underwater Image Enhancement. In Proceedings of the International Conference on Image Processing, Athens, Greece, 7–10 October 2018. [Google Scholar]
  18. Sun, X.; Liu, L.; Li, Q.; Dong, J.; Lima, E.; Yin, R. Deep Pixel to Pixel Network for Underwater Image Enhancement and Restoration. IET Image Process. 2018, 13, 469–474. [Google Scholar] [CrossRef]
  19. Uplavikar, P.; Wu, Z.; Wang, Z. All-In-One Underwater Image Enhancement using Domain-Adversarial Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshop, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  20. Fabbri, C.; Islam, M.J.; Sattar, J. Enhancing Underwater Imagery Using Generative Adversarial Networks. In Proceedings of the IEEE International Conference on Robotics and Automation, Brisbane, Australia, 21–25 May 2018. [Google Scholar]
  21. Yang, M.; Hu, K.; Du, Y.; Wei, Z.; Sheng, Z.; Hu, J. Underwater image enhancement based on conditional generative adversarial network. Signal Process. Image Commun. 2020, 81, 115723. [Google Scholar] [CrossRef]
  22. Li, H.; Li, J.; Wang, W. A Fusion Adversarial Underwater Image Enhancement Network with a Public Test Dataset. Comput. Sci. 2019, 95. [Google Scholar] [CrossRef]
  23. Hu, K.; Zhang, Y.; Weng, C.; Wang, P.; Deng, Z.; Liu, Y. An Underwater Image Enhancement Algorithm Based on Generative Adversarial Network and Natural Image Quality Evaluation Index. J. Mar. Sci. Eng. 2021, 9, 691. [Google Scholar] [CrossRef]
  24. Chen, Y.; Lai, Y.; Liu, Y. CartoonGAN: Generative Adversarial Networks for Photo Cartoonization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  25. Chiang, J.Y.; Chen, Y. Underwater Image Enhancement by Wavelength Compensation and Dehazing. IEEE Trans. Image Process. 2012, 21, 1756–1769. [Google Scholar] [CrossRef] [PubMed]
  26. Zhu, J.; Park, T.; Isola, P.; Efros, A. Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  27. Huang, G.; Liu, Z.; Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  28. Liu, R.; Fan, X.; Zhu, M.; Hou, M.; Luo, Z. Real-world Underwater Enhancement: Challenges, Benchmarks, and Solutions under Natural Light. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 4861–4875. [Google Scholar] [CrossRef]
  29. Hautiere, N.; Tarel, J.P.; Aubert, D.; Dumont, E. Blind contrast enhancement assessment by gradient ratioing at visible edges. Image Anal. Stereol. 2008, 27, 87–95. [Google Scholar] [CrossRef]
  30. Gadre, S.R. Information entropy and thomas-Fermi theory. Phys. Rev. A 1984, 30, 620–621. [Google Scholar] [CrossRef]
  31. Panetta, K.; Gao, C.; Agaian, S. Human-visual-systeminspired Underwater Image Quality Measures. IEEE J. Ocean. Eng. 2016, 41, 541–551. [Google Scholar] [CrossRef]
  32. Pan, P.; Yuan, F.; Cheng, E. Underwater Image De-scattering and Enhancing using Dehazenet and HWD. J. Mar. Sci. Technol. 2018, 26, 531–540. [Google Scholar]
  33. Berman, D.; Treibitz, T.; Avidan, S. Diving into Haze-Lines: Color Restoration of Underwater Images. In Proceedings of the British Machine Vision Conference, London, UK, 4–7 September 2017. [Google Scholar]
  34. Iqbal, K.; Odetayo, M.; James, A.; Salam, R.A.; Talib, A.Z.H. Enhancing the low quality images using Unsupervised Colour Correction Method. In Proceedings of the 2010 IEEE International Conference on Systems, Man and Cybernetics, Istanbul, Turkey, 10–13 October 2010. [Google Scholar]
  35. Islam, M.J.; Xia, Y.; Sattar, J. Fast Underwater Image Enhancement for Improved Visual Perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The Unpaired underwater images dataset. The top row displays the degraded underwater images, and the bottom row shows the clear ones. Both rows are unpaired.
Figure 1. The Unpaired underwater images dataset. The top row displays the degraded underwater images, and the bottom row shows the clear ones. Both rows are unpaired.
Information 13 00001 g001
Figure 2. A brief illustration of UW-CycleGAN, which consists of two generators and two discriminators. The generator G (or F) is a downsample–upsample framework with DenseNet blocks. The discriminator D Y (or D X ) comprises five full convolution layers.
Figure 2. A brief illustration of UW-CycleGAN, which consists of two generators and two discriminators. The generator G (or F) is a downsample–upsample framework with DenseNet blocks. The discriminator D Y (or D X ) comprises five full convolution layers.
Information 13 00001 g002
Figure 3. (a) input image x, (b) generated image G ( x ) , and (c) reconstructed image F ( G ( x ) ) .
Figure 3. (a) input image x, (b) generated image G ( x ) , and (c) reconstructed image F ( G ( x ) ) .
Information 13 00001 g003
Figure 4. Architecture of the generator and DenseNet blocks networks. (a) Generator network structure and (b) DenseNet block, where “Conv” denotes the convolution layer and “Deconv” denotes the deconvolution layer.
Figure 4. Architecture of the generator and DenseNet blocks networks. (a) Generator network structure and (b) DenseNet block, where “Conv” denotes the convolution layer and “Deconv” denotes the deconvolution layer.
Information 13 00001 g004
Figure 5. Visual quality comparisons of the enhanced underwater images on the URPC dataset.
Figure 5. Visual quality comparisons of the enhanced underwater images on the URPC dataset.
Information 13 00001 g005
Figure 6. Visual quality comparisons of the enhanced underwater images on the EUVP dataset.
Figure 6. Visual quality comparisons of the enhanced underwater images on the EUVP dataset.
Information 13 00001 g006
Figure 7. Ablation experiments: (a) input image, (b) w/o L C o n t e n t , (c) w/o L B l u r , (d) G R e s N e t , and (e) UW-CycleGAN.
Figure 7. Ablation experiments: (a) input image, (b) w/o L C o n t e n t , (c) w/o L B l u r , (d) G R e s N e t , and (e) UW-CycleGAN.
Information 13 00001 g007
Table 1. Quality evaluation of all methods on the URPC dataset.
Table 1. Quality evaluation of all methods on the URPC dataset.
 AG ↑IE ↓UIQM ↑
Deunderwater7.50477.81785.1460
HL7.30217.40334.0719
UCM4.91027.39553.8221
FUnIE-GAN-UP5.94447.38194.2130
CartoonGAN4.90797.25674.4997
CycleGAN6.47377.27854.8380
UW-CycleGAN7.63457.18245.1689
Table 2. Quality evaluation of all methods on the EUVP dataset.
Table 2. Quality evaluation of all methods on the EUVP dataset.
 AG ↑IE ↓UIQM ↑
Deunderwater2.49457.78301.5500
HL2.05657.42711.5769
UCM2.54897.24512.0124
FUnIE-GAN-UP2.60147.34630.9782
CartoonGAN2.72246.74222.0883
CycleGAN2.99696.74522.6075
UW-CycleGAN3.13706.48272.7497
Table 3. Ablation Experiments: Quality evaluation of all methods on URPC dataset.
Table 3. Ablation Experiments: Quality evaluation of all methods on URPC dataset.
 AG ↑IE ↓UIQM ↑
w/o L C o n t e n t 7.35677.37275.1268
w/o L B l u r 7.54907.27505.1530
G R e s N e t 6.99847.28645.0271
UW-CycleGAN7.63457.18245.1689
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Du, R.; Li, W.; Chen, S.; Li, C.; Zhang, Y. Unpaired Underwater Image Enhancement Based on CycleGAN. Information 2022, 13, 1. https://doi.org/10.3390/info13010001

AMA Style

Du R, Li W, Chen S, Li C, Zhang Y. Unpaired Underwater Image Enhancement Based on CycleGAN. Information. 2022; 13(1):1. https://doi.org/10.3390/info13010001

Chicago/Turabian Style

Du, Rong, Weiwei Li, Shudong Chen, Congying Li, and Yong Zhang. 2022. "Unpaired Underwater Image Enhancement Based on CycleGAN" Information 13, no. 1: 1. https://doi.org/10.3390/info13010001

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop