Next Article in Journal
Viscous Flow Fields Induced by the Run-Up of Periodic Waves on Vertical and Sloping Seawalls
Next Article in Special Issue
High Speed and Precision Underwater Biological Detection Based on the Improved YOLOV4-Tiny Algorithm
Previous Article in Journal
Numerical Simulation of Scalar Mixing and Transport through a Fishing Net Panel
Previous Article in Special Issue
Identification of Orbital Angular Momentum by Support Vector Machine in Ocean Turbulence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Image Enhancement Based on Color Correction and Detail Enhancement

1
School of Information and Control Engineering, Qingdao University of Technology, Qingdao 266520, China
2
Department of Computer Science and Technology, Qingdao University, Qingdao 266071, China
3
Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford GU2 7XH, UK
*
Authors to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2022, 10(10), 1513; https://doi.org/10.3390/jmse10101513
Submission received: 25 September 2022 / Revised: 7 October 2022 / Accepted: 11 October 2022 / Published: 17 October 2022
(This article belongs to the Special Issue Underwater Engineering and Image Processing)

Abstract

:
To solve the problems of underwater image color deviation, low contrast, and blurred details, an algorithm based on color correction and detail enhancement is proposed. First, the improved nonlocal means denoising algorithm is used to denoise the underwater image. The combination of Gaussian weighted spatial distance and Gaussian weighted Euclidean distance is used as the index of nonlocal means denoising algorithm to measure the similarity of structural blocks. The improved algorithm can retain more edge features and texture information while maintaining noise reduction ability. Then, the improved U-Net is used for color correction. Introducing residual structure and attention mechanism into U-Net can effectively enhance feature extraction ability and prevent network degradation. Finally, a sharpening algorithm based on maximum a posteriori is proposed to enhance the image after color correction, which can increase the detailed information of the image without expanding the noise. The experimental results show that the proposed algorithm has a remarkable effect on underwater image enhancement.

1. Introduction

Projects such as underwater biometrics [1], garbage collection [2], and resource exploration [3] all require clear images, but capturing a clear scene in the underwater world is not a simple task. There are a large number of media that can absorb light underwater; light will experience different degrees of attenuation when propagating underwater. Underwater images usually have problems such as color deviation, low contrast, and blurred details, which affect the subsequent research work on images. Therefore, it is necessary to study underwater image enhancement technology.
The current popular underwater image enhancement algorithms are divided into physical model algorithms and non-physical model algorithms. The algorithm based on the physical model mathematically models the degradation process of the underwater image and obtains a clear image by estimating the model parameters. Common physical model algorithms include the following. In 2010, Chao et al. [4] used the dark channel prior (DCP) to derive the background light and transmittance of the image and then restored the underwater image based on a physical model. In 2012, Chiang et al. [5] proposed a restoration method by compensating for the light absorbed by the underwater medium. Because there are a large number of media that can absorb light underwater according to the loss process of light underwater, a clear image is inverted by compensating light to achieve the effect of the restoration. In 2015, Galdran et al. [6] proposed an automatic red channel underwater image restoration method, which is an algorithm that combines the image defogging model with the attenuation rate of light in water.
The non-physical model algorithm does not consider the imaging process and model and improves the quality of the underwater image by image processing. Common algorithms based on non-physical models are: In 2007, Iqbal et al. [7] proposed a sliding stretch method based on color space. The color correction is performed by histogram equalization [8] of the underwater image in RGB color space, and the contrast is adjusted by a stretching operation in the HIS color space. In 2012, Ancuti et al. [9] combined algorithms such as contrast enhancement, white balance, and extraction of regions of interest. These algorithms are used to enhance the underwater images, respectively, and then the processed images are fused according to a certain proportion. In 2013, Drews et al. [10] proposed the dark channel prior (UDCP) algorithm. The algorithm restores the underwater image by adjusting the blue and green channels of the image. In 2018, Peng et al. [11] used depth-dependent color change, scene light differentiation, and adaptive color correction to restore underwater images.
In recent years, deep learning [12] has been widely used in various fields and achieved great success. In 2018, Li et al. [13] applied the generative adversarial network [14] (GAN) to the problem of underwater image enhancement and generated a large number of datasets of underwater images and atmospheric images. In 2019, Li et al. [15] designed Water-Net, a neural network that enhances underwater images. In 2019, Wang et al. [16] proposed the UWGAN algorithm. GAN is used to generate a large number of datasets of underwater images and atmospheric images, and the medical image segmentation network U-Net [17] is used to train the model of mapping from underwater images to atmospheric images.
Many algorithms nowadays often lose the detailed information of the image in the process of color correction, and it is difficult to solve the problems of color deviation, low contrast, and detail loss of underwater images at the same time. To solve this problem, an algorithm based on color correction and detail enhancement is proposed. The algorithm transforms the underwater image enhancement problem into three problems: underwater image denoising, color correction, and detail enhancement, and designs algorithms for each problem. Experimental results show that the algorithm has a good effect on color correction and detail enhancement.
The innovations of this study are as follows:
(1) According to the underwater image noise problem, some improvements have been made to the nonlocal means denoising [17] (NL-means). The Gaussian weighted spatial distance and the Gaussian weighted Euclidean distance are combined as the index of NL-means to measure the similarity of structural blocks. The improved NL-means can preserve the texture features and edge information of underwater targets while maintaining the denoising function of the original algorithm.
(2) We use U-Net [18] to correct the color deviation of underwater images. Some improvements have been made to U-Net for underwater image problems. Introducing residual structure [19] and attention mechanism [20] into U-Net can effectively enhance feature extraction ability and prevent network degradation.
(3) According to the underwater image degradation model, an underwater image sharpening algorithm based on maximum a posteriori (MAP) is proposed. The algorithm can increase the detailed information of the image without expanding the noise.

2. Materials and Methods

This section studies the basic principle of light propagation in water and image enhancement from three aspects: image denoising, color correction, and image sharpening.

2.1. Underwater Image Imaging Model

The imaging of underwater images in the camera can be divided into direct component, backscattering component, and forward-scattering component [21].
(1) The direct component is the part where the light reaches the camera after being reflected by the target object in the process of underwater propagation. Its expression is:
E d c x , y = E c x , y exp a c d x , y
where x , y are the coordinates of image pixels, c is the image of red, green, and blue three color channels, E c x , y is the reflected light of the target, a c is the attenuation coefficient, and d x , y is the distance between camera and target.
(2) The backscattering component is the part of light reaching the camera after scattering by the medium in water. Its expression is:
E f c x , y = g c x , y E d c x , y
where g c x , y is the point-spread function, and ⊗ is the convolution operations.
(3) The forward-scattering component is the part where the light encounters the target object, is reflected by the target object, and then is scattered by the medium to reach the camera. The expression is:
E b c x , y = B c 1 exp a c d x , y
where B c is the backlight.

2.2. Underwater Image Denoising

Due to the particularity of underwater impurities and illumination conditions, underwater images usually contain noise. NL-means can remove the noise of underwater images, but also reduce the texture features and edge information of images. The proposed algorithm makes some improvements to NL-means and combines Gaussian weighted spatial distance with Gaussian weighted Euclidean distance as a new index to measure the similarity of structural blocks. The improved NL-means can retain the texture features and edge information of underwater targets while maintaining the denoising function of the original algorithm.

2.2.1. NL-Means

NL-means is a good denoising algorithm. First, the entire image is divided into several regions. Then, it finds a region similar to it for each region. Finally, the average of these similar regions is used to replace the pixel values of these regions. The expression of NL-means is:
x i , j = k , g Ω i , j ω k , g , i , j y k , g k , g Ω i , j ω k , g , i , j
where Ω i , j is the local region centered on ( i , j ) , and the weight value ω k , g , i , j is the distance between the image structure blocks at the exponential function mapping ( k , g ) and ( i , j ) , which is expressed as:
ω k , g , i , j = exp d y k , g , y i , j h 2 r 2
where y i , j is a vector centered on ( i , j ) , h 2 is the filter intensity coefficient, and d y k , g , y i , j is the similarity measure of structural blocks, which is represented by Euclidean distance:
d y k , g , y i , j = y k , g y i , j 2 2

2.2.2. Improved NL-Means

The weighted kernel function of the NL-means algorithm can make the regions with high similarity obtain larger weights and the regions with low similarity obtain smaller weights. The ideal kernel function can output larger weights when the similar neighborhood distance is small, and the output decreases rapidly as the distance increases. Since the Gaussian kernel function increases less weight when the Euclidean distance is small, the image noise reduction performance will be reduced due to the frequent change of signal strength when it is used alone, which will affect the subsequent U-Net to extract the feature information in the image. Inspired by the related noise reduction algorithm (Qiuyu Song, 2022) [22], the Gaussian weighted spatial distance and the Gaussian weighted Euclidean distance are combined as a new index to measure the similarity of structural blocks in the image, which effectively solves the problem of decreasing noise reduction ability.
The Gaussian weighted spatial distance can be expressed as:
d s y k , g , y i , j = y k , g y i , j 2
The Gaussian weighted Euclidean distance can be expressed as:
d y k , g , y i , j = y k , g y i , j 2 2
Gaussian weighted spatial distance is combined with Gaussian weighted Euclidean distance as a new index to measure the similarity of structural blocks:
ω k , g , i , j = exp y k , g y i , j 2 2 × y k , g y i , j 2 h 2 r 2

2.3. Underwater Image Color Correction

To solve the problem of color deviation of underwater images, an improved U-Net is proposed to fully extract the features of underwater images and adaptively learn the mapping relationship between underwater images and atmospheric images. The improved U-Net can increase the feature extraction ability and improve the network accuracy.

2.3.1. U-Net

U-Net is an end-to-end network proposed by OlafRonneberger et al. [17] in 2015, which is mainly used for the semantic segmentation of medical images. U-Net is based on FCN [23] and only contains the convolution layer and pooling layer. The network has a fully symmetrical U-shaped structure with the same number of encoders and decoders on both sides. The model undergoes four down-samplings, and each coding layer contains two convolution operations to extract information from the image. Then down-sampling is performed by maximum pooling. The model also undergoes four up-sampling processes. First, concat the feature maps of the same depth using a jump connection. Then, the features in the image are extracted after two convolution operations. Finally, the features are passed up through up-sampling. Through the U-shaped structure, the down-sampling process extracts features, and the up-sampling process transmits features upwards so that the network extracts features with higher accuracy and better effect.

2.3.2. Improved U-Net

As shown in Figure 1, the proposed underwater image color correction algorithm is based on U-Net and makes the following improvements to the network:
(1) Generally, deeper features can be extracted as the network deepens. However, due to the problem of network degradation, the effect of a deep network on feature extraction may not be as good as that of the shallow network. Changing the two convolution structures of each layer in the U-Net network to a residual structure can effectively prevent network degradation during color correction and ensure the ability of the network to extract features.
(2) Since U-Net has invalid semantic information when jumping connections, CBAM is added to the feature map of each layer in the encoder after passing through the convolutional layer. The feature map recovers the features of the image during up-sampling by CBAM. CBAM has a good resource allocation function, which can allocate more resources to more important features under the condition of limited resources, suppress other invalid features, and improve network accuracy.

2.3.3. CBAM

CBAM is a very efficient attention mechanism including a channel attention module (CAM) and spatial attention module (SAM). As shown in Figure 2, the feature map goes through CAM and then SAM. CAM improves attention to the target category, and SAM improves attention to the target location.
As shown in Figure 3, in CAM, the input features are first processed by maximum pooling and average pooling. In the backpropagation process, the maximum pooling is used to obtain the maximum value of each neighborhood, ignoring the influence of the non-maximum value. Average pooling calculates the average value of each neighborhood so that each pixel is involved in the calculation. Then the results are input into the shared fully connected layer to compress the spatial dimension of the features, and the obtained results are added at the pixel level. Then, the channel attention weight is obtained by the Sigmoid activation function. Finally, the channel attention weight is a matrix multiplied by the original image to improve the attention to the target category.
As shown in Figure 4, in SAM, the input features are first processed by maximum pooling and average pooling. In the backpropagation process, the maximum pooling is used to obtain the maximum value of each neighborhood, ignoring the influence of the non-maximum value. Average pooling calculates the average value of each neighborhood so that each pixel is involved in the calculation. Then, concat the results of these two outputs at the channel level. Then, the result is subjected to a convolution operation to reduce the number of channels of the feature map. Next, the spatial attention weight is obtained by the Sigmoid activation function. Finally, the channel attention weight is a matrix multiplied by the original image to improve the attention to the target position.

2.4. Underwater Image Detail Enhancement

To solve the problem of lack of details in underwater images, this paper transforms the image detail enhancement process into the maximum likelihood problem of underwater image deblurring probability model. The blurred kernel and the clear image are alternately updated by the MAP method to make the clear image closer to the real image.

2.4.1. Image Degradation Model

The image degradation model [24] is a model of the blurring process of the image underwater. The formula is as follows:
g ( x , y ) = k ( x , y ) f ( x , y ) + n ( x , y )
where f ( x , y ) is the input image, k ( x , y ) is a degradation function, n ( x , y ) is the random noise, and g ( x , y ) is the blurred image.
If the noise interference is ignored, the model can be simplified as:
g ( x , y ) = k ( x , y ) f ( x , y )
The usual solution to the image deblurring problem is: First, build a probabilistic model of image deblurring. Then, according to this model, the conditional probability density of the blurred image concerning the clear image and the degradation function is obtained. Finally, by maximizing this conditional probability, the most suitable clear image and degradation function are found. However, due to the particularity of underwater impurities and illumination conditions, underwater images usually contain a large amount of noise. While maximizing this conditional probability, the noise will also be amplified. This paper uses the MAP method to solve this problem. According to MAP, the image blur model can be expressed as:
min f , k F ( g ; k , f ) + α ρ f ( f ) + β ρ k ( k )
where F ( ) is the negative logarithm of the conditional probability density, representing the data item, ρ f and ρ k are priors for clear images and blur kernels, representing the regular term, and α and β are the weight values of the regular term.

2.4.2. The Process of Sharpening Algorithm

As shown in Figure 5, the proposed underwater image sharpening algorithm flow mainly includes the blur kernel estimation part and the clear image estimation part. The update of the blur kernel and the clear image is an alternating process.
When updating the blur kernel, the current clear image is first blurred. Then, the adaptive threshold method is used to increase the salient region information of the image. Finally, the blur kernel is estimated according to our previously designed deblurring model.
When updating the clear image, the regularization term optimization model is first added to the deblurring model to increase the generalization ability of the model. Then, the bilinear interpolation method is used to up-sample the clear image we estimate. Finally, the obtained clear image is used as the input to update the blur kernel, and the clear image and the blur kernel are continuously updated to make the clear image closer to the real image.

2.4.3. Saliency Region Extraction

The salient region refers to the contour information in the image. According to [25], effective extraction of salient regions of clear images can improve the accuracy of estimating blur kernels. It is a difficult task to extract the salient region of the image. This paper extracts the salient region from four aspects: brightness, color, direction, and edge [26].
Assuming that the three color channels of the input image are r, g, and b, the brightness feature can be expressed as:
I = ( r + g + b ) / 3
For color features, the saliency method of frequency domain harmonic [27] is used to obtain the color feature map of the image. First, we change the color to a uniform CIELab color space. Then, the transformed image is processed by Gaussian filtering. Finally, the square of the difference between the original image and the filtered image is used as the saliency map of the color. The formula is as follows:
C ( x , y ) = I u I w h c 2
where I u is the transformed image, and I w h c is the filtered image.
For the directional features, Gabor filters [28] in four directions of 0°, 45°, 90°, and 135°are used to filter the grayscale images, respectively, and the feature maps in four directions are obtained. The expression is:
sample x 0 , y 0 ; θ , φ = I ( x , y ) Gabor x x 0 , y y 0 ; θ , φ d x d y
where X 0 and y 0 are the center coordinates of the receptive field, and I ( x , y ) is the input image.
For the edge features of the image, Canny edge detection [29] is used. First, Gaussian filtering is used to smooth the image and eliminate the noise in the image. Then, the gradient of the image is calculated to find possible edges. Then, non-maximum suppression is used to eliminate the edge of false detection. Finally, the double threshold method is used to filter out the edge information in the image.

2.4.4. The Estimation of Fuzzy Kernel

In the estimation process of fuzzy kernel to prevent nonconvex optimization problems and improve the estimation speed of the fuzzy kernel, this paper uses regularization terms to optimize the deblurring model. The expression is as follows:
min k S k g 2 2 + ρ k 2 2
where ρ is used to balance the relative strength between data items and regular items.
By using the least square method, we can obtain:
k = F 1 F ¯ x S F x g + F ¯ y S F y g F x S 2 + F y S 2 + ρ
where F ( ) is the fast Fourier, F 1 ( ) is the inverse transform, and F ¯ ( ) is the conjugate operation.

2.4.5. The Estimation of Clear Images

When estimating the clear image, we use the normalized prior term as the regularization term to optimize the equation. The formula is as follows:
min f μ 2 f k g 2 2 + f 1 f 2
where f is the gradient value of the clear image, and μ is the weight value of the data item.

3. Results and Discussion

The experiment was completed in the Python3.6 environment. The CPU is Intel (R) Core (TM) i5-6300 HQ CPU 8.00 GB RAM, using GPU acceleration, GPU is GTX1080, and the deep learning framework is PyTorch. A total of 2000 pairs of underwater image training networks were selected from the EUVP dataset. It covers images under different underwater scenes and lighting conditions. Adam [30] was used as the optimization algorithm during the training process, the batch size was set to 4, the initial learning rate was set to 0.001, and the learning rate was divided by 10 for every 30 epochs, eventually training 200 epochs.
To verify the effectiveness of the proposed algorithm, some representative images in the EUVP dataset were selected to compare the proposed algorithm with some classical algorithms in terms of color correction, detail enhancement, and image quality. These algorithms included IBLA [31], UDCP [10], ULAP [32], RGHS [33], Sea-thru [34], UWGAN [18], and FunieGAN [35].

3.1. Color Correction Experiment

To verify the effectiveness of the proposed algorithm for color correction, some images of blue and green scenes on the EUVP dataset were selected for color correction using the proposed algorithm and the classical algorithm.
As shown in Figure 6, (a), (b), (e), (f), (g) and (i) show the comparison results of images in the blues scene, and (c), (d), (h) and (j) show the comparison results of images in the green scene. It can be seen from the comparison results that the IBLA algorithm had a better enhancement effect for the image in the blue scene, and had a certain defogging effect for the image in the green scene, which can increase the contrast of the image, but at the same time cause the loss of image details. The UDCP algorithm can solve the fog blur problem of underwater images to a certain extent, but the processed images were darker overall, the contrast was low, and the effect was not ideal. The ULAP algorithm had a good enhancement effect on the image in the blue scene. The image processing in the green scene was reddish, and there was a certain color deviation from the real image. The RGHS algorithm had a certain solution to the problem of underwater image fog blur, but the overall effect was not obvious. The Sea-thru algorithm had a good effect on image enhancement in green scenes. For images in blue scenes, the details of the image can be increased, but the overall image color was blue, and there was a certain color deviation from the real image. The UWGAN algorithm had a good defogging effect for low contrast images while improving image contrast and increasing image details, but for high contrast images, the effect was not good. The FunieGAN algorithm had a certain effect of defogging and increasing image details, but the effect was not obvious. The algorithm of this study can effectively solve the fog blur problem of underwater images for images in blue and green scenes, and the image color was closer to the true value image. The results show that the proposed algorithm has a good ability to correct image color deviation for most underwater scene images.

3.2. Detail Enhancement Experiment

To verify that the proposed algorithm has the effect of enhancing image details, three images in the EUVP dataset were selected to compare the number of visible edges between the original image and the enhanced image. As shown in Figure 7, (a) is the edge detection of the original image and (b) is the edge detection of the enhanced image. The edges of the image contain a lot of information, and the number of edges in (b) is much higher than that in (a), which means that the proposed algorithm enhances more image details. Therefore, the proposed algorithm has the effect of increasing image details.
To verify the advantages of the proposed algorithm in detail enhancement, the visible edge growth rate of the image was used as the evaluation index to compare the proposed algorithm with the classical algorithm. Five underwater images were selected in the EUVP dataset, and the value of the visible edge growth rate e was calculated and compared. The visible edges contain a lot of detail information. The more the number of visible edge restoration means, the better the image enhancement effect. As shown in Table 1, the e values of IBLA and RGHS are lower, and the increase in the number of edges is not obvious. The e value of the proposed algorithm is the highest, which means that it is superior to other algorithms in detail enhancement. Therefore, the proposed algorithm can better restore the details of underwater images.

3.3. Image Quality Assessment

To verify the quality of the enhanced image, the peak signal-to-noise ratio (PSNR), structural similarity (SSIM), underwater image quality assessment metrics (UCIQE), and underwater image quality metrics (UIQM) [36] were used to quantitatively evaluate the measurement of the above ten images. PSNR is used to judge the degree of distortion between the enhanced image and the true value image. The greater the PSNR, the better the enhancement effect of the algorithm on the image. SSIM is an index of comprehensive contrast, brightness, and structural similarity. The larger the SSIM value, the more similar the enhanced image is to the true value image, which indicates that the effect of the algorithm is better. UCIQE is an index to judge the overall quality of the image. The larger the value, the better the effect of the algorithm. UIQM is an indicator of the color, clarity, and contrast of the integrated image. The larger the value, the better the effect of the algorithm.
Table 2 shows the scores of SSIM, PSNR, UIQM, and UCIQE of various algorithms on the EUVP dataset. The UWGAN algorithm has the highest SSIM value and the best enhancement effect on the structural similarity of the image, but the UIQM value is very low, and the color and clarity are poor. The UCIQE value of the UDCP algorithm is the highest, and the overall image quality is high. However, the PSNR value is very low, and the image distortion is large. The PSNR and UIQM values of the proposed algorithm are higher than other algorithms, the distortion degree of the image is the smallest, and the enhanced image is closer to the original image. The algorithm proposed by SSIM and UCIQE values is not the highest, but it is superior to most algorithms. The results show that the overall quality of the image processed by the proposed algorithm is good, close to the real image, and in line with human visual senses.

3.4. Running Time Experiment

To verify the real-time performance of the algorithm, the running time of the proposed algorithm was compared with that of the classical algorithm. The experiment used the above ten images with a pixel size of 256 × 256 . The running environment of the computer was Intel (R) Core (TM) i5-6300HQ CPU 8.00 GB RAM. All algorithms were tested on the same computer. IBLA, UDCP, ULAP, RGHS, and Sea-thru are traditional enhancement algorithms, and UWGAN and FunieGAN are deep learning algorithms. As shown in Table 3, the ULAP algorithm ran the fastest. The proposed algorithm is second only to ULAP and RGHS algorithms and is faster than other algorithms. The results show that the proposed algorithm has a faster running speed and can meet the real-time requirements of underwater image enhancement.

3.5. Validation of Algorithm Effectiveness

To verify the effectiveness of adding an attention mechanism and a residual module in the U-Net, U-Net and improved U-Net were trained using the EUVP dataset. Color correction was performed on the above 10 images using the trained results. SSIM, PSNR, UIQM, and UCIQE were used as evaluation indexes of correction effect.
As shown in Table 4, (a) is the correction effect of U-Net, (b) is the correction effect after adding the residual structure in U-Net, (c) is the correction effect of adding the attention mechanism in U-Net, and (d) is the correction effect of increasing both the attention mechanism and the residual structure. It can be seen that adding residual structure and attention mechanism in U-Net can effectively improve the SSIM, PSNR, UIQM, and UCIQE of images. Increasing the attention mechanism in U-Net works better than increasing the residual structure. Therefore, adding a residual structure and an attention mechanism in U-Net has a certain effect on the color correction of underwater images.

3.6. Ablation Study

To understand the role of each component in the proposed algorithm, an ablation study was performed using the above ten images. SSIM, PSNR, UIQM, and UCIQE were used to measure the effect of each experiment. As shown in Table 5, (a) only the improved NL-means was used to denoise the underwater image, (b) only the improved U-Net was used to correct the color of the underwater image, (c) only the proposed sharpening algorithm was used to process the image, (d) the improved NL-means was used to denoise the underwater image and the improved U-Net was used to correct the color of the underwater image, and (e) the complete algorithm was used. The following conclusions can be drawn:
(1) Compared with the individual experimental results of each component, the complete model enhances the image best, which means that using these three components together is effective.
(2) The three components of the proposed algorithm have the effect of improving image performance. The improved U-Net has the greatest effect on improving SSIM and PSNR. The proposed sharpening algorithm has the greatest effect on improving UIQM and UCIQE.

3.7. Application Test

The proposed underwater image enhancement technology can be applied to underwater key point detection and image segmentation. To verify the effectiveness of the proposed algorithm, the underwater key point detection and image segmentation of the original image and the enhanced image were compared.
Underwater key point detection is an important technology for underwater image detection and recognition. To verify that the proposed algorithm has an enhanced effect on key point detection, SIFT key point detection [37] was performed on the original image and the enhanced image of the proposed algorithm, and the number of key points were compared. As shown in Figure 8, the number of key points of the enhanced image is much more than that of the original image. The results show that the proposed algorithm has a good effect on enhancing the details of underwater images, which is conducive to the detection and recognition of underwater images.
Underwater image segmentation divides the image into different regions according to the characteristics of the image. The fast FCM clustering algorithm [38] was used to segment the original image and the enhanced image. As shown in Figure 9, (a) is the segmentation effect of the original image, and (b) is the segmentation effect of the enhanced image. Using this algorithm to enhance the image, the segmentation effect is more accurate, especially for the segmentation of foreground and background. The results show that the proposed algorithm provides a more accurate segmentation effect for underwater image segmentation.

4. Conclusions

In this work, a new enhancement algorithm was proposed. The algorithm enhances underwater images mainly from three aspects: image denoising, color correction, and detail enhancement. First, some improvements were made to the NL-means denoising algorithm. The improved NL-means can preserve edge and texture information while denoising. Then, the improved U-Net was used to correct the color of underwater images. Introducing residual structure and attention mechanism into U-Net can effectively enhance feature extraction ability and prevent network degradation. Finally, aiming at the problem of the lack of underwater image details, an image sharpening algorithm based on MAP was designed. The algorithm can increase image detail information without expanding noise.
The proposed algorithm was compared with some classical algorithms. The results show that the algorithm has a significant effect on the enhancement of underwater images, which provides help for future research of underwater image enhancement.

Author Contributions

Z.W. and Y.J. contributed to writing the original draft, revising and editing the manuscript, data collection, data analysis, statistics, and data interpretation; L.S. contributed to revising and editing the manuscript; J.S. contributed to conceptualization of the study, data interpretation, and revising and editing the manuscript. All authors approved the submitted version of the manuscript.

Funding

This work was supported by National Natural Science Foundation of China (No. 61501278).

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Knausgård, K.M.; Wiklund, A.; Sørdalen, T.K.; Halvorsen, K.T.; Kleiven, A.R.; Jiao, L.; Goodwin, M. Temperate fish detection and classification: A deep learning based approach. Appl. Intell. 2022, 52, 6988–7001. [Google Scholar] [CrossRef]
  2. Xue, B.; Huang, B.; Wei, W.; Chen, G.; Li, H.; Zhao, N.; Zhang, H. An Efficient Deep-Sea Debris Detection Method Using Deep Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 12348–12360. [Google Scholar] [CrossRef]
  3. Bailey, G.N.; Flemming, N.C. Archaeology of the continental shelf: Marine resources, submerged landscapes and underwater archaeology. Quat. Sci. Rev. 2008, 27, 2153–2165. [Google Scholar] [CrossRef] [Green Version]
  4. Chao, L.; Wang, M. Removal of water scattering. In Proceedings of the 2010 2nd International Conference on Computer Engineering and Technology, Kuala Lumpur, Malaysia, 7–10 May 2010; Volume 2, pp. V2-35–V2-39. [Google Scholar]
  5. Chiang, J.Y.; Chen, Y.C. Underwater image enhancement by wavelength compensation and dehazing. IEEE Trans. Image Process. 2011, 21, 1756–1769. [Google Scholar] [CrossRef]
  6. Galdran, A.; Pardo, D.; Picón, A.; Alvarez-Gila, A. Automatic red-channel underwater image restoration. J. Vis. Commun. Image Represent. 2015, 26, 132–145. [Google Scholar] [CrossRef] [Green Version]
  7. Iqbal, K.; Salam, R.A.; Osman, A.; Talib, A.Z. Underwater Image Enhancement Using an Integrated Colour Model. IAENG Int. J. Comput. Sci. 2007, 34, 1–6. [Google Scholar]
  8. Kaur, M.; Kaur, J.; Kaur, J. Survey of contrast enhancement techniques based on histogram equalization. Int. J. Adv. Comput. Sci. Appl. 2011, 2, 137–141. [Google Scholar] [CrossRef]
  9. Ancuti, C.; Ancuti, C.O.; Haber, T.; Bekaert, P. Enhancing underwater images and videos by fusion. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 81–88. [Google Scholar]
  10. Drews, P.; Nascimento, E.; Moraes, F.; Botelho, S.; Campos, M. Transmission estimation in underwater single images. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Sydney, NSW, Australia, 2–8 December 2013; pp. 825–830. [Google Scholar]
  11. Peng, Y.T.; Cao, K.; Cosman, P.C. Generalization of the dark channel prior for single image restoration. IEEE Trans. Image Process. 2018, 27, 2856–2868. [Google Scholar] [CrossRef] [PubMed]
  12. LeCun, Y.; Bengio, Y.; Geoffrey, H. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  13. Li, C.; Guo, J.; Guo, C. Emerging from water: Underwater image color correction based on weakly supervised color transfer. IEEE Signal Process. Lett. 2018, 25, 323–327. [Google Scholar] [CrossRef] [Green Version]
  14. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  15. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An underwater image enhancement benchmark dataset and beyond. IEEE Trans. Image Process. 2019, 29, 4376–4389. [Google Scholar] [CrossRef] [Green Version]
  16. Wang, N.; Zhou, Y.; Han, F.; Zhu, H.; Yao, J. UWGAN: Underwater GAN for real-world underwater color restoration and dehazing. arXiv 2019, arXiv:1912.10269. [Google Scholar]
  17. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  18. Jiang, Q.; Wang, G.; Ji, T.; Wang, P. Underwater image denoising based on non-local methods. In Proceedings of the 2018 OCEANS-MTS/IEEE Kobe Techno-Oceans (OTO), Kobe, Japan, 28–31 May 2018; pp. 1–5. [Google Scholar]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  20. Niu, J.Y.; Xie, Z.H.; Li, Y.; Cheng, S.J.; Fan, J.W. Scale fusion light CNN for hyperspectral face recognition with knowledge distillation and attention mechanism. Appl. Intell. 2022, 52, 6181–6195. [Google Scholar] [CrossRef]
  21. Luo, Z.; Tang, Z.; Jiang, L.; Ma, G. A referenceless image degradation perception method based on the underwater imaging model. Appl. Intell. 2022, 52, 6522–6538. [Google Scholar] [CrossRef]
  22. Song, Q.; Wu, C.; Tian, X.; Song, Y.; Guo, X. A novel self-learning weighted fuzzy local information clustering algorithm integrating local and non-local spatial information for noise image segmentation. Appl. Intell. 2022, 52, 6376–6397. [Google Scholar] [CrossRef]
  23. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  24. Cho, S.; Lee, S. Fast motion deblurring. ACM Trans. Graph. 2009, 28, 1–8. [Google Scholar] [CrossRef]
  25. Xu, L.; Jia, J. Two-phase kernel estimation for robust motion deblurring. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2010; pp. 157–170. [Google Scholar]
  26. Rao, R.P.; Ballard, D.H. An active vision architecture based on iconic representations. Artif. Intell. 1995, 78, 461–505. [Google Scholar] [CrossRef] [Green Version]
  27. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1597–1604. [Google Scholar]
  28. Daugman, J.G. Complete discrete 2-D Gabor transforms by neural networks for image analysis and compression. IEEE Trans. Acoust. Speech Signal Process. 1988, 36, 1169–1179. [Google Scholar] [CrossRef] [Green Version]
  29. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 6, 679–698. [Google Scholar] [CrossRef]
  30. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  31. Peng, Y.T.; Cosman, P.C. Underwater image restoration based on image blurriness and light absorption. IEEE Trans. Image Process. 2017, 26, 1579–1594. [Google Scholar] [CrossRef] [PubMed]
  32. Song, W.; Wang, Y.; Huang, D.; Tjondronegoro, D. A rapid scene depth estimation model based on underwater light attenuation prior for underwater image restoration. In Pacific Rim Conference on Multimedia; Springer: Cham, Switzerland, 2018; pp. 678–688. [Google Scholar]
  33. Huang, D.; Wang, Y.; Song, W.; Sequeira, J.; Mavromatis, S. Shallow-water image enhancement using relative global histogram stretching based on adaptive parameter acquisition. In International Conference on Multimedia Modeling; Springer: Cham, Switzerland, 2018; pp. 453–465. [Google Scholar]
  34. Akkaynak, D.; Treibitz, T. Sea-thru: A method for removing water from underwater images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1682–1691. [Google Scholar]
  35. Islam, M.J.; Xia, Y.; Sattar, J. Fast underwater image enhancement for improved visual perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
  36. Panetta, K.; Gao, C.; Agaian, S. Human-visual-system-inspired underwater image quality measures. IEEE J. Ocean. Eng. 2015, 41, 541–551. [Google Scholar] [CrossRef]
  37. Chou, W. Maximum a posterior linear regression with elliptically symmetric matrix variate priors. In Proceedings of the Sixth European Conference on Speech Communication and Technology, Budapest, Hungary, 5–9 September 1999. [Google Scholar]
  38. Lei, T.; Jia, X.; Zhang, Y.; Liu, S.; Meng, H.; Nandi, A.K. Superpixel-based fast fuzzy C-means clustering for color image segmentation. IEEE Trans. Fuzzy Syst. 2018, 27, 1753–1766. [Google Scholar] [CrossRef]
Figure 1. The structure of improved U-Net.
Figure 1. The structure of improved U-Net.
Jmse 10 01513 g001
Figure 2. The structure of CBAM.
Figure 2. The structure of CBAM.
Jmse 10 01513 g002
Figure 3. The structure of CAM.
Figure 3. The structure of CAM.
Jmse 10 01513 g003
Figure 4. The structure of SAM.
Figure 4. The structure of SAM.
Jmse 10 01513 g004
Figure 5. Flowchart of sharpening algorithm.
Figure 5. Flowchart of sharpening algorithm.
Jmse 10 01513 g005
Figure 6. The comparison results of color correction effects between this research algorithm and other algorithms: (a,b,eg,i) are blue scene images, (c,d,h,j) are green scene images.
Figure 6. The comparison results of color correction effects between this research algorithm and other algorithms: (a,b,eg,i) are blue scene images, (c,d,h,j) are green scene images.
Jmse 10 01513 g006
Figure 7. Visibility edge of gray: (a) original image; and (b) enhanced image by the proposed algorithm.
Figure 7. Visibility edge of gray: (a) original image; and (b) enhanced image by the proposed algorithm.
Jmse 10 01513 g007
Figure 8. SIFT key point detection: (a) original image; (b) enhanced image by the proposed algorithm.
Figure 8. SIFT key point detection: (a) original image; (b) enhanced image by the proposed algorithm.
Jmse 10 01513 g008
Figure 9. Underwater image segmentation: (a) original image; (b) enhanced image by the proposed algorithm.
Figure 9. Underwater image segmentation: (a) original image; (b) enhanced image by the proposed algorithm.
Jmse 10 01513 g009
Table 1. Visible edge growth rate of different enhancement algorithms.
Table 1. Visible edge growth rate of different enhancement algorithms.
ImageIBLAUDCPULAPRGHSSea-thruUWGANFunieGANOurs
11.4521.4771.3431.2421.5021.6071.6221.375
21.4051.3311.2720.9921.4221.5071.4231.221
30.7922.2201.7211.5232.2042.3322.2152.274
40.5410.6051.0231.0051.6521.5521.4761.775
51.3051.4271.2751.1211.4451.3071.5022.513
Average1.0991.4121.3271.1771.6451.6611.6481.832
Table 2. Underwater image quality evaluation of different enhancement algorithms.
Table 2. Underwater image quality evaluation of different enhancement algorithms.
ScoresInputIBLAUDCPULAPRGHSSea-thruUWGANFunieGANOurs
SSIM0.7940.6940.5790.7560.7590.8040.8270.7790.745
PSNR17.21616.63113.12817.53216.48815.92114.74315.30117.637
UIQM1.3772.7722.9872.2702.2081.0661.8752.2404.035
UCIQE0.3790.4590.4970.4520.4470.3780.4760.4310.429
Table 3. Running time of different enhancement algorithms.
Table 3. Running time of different enhancement algorithms.
MethodsIBLAUDCPULAPRGHSSea-thruUWGANFunieGANOurs
Time (s)7.47743.14250.60911.44073.30121.50141.72561.4770
Table 4. The effect evaluation of improved U-Net.
Table 4. The effect evaluation of improved U-Net.
NetworkSSIMPSNRUIQMUCIQE
(a)0.70317.5211.4220.383
(b)0.71117.7551.4510.385
(c)0.72317.9681.5700.390
(d)0.73118.2011.7820.392
Table 5. The quality evaluation of different components of the proposed algorithm.
Table 5. The quality evaluation of different components of the proposed algorithm.
ExperimentSSIMPSNRUIQMUCIQE
(a)0.72417.4151.4270.388
(b)0.73118.2011.7820.392
(c)0.71617.3381.8450.401
(d)0.73318.2722.3290.413
(e)0.74518.6374.0350.429
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wu, Z.; Ji, Y.; Song, L.; Sun, J. Underwater Image Enhancement Based on Color Correction and Detail Enhancement. J. Mar. Sci. Eng. 2022, 10, 1513. https://doi.org/10.3390/jmse10101513

AMA Style

Wu Z, Ji Y, Song L, Sun J. Underwater Image Enhancement Based on Color Correction and Detail Enhancement. Journal of Marine Science and Engineering. 2022; 10(10):1513. https://doi.org/10.3390/jmse10101513

Chicago/Turabian Style

Wu, Zeju, Yang Ji, Lijun Song, and Jianyuan Sun. 2022. "Underwater Image Enhancement Based on Color Correction and Detail Enhancement" Journal of Marine Science and Engineering 10, no. 10: 1513. https://doi.org/10.3390/jmse10101513

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop