Next Article in Journal
Research on Fatigue Strength of Polar Icebreaker Structures Considering Ice Loads Based on Discrete Ice Element Model
Previous Article in Journal
CFD Investigation on Effect of Ship–Helicopter Coupling Motions on Aerodynamic Flow Field and Rotor Loads
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Image Enhancement Using Dynamic Color Correction and Lightweight Attention-Embedded SRResNet

1
Institute of Oceanographic Instrumentation, Qilu University of Technology (Shandong Academy of Sciences), Qingdao 266300, China
2
State Key Laboratory of Physical Oceanography, Qingdao 266300, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2025, 13(8), 1546; https://doi.org/10.3390/jmse13081546
Submission received: 13 July 2025 / Revised: 26 July 2025 / Accepted: 30 July 2025 / Published: 12 August 2025
(This article belongs to the Section Ocean Engineering)

Abstract

An enhancement method integrating dynamic color correction with a lightweight residual network is proposed to resolve the challenges of color bias and insufficient contrast in underwater imaging. The dynamic color correction module is implemented based on the gray-world assumption, adaptively adjusting inter-channel color shifts to mitigate blue-green dominance in acquired images. Subsequently, the corrected images are processed through an improved SRResNet architecture incorporating lightweight residual blocks with embedded channel–spatial attention mechanisms, enhancing the responses of feature channels and the saliency of spatial regions Model complexity is reduced through depthwise separable convolutions and channel dimension reduction, ensuring computational efficiency. Validation on UIEB and RUIE datasets demonstrates superior qualitative and quantitative performance, achieving PSNR gains of 0.92–5.95 dB and UCIQE improvements of 0.14–0.74, compared with the established methodologies. Ablation studies quantify the contributions of the color correction and attention mechanisms to the overall enhancement efficiency, verifying the network’s effectiveness.

1. Introduction

The ocean’s abundant resources and biodiversity have established it as a strategically vital region for all nations. Underwater imaging plays a pivotal role in uncovering critical features such as seabed topography, sediment distribution, and biodiversity patterns. The effective exploration of marine resources and the assessment of ecosystem health depend on high-quality underwater imagery to support scientific research and inform engineering decision-making. However, the complex and dynamic underwater environment considerably alters light propagation, presenting major challenges for imaging systems. As shown in Figure 1, intense scattering and absorption in seawater generate blue-green channel shifts and contrast reduction, resulting in substantial image quality degradation [1]. This deterioration not only impairs visual clarity but also obstructs tasks such as target detection, identification, and 3D reconstruction. Consequently, underwater image enhancement (UIE) techniques, aiming at rectifying color casts and enhancing contrast, have garnered significant attention and extensive research [2].
Drews-Jr et al. [3] proposed the Underwater Dark Channel Prior (UDCP), utilizing the lowest 0.1% of blue and green pixels to estimate transmission and ambient light, effectively removing haze by mitigating red-light attenuation. Retinex-based methods [4] separate illumination and reflectance through multi-scale filtering to correct color casts and enhance local contrast. However, they necessitate manual or semi-automatic parameter tuning, rendering them unstable under varying noise and lighting conditions. Pizer et al. [5] introduced Contrast-Limited Adaptive Histogram Equalization (CLAHE), which restricts noise amplification by clipping local histograms, substantially improving local contrast and detail clarity. Schechner and Karpel [6] employed polarization imaging to separate backscatter from transmitted light, effectively suppressing noise and improving visibility in turbid waters. Ancuti et al. [7] developed a Laplacian pyramid fusion framework that initially enhances various preprocessed outputs (e.g., white balance, gamma correction) independently and subsequently fuses them adaptively based on naturalness and contrast, balancing global color consistency with detail preservation. Peng et al. [8] presented a Physical Dicolor Model (PDM) histogram equalization approach that restores color distributions using an underwater imaging model before applying histogram equalization to enhance contrast, achieving both physical consistency and visual appeal. Although these methods obviate the need for large-scale annotations, their dependence on parameter estimation and susceptibility to environmental variations frequently result in inconsistent performance.
Recently, end-to-end deep learning has dominated underwater image enhancement by leveraging large labeled or synthetic datasets to learn direct degradation-enhancement mappings, enabling robust color bias removal and concurrent contrast and detail restoration. Li et al. [2] introduced WaterNet, employing a gated network to fuse white balance, histogram equalization and gamma correction outputs for adaptive color-cast correction. Anwar et al. [9] trained a lightweight UWCNN on synthetic datasets encompassing diverse degradations, circumventing physical-model errors while efficiently restoring fine details. Islam et al. [10] proposed FUnIE-GAN, integrating multi-scale perceptual losses into a GAN framework to generate natural textures and colors in real time on embedded platforms. Wei et al. [11] designed UHD-SFNet with dual branches—frequency-domain texture extraction via 1D convolutions and spatial color processing via U-RSGNet—whose fusion enables ultra-high-resolution reconstructions. Sun et al. [12] realized image enhancement through an end-to-end adaptive approach based on deep convolutional neural networks with an encoding–decoding structure. A key advantage of this method is that it eliminates the need to account for physical environmental factors. Wang et al. [13] presented UIEC2-Net, simultaneously extracting RGB and HSV features with adaptive attention fusion for end-to-end luminance and chromatic correction. Li et al. [14] introduced Emerging from Water, a weakly supervised CycleGAN-based color-transfer method utilizing adversarial, cycle-consistency and SSIM losses to learn underwater-to-air mappings without paired data. Lin et al. [15] proposed a dilated generative adversarial network (DGAN). Notably, this network not only enables the discriminator to judge the authenticity of the entire image, but also equips it with the capability to perform pixel-level classification for each constituent pixel. Xiang et al. [16] proposed a four-input deep learning model with polarized residual dense networks to effectively recover underwater mode images. Fu et al. [17] proposed an unsupervised underwater image restoration method. This approach leverages the homology between the original underwater image and its undegraded counterpart to effectively enhance image quality. Qin et al. [18] proposed MCRNet, fusing RGB/HSV/Lab features via self-attention to substantially improve color correction and contrast in severely degraded images. Shin et al. [19] proposed a three-stage cooperative ambient light estimation network. This network incorporates multiscale feature extraction, feature fusion, and non-linear regression to achieve accurate estimation of Ac. Hao et al. [20] proposed a two-stage underwater image restoration method based on physical modeling and causal intervention. By introducing a novel underwater image degradation formula and constructing an image restoration network integrated with multi-scale convolution and skip connections, this method achieves high-quality image restoration. Spandana et al. [21] put forward a multi-scale underwater image enhancement framework based on attention mechanisms and residual learning. Through designing an adaptive feature weighting module and building a cross-layer information interaction network, this framework realizes effective improvement of underwater image clarity. Despite excelling in color correction, contrast restoration, and detail preservation, these deep models remain constrained by large parameter counts, high computational costs, and substantial dependence on labeled data.
To address these issues, an underwater image enhancement method using dynamic color correction and lightweight attention-embedded SRResNet is presented in this paper. Dynamic color restoration is grounded in the gray-world hypothesis to adaptively rectify global color bias. A lightweight residual network—specifically, an enhanced SRResNet variant—was created for fine-grained detail refinement and contrast augmentation. By synergistically combining traditional color-constancy techniques for rapid bias mitigation with deep-learning-based high-frequency detail recovery, the presented approach substantially elevates visual fidelity. The contributions of this paper are listed as follows:
  • Dynamic Color-Recovery Module: Derives per-channel correction factors directly from image mean values—eliminating prior assumptions—to achieve effective color cast removal.
  • Lightweight Residual Blocks: An SRResNet-derived architecture leveraging depthwise separable convolutions with reduced channel dimensions and layer counts to minimize computational overhead, while preserving feature propagation through residual connections.
  • Comprehensive Experimental Validation: Conducts comparative and ablation studies on the UIEB dataset against traditional and established deep-learning methods, employing qualitative and quantitative metrics to substantiate the method’s superiority in color correction and detail enhancement.
The remainder of this paper is structured as follows: Section 2 details the proposed method; Section 3 presents experimental results and analyses; Section 4 concludes and discusses future directions.

2. Proposed Method

The proposed enhancement framework comprises two principal components, as illustrated in Figure 2. This section details the architectural design of (1) the Dynamic Color Recorrection module and (2) the lightweight residual-based SRResNet backbone, followed by the theoretical rationale for the training strategy and loss function design.

2.1. Dynamic Color Recorrection Module

Underwater images frequently manifest significant blue-green color casts due to wavelength-dependent light attenuation in aquatic environments. To address this phenomenon, a Dynamic Color Recorrection (DCR) module is introduced, grounded in the gray-world assumption that the mean intensities of all color channels in a natural image converge to equilibrium. This module adaptively rectifies color biases through a pixel-wise correction mechanism, as illustrated in Figure 3.
Given an input underwater image I R H × W × 3 , the proposed DCR module quantifies and rectifies color deviations through a four-step pipeline. First, the global mean intensities of the R, G, and B channels are computed to characterize color distribution:
      I ¯ c = 1 H × W H i = 1 W j = 1 I c ( i , j ) , c { R , G , B }
Here, I ¯ c denotes the mean luminance of channel c , with its deviation directly reflecting color cast severity. Following the gray-world hypothesis, the ideal average intensity for the three channels is identical. Thus, the arithmetic mean of these channel averages is defined as the target grayscale reference:
I ¯ g = I ¯ R + I ¯ G + I ¯ B 3
This value establishes the neutral gray baseline for subsequent gain calculations. To ensure numerical stability, a lower-bound constraint is applied to the gray scale reference I ¯ g = m a x ( I ¯ g , ϵ ) , incorporating a constant ϵ = 0.01 for stability and calculating per-channel correction gains:
g c = I ¯ g I ¯ c , c { R , G , B }
The gain factor g c quantifies required compensation: if a channel’s mean intensity is below the target gray level (for instance, the red channel under a blue-green cast), g c exceeds unity to boost its response; otherwise, g c attenuates the channel. The corrected image is generated by applying these channel-specific gains to each pixel:
I ^ c ( i , j ) = g c I c ( i , j ) , c { R , G , B }
To prevent pixel values from exceeding valid ranges after channel-wise gain adjustment—which may lead to overexposure or color distortion—a clipping operation is introduced following color correction. This step constrains each channel’s pixel values within the valid range of [0, 1], and is expressed as follows:
I c i , j = min max I ^ c i , j , 0 , 1 , c { R , G , B }
The output I R H × W × 3 exhibits balanced color chromaticity while preserving the global luminance structure. This dynamic gain strategy integrates the classical gray-world color constancy principle with an adaptive correction mechanism, effectively eliminating underwater color casts without relying on prior scene-specific knowledge.

2.2. Lightweight Residual Enhancement Network

Although dynamic color correction mitigates color casts, the enhanced image I often exhibits persistent low contrast and blurred details due to residual degradation. To address these limitations, a lightweight residual enhancement network based on SRResNet is proposed, as illustrated in Figure 4. The residual network integrates three key design principles: depthwise separable convolutions to reduce computational costs to approximately 1/8 of standard convolutional operations; a dual attention mechanism to amplify salient features; and residual connections to maintain information flow while mitigating gradient vanishing, thereby rendering the network suited for underwater image enhancement tasks that demand high-fidelity detail preservation.
The network initiates feature extraction through a head convolutional layer:
F 0 = σ ( W h I + b h )
where W h R 9 × 9 × 3 × 64 denotes the convolutional kernel, and σ represents the activation function. The 9 × 9 kernel captures spatial features within an extended receptive field, while ReLU function introduces non-linear transformation to enhance feature discriminability. The bias term b h adaptively modulates activation threshold across channels, enabling robust adaptation to diverse input distributions.
Subsequently, the initial feature map F 0 undergoes progressive refinement through eight cascaded lightweight residual blocks. Each residual block R e s B l o c k i produces output defined as follows:
F i = R e s B l o c k i F i 1 , i = 1,2 , , N
This cascaded structure enables progressive extraction of multilevel feature representations, transitioning from low-level textures to high-level semantic information, ultimately producing the enhanced image through the tail fusion module:
I E = W t F N + b t
where W t maps the high-dimensional features back into the three-channel RGB color space, and b t is the bias parameter of the tail convolutional layer used to adjust the reference values of RGB channels in the output image.

Core Design of Lightweight Residual Block

The lightweight residual block serves as the core building block of the enhancement network. Integrating depthwise separable convolution with an attention mechanism, achieves computational complexity of approximately 1/8 that of standard convolutions while maintaining almost equivalent performance. Given an input feature map F i n , the block performs sequential feature transformation through two depthwise separable convolutional layers, where the first layer simultaneously performs spatial feature extraction and channel fusion, formulated as follows:
F 1 = σ ( B N ( W p w 1 ( W d w 1 g r o u p s F i n ) ) )
where W p w is the depthwise convolution kernel ( 3 × 3 × 1 , groups = 64) that independently processes the spatial information per channel, and W p w is the pointwise convolution kernel ( 1 × 1 × 64 × 64 ) for cross-channel fusion. The second layer preserves linear representation by omitting non-linear activation, formulated as follows:
F 2 = B N ( W p w 2 ( W d w 2 g r o u p s F 1 ) )
To enhance feature representation, a tandem structure combining channel attention and spatial attention is introduced. The structure of the channel attention is illustrated in Figure 5, where the channel attention mechanism captures channel-wise statistics via global average pooling, as follows:
F c a = F 2 σ ( M L P ( G A P ( F 2 ) ) + M L P ( G M P ( F 2 ) ) )
GAP and GMP denote global average pooling and global max pooling, respectively, where MLP represents a two-layer fully connected network with shared weights across channels, and indicates element-wise multiplication. The spatial attention module generates spatial weight maps by aggregating channel-wise statistical information as follows:
F s a = F c a σ ( C o n v 7 × 7 ( [ A v g P o o l ( F c a ) ; M a x P o o l ( F c a ) ] )
Figure 6 shows the structure of spatial attention. AvgPool and MaxPool represent channel-wise global average pooling and global max pooling operations, respectively, which aggregate spatial information by computing mean and maximum statistics across spatial dimensions. Their concatenated outputs are processed through a 7 × 7 convolution followed by the activation of σ to generate the spatial attention map. This dual attention design enables simultaneous focus on critical feature channels and spatial locations, enhancing feature discriminability. The enhanced features are fused with the original input through the residual concatenation as follows:
F o u t = F i n + F s a
Residual connections mitigate the vanishing gradient issues in deep networks while preserving intact information flow, thereby enabling the network to concentrate on learning enhancement-specific residuals rather than reproducing input features.

2.3. Loss Functions and Optimization Strategy

A hybrid loss function is designed to jointly optimize pixel-level accuracy and structural similarity:
L t = λ L L 1 + ( 1 λ ) L S S I M
where L L 1 = | | I E I g t | | 1 ensures pixel-level accuracy, and L S S I M = 1 S S I M ( I E , I g t ) preserves structural consistency. Here, I E denotes the enhanced image output by the mode, I g t represents the corresponding high-quality underwater reference image, and λ is the weighting coefficient. This formulation guarantees that the enhanced output maintains both pixel-level proximity to the ground truth and structurally coherent characteristics, thereby ensuring multi-scale perceptual quality preservation. The complete underwater image enhancement process that integrates dynamic color restoration and residual learning can be formulated as follows:
I E = F R ( F c ( I ) )
where F c denotes the dynamic color restoration function, and F R represents the lightweight residual network mapping. This two-stage framework synergistically combines physical priors of underwater image degradation with data-driven refinement, achieving efficient high-fidelity enhancement.

3. Experimental Validation

3.1. Dataset Selection and Experimental Configuration

Underwater datasets constitute a pivotal foundation for deep-learning-based image enhancement, enabling model training and performance evaluation. Underwater datasets are typically constructed via two paradigms: (1) synthetic data generation using style transfer techniques such as CycleGAN [22] to generate clear reference images from underwater scenes, offering cost-effective scalability; and (2) real-world data acquisition involving underwater scenes capture followed by traditional enhancement for reference generation, which better preserves complex real-world degradation patterns despite higher resource requirements.
Evaluation is performed on two widely used UIE datasets: UIEB [2] and RUIE [23]. UIEB comprises 950 real underwater images, including 890 paired samples for training and 60 challenging unpaired cases exhibiting greenish or bluish casts, low contrast, and turbidity. RUIE provides extended quality variations for robustness validation. Performance is assessed using both full-reference metrics (PSNR, SSIM) and no-reference metrics (UCIQE, UIQM).
Experiments were conducted on a platform equipped with an AMD R5-5600 CPU, 16 GB RAM, and an AMD Radeon RX 6750 GPU (AMD Inc., Santa Clara, CA, USA). The training input images were cropped to 512 × 512 pixels. The network architecture was implemented using PyTorch (version 2.0.1+cpu), with model optimization performed via the Adam optimizer. Data augmentation was achieved through horizontal flipping and random in-plane rotations of 90° and 180°. The training protocol was configured with a batch size of 8, 80 epochs, and an initial learning rate of 0.0002.

3.2. Qualitative Comparison

Qualitative evaluation was conducted on the UIEB dataset (including the Challenging Set) and the raw images from the RUIE dataset. The proposed model was compared against several existing underwater image enhancement algorithms, encompassing traditional methods such as UDCP [3] and MLLE [24], and deep-learning-based approaches including DICAM [25], FA+Net [26], DeepWater [27], and PUIE-Net [28]. Most of the comparison methods were implemented using publicly available code or pretrained models released by the original authors, and re-evaluated under the same test conditions for fairness. Representative qualitative results are presented in Figure 7 and Figure 8.
As illustrated in Figure 7 and Figure 8, all examples were selected from the UIEB test set and the RUIE dataset. The first row displays the original underwater images, while rows 2–8 present enhancement results from seven comparative methods in the following order: UDCP [3], MLLE [24], DICAM [25], FA+Net [26], DeepWater [9], PUIE-Net [27], and the proposed method. Traditional approaches such as UDCP and MLLE exhibit some limitations, including color overcorrection, detail loss, and inadequate handling of complex underwater degradation, often resulting in unnatural outputs due to excessive compensation in blue and green channels. In contrast, deep-learning-based methods demonstrate improved detail restoration, where DICAM and FA+Net effectively enhance contrast but may introduce residual color casts or oversaturated regions, while PUIE-Net achieves more natural color reproduction yet maintains a slight greenish tint with constrained contrast improvement.
The experimental results demonstrate that the proposed method not only effectively eliminates blue-green color casts but also simultaneously enhances image contrast and texture details, thereby producing underwater images with significantly improved clarity and naturalness. The method exhibits high fidelity in structural detail preservation, generating output images that achieve better visual consistency with ground truth references. Across multiple comparative trials, the method consistently yields enhanced results characterized by natural color reproduction and rich detail representation, confirming its superior subjective visual quality.

3.3. Quantitative Comparison

To quantitatively evaluate the performance of the aforementioned methods, full-reference metrics including Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM) are employed. PSNR [28] quantifies the ratio of maximum signal power to corrupting noise power, calculated based on pixel-wise differences between reference and distorted images:
P S N R = 10 l o g 10 ( M A X I 2 M S E )
where M A X I 2 is the maximum possible pixel value in the image (usually 255 for 8-bit images), and M S E is the mean square error. SSIM [29] integrates luminance, contrast, and structural information into a comprehensive quality assessment of the perceived image, grounded in human visual perception, defined as follows:
S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
μ x , μ y denote the mean values of images x and y , respectively. σ x 2 and σ y 2 represent the variances of images x and y , respectively; σ x y stands for the covariance between images   x and y . Additionally, C 1 , C 2   are constants introduced to stabilize the denominator. Higher PSNR and SSIM values indicate superior enhancement performance. For no-reference assessment, Underwater Color Image Quality Evaluation (UCIQE) and Underwater Image Quality Measure (UIQM) are adopted. UCIQE [30] combines chroma, saturation, and contrast through a linear combination as follows:
U C I Q E = c 1 × σ c + c 2 × c o n + c 3 × μ s
where σ c is the standard deviation of chromaticity, c o n denotes luminance contrast, μ s denotes the mean value of saturation, and c 1 , c 2 , c 3 are weighting factors. UIQM [31] evaluates the colorfulness, sharpness, and contrast of underwater images using the following formula:
U I Q M = b 1 × U I C M + b 2 × U I S M + b 3 × U I C o n M
U I C M denotes the color metric of the underwater image, U I S M denotes the sharpness metric of the underwater image, U I C o n M denotes the contrast metric of the underwater image, and b 1 , b 2 , b 3 are weighting factors. Higher UIQM values reflect better perceptual quality of the underwater images. All methods were evaluated on 50 images from both the UIEB and RUIE datasets, with average scores reported in Table 1. The highest values in each category are highlighted in red.
Table 1 summarizes the comparative performance of the evaluated methods on the UIEB and RUIE test sets. On the UIEB dataset, the proposed approach achieves the highest PSNR of 20.093 dB, demonstrating superior noise suppression capability while preserving fine structural details for high-fidelity image reconstruction. It attains leading scores of 0.965 in UCIQE and 1.563 in UIQM, where UCIQE quantifies the efficacy of color correction and contrast enhancement, and UIQM evaluates comprehensive colorfulness, sharpness, and contrast—performance metrics that this method outperforms others in. The substantial margin over competing methods validates the effectiveness of the proposed color-compensation and contrast-enhancement strategies in addressing light attenuation and color-cast distortions.
For the more challenging RUIE dataset, the framework exhibits robust generalization performance. It maintains a PSNR of 18.378 dB, comparable to UIEB results, while achieving the highest UCIQE score among all methods. While SSIM is slightly lower than some competing methods, this can be attributed to the dynamic color restoration module, which applies channel-wise gain compensation prior to structural enhancement. This correction improves global color fidelity and visibility but may slightly shift pixel-wise structural alignment compared to the ground truth, thus lowering SSIM. Nonetheless, the proposed framework consistently delivers visually superior results across diverse underwater conditions. Notably, under extreme degradation caused by suspended particles and light scattering, the method effectively reconstructs natural color tones and restores fine details, confirming its robustness and adaptability in complex underwater environments.

3.4. Ablation Studies

To validate the contributions of key components, four ablation experiments were systematically conducted to evaluate the individual and combined effects of the Dynamic Color Recorrection (DCR) module and the Convolutional Block Attention Module (CBAM). All models were trained under identical hyperparameter configurations on the UIEB dataset and evaluated across both UIEB and RUIE test sets. Qualitative results are summarized in Table 2, with performance metrics including PSNR, SSIM, UCIQE, and UIQM.
Table 2 reports the quantitative performance of ablation configurations on the UIEB and RUIE test sets. The baseline model, excluding both DCR and CBAM, exhibits the lowest performance, confirming inadequate detail reconstruction and color correction capabilities. The DCR module demonstrates significant improvements in PSNR and related metrics, highlighting its critical role in color restoration. Sole integration of the CBAM substantially elevates UCIQE scores, validating its effectiveness in enhancing salient features through attention mechanisms. The full model, incorporating both DCR and CBAM, achieves optimal performance across all metrics, underscoring the synergistic advantages of their combined implementation.

4. Conclusions

This paper presents a novel underwater image enhancement framework that synergistically combines gray-world-based dynamic color restoration with a lightweight residual network. The proposed Dynamic Color Recorrection module adaptively corrects underwater color casts, while the subsequent residual network integrates spatial and channel attention mechanisms to enhance image details and contrast. Experimental results on the UIEB and RUIE datasets demonstrate that this method effectively suppresses blue-green color casts, significantly improves image clarity and contrast, and outperforms both traditional and deep-learning-based methods across PSNR, SSIM, UCIQE, and UIQM metrics. The framework generates visually natural results while maintaining computational efficiency. Future work will focus on real-time inference optimization, deployment in lightweight underwater robotic systems, and extended validation in diverse marine environments.

Author Contributions

K.Z., Conceptualization, Methodology and Writing—original draft. Y.Z., Resources, Conceptualization and Funding acquisition. D.Y., Conceptualization, Methodology and Resources. X.F., Software, Validation and Supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This study was financially supported by National Key R&D Program of China (No. 2024YFE0108700), Taishan Scholars Program, Key R&D Program Project of Shandong Province (No. 2023CXPT015), Qingdao Science and Technology Innovation Project (No. 24-1-3-hygg-20-hy), and Science and Education Production Project of Qilu University of Technology (Shandong Academy of Sciences) (No. 2024GH10).

Data Availability Statement

Data will be available upon a suitable request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Raihan, J.A.; Abas, P.A.; De Silva, L.C. Review of underwater image restoration algorithms. IET Image Process. 2019, 13, 1587–1596. [Google Scholar] [CrossRef]
  2. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An Underwater Image Enhancement Benchmark Dataset and Beyond. IEEE Trans. Image Process. 2019, 29, 4376–4389. [Google Scholar] [CrossRef] [PubMed]
  3. Drews-Jr, P.L.; Nascimento, E.; Moraes, F.; Botelho, S.; Campos, M. Transmission estimation in underwater single images. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Sydney, Australia, 2–8 December 2013; pp. 825–830. [Google Scholar]
  4. Land, E.H.; McCann, J.J. Lightness and Retinex Theory. J. Opt. Soc. Am. 1971, 61, 1–11. [Google Scholar] [CrossRef] [PubMed]
  5. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; Romeny, B.T.H.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  6. Schechner, Y.Y.; Karpel, N. Recovery of underwater visibility and structure by polarization analysis. IEEE J. Ocean. Eng. 2005, 30, 570–587. [Google Scholar] [CrossRef]
  7. Ancuti, C. Color Balance and Fusion for Underwater Image Enhancement. IEEE Trans. Image Process. 2017, 27, 379–393. [Google Scholar] [CrossRef]
  8. Peng, Y.T.; Chen, Y.R.; Chen, Z.; Wang, J.H. Underwater image enhancement based on histogram-equalization approximation using physics-based dichromatic modeling. Sensors 2022, 22, 2168. [Google Scholar] [CrossRef]
  9. Anwar, S.; Li, C.; Porikli, F. Deep Underwater Image Enhancement. arXiv 2018, arXiv:1807.03528. [Google Scholar]
  10. Islam, M.J.; Xia, Y.; Sattar, J. Fast underwater image enhancement for improved visual perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
  11. Wei, Y.; Zheng, Z.; Jia, X. UHD underwater image enhancement via frequency-spatial domain aware network. In Proceedings of the Asian Conference on Computer Vision, Macao, China, 4–8 December 2022; pp. 299–314. [Google Scholar]
  12. Sun, X.; Liu, L.P.; Dong, J.Y. Underwater image enhancement with encoding-decoding deep CNN networks. In Proceedings of the IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computed, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation, San Francisco, CA, USA, 4–8 August 2017; IEEE Computer Society Press: Los Alamitos, CA, USA, 2017; pp. 1–6. [Google Scholar]
  13. Wang, Y.; Guo, J.; Gao, H.; Yue, H. UIEC2-Net: CNN-based Underwater Image Enhancement Using Two Color Space. Signal Process. Image Commun. 2021, 96, 116250. [Google Scholar] [CrossRef]
  14. Li, C.; Guo, J.; Guo, C. Emerging from Water: Underwater Image Color Correction Based on Weakly Supervised Color Transfer. IEEE Signal Process. Lett. 2018, 25, 323–327. [Google Scholar] [CrossRef]
  15. Lin, J.C.; Hsu, C.B.; Lee, J.C. Dilated generative adversarial networks for underwater image restoration. J. Mar. Sci. Eng. 2022, 10, 500. [Google Scholar] [CrossRef]
  16. Xiang, Y.; Yang, X.; Ren, Q.; Wang, G.; Gao, J.; Chew, K.H.; Chen, R.P. Underwater polarization imaging recovery based on polarimetric residual dense network. IEEE Photon. J. 2022, 14, 7860206. [Google Scholar] [CrossRef]
  17. Fu, Z.; Lin, H.; Yang, Y.; Chai, S.; Sun, L.; Huang, Y.; Ding, X. Unsupervised underwater image restoration: From a homology perspective. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; AAAI Press: Palo Alto, CA, USA, 2022; pp. 643–651. [Google Scholar]
  18. Qin, N.; Wu, J.; Liu, X. MCRNet: Underwater image enhancement using multi-color space residual network. Biomim. Intell. Robot. 2024, 4, 100169. [Google Scholar] [CrossRef]
  19. Shin, Y.S.; Cho, Y.; Pandey, G.; Kim, A. Estimation of ambient light and transmission map with common convolutional architecture. In Proceedings of the OCEANS 2016 MTS/IEEE Monterey, Monterey, CA, USA, 19–23 November 2016; IEEE: New York, NY, USA, 2016; pp. 1–7. [Google Scholar]
  20. Hao, J.; Yang, H.; Hou, X.; Zhang, Y. Two-stage underwater image restoration algorithm based on physical model and causal intervention. IEEE Signal Process. Lett. 2023, 30, 120–124. [Google Scholar] [CrossRef]
  21. Spandana, C.; Srisurya, I.V.; Priyadharshini, A.R.; Krithika, S.; Nandhini, S.A. Underwater image enhancement and restoration using cycle GAN. In Proceedings of the International Conference on Innovative Computing and Communication, New Delhi, India, 16–17 February 2023; Springer Nature: Singapore, 2023; pp. 99–110. [Google Scholar]
  22. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  23. Liu, R.; Fan, X.; Zhu, M. Real-world underwater enhancement: Challenges, benchmarks, and solutions under natural light. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 4861–4875. [Google Scholar] [CrossRef]
  24. Zhang, W.; Zhuang, P.; Sun, H. Underwater image enhancement via minimal color loss and locally adaptive contrast enhancement. IEEE Trans. Image Process. 2022, 31, 3997–4010. [Google Scholar] [CrossRef]
  25. Tolie, H.F.; Ren, J.; Elyan, E. DICAM: Deep inception and channel-wise attention modules for underwater image enhancement. Neurocomputing 2024, 584, 127585. [Google Scholar] [CrossRef]
  26. Jiang, J.; Ye, T.; Bai, J. Five A + Network: You Only Need 9K Parameters for Underwater Image Enhancement. arXiv 2023, arXiv:2305.08824. [Google Scholar]
  27. Fu, Z.; Li, X.; Ma, H.; Yan, P.; Li, Q.; Chen, W. Uncertainty Inspired Underwater Image Enhancement. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer Nature: Cham, Switzerland, 2022; pp. 465–482. [Google Scholar]
  28. Kumar, N.N.; Ramakrishna, S. An Impressive Method to Get Better Peak Signal Noise Ratio (PSNR), Mean Square Error (MSE) Values Using Stationary Wavelet Transform (SWT). Glob. J. Comput. Sci. Technol. Graph. Vis. 2012, 12, 34–40. [Google Scholar]
  29. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  30. Yang, M.; Sowmya, A. An underwater color image quality evaluation metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef]
  31. Panetta, K.; Gao, C.; Agaian, S. Human-visual-system-inspired underwater image quality measures. IEEE J. Ocean Eng. 2015, 41, 541–551. [Google Scholar] [CrossRef]
Figure 1. Underwater degradation image.
Figure 1. Underwater degradation image.
Jmse 13 01546 g001
Figure 2. Workflow architecture of the proposed underwater image enhancement framework.
Figure 2. Workflow architecture of the proposed underwater image enhancement framework.
Jmse 13 01546 g002
Figure 3. DCR module workflow architecture.
Figure 3. DCR module workflow architecture.
Jmse 13 01546 g003
Figure 4. Architecture of the proposed lightweight residual enhancement network.
Figure 4. Architecture of the proposed lightweight residual enhancement network.
Jmse 13 01546 g004
Figure 5. Architecture of channel attention module.
Figure 5. Architecture of channel attention module.
Jmse 13 01546 g005
Figure 6. Architecture of spatial attention module.
Figure 6. Architecture of spatial attention module.
Jmse 13 01546 g006
Figure 7. Enhancement results on the UIEB dataset. (a) RAW input; (b) UDCP [3]; (c) MLLE [24]; (d) DICAM [25]; (e) FA+Net [26]; (f) DeepWater [9]; (g) PUIE-Net [27]; (h) proposed method.
Figure 7. Enhancement results on the UIEB dataset. (a) RAW input; (b) UDCP [3]; (c) MLLE [24]; (d) DICAM [25]; (e) FA+Net [26]; (f) DeepWater [9]; (g) PUIE-Net [27]; (h) proposed method.
Jmse 13 01546 g007
Figure 8. Enhancement results on the RUIE dataset. (a) RAW input; (b) UDCP [3]; (c) MLLE [24]; (d) DICAM [25]; (e) FA+Net [26]; (f) DeepWater [9]; (g) PUIE-Net [27]; (h) proposed method.
Figure 8. Enhancement results on the RUIE dataset. (a) RAW input; (b) UDCP [3]; (c) MLLE [24]; (d) DICAM [25]; (e) FA+Net [26]; (f) DeepWater [9]; (g) PUIE-Net [27]; (h) proposed method.
Jmse 13 01546 g008
Table 1. Quantitative comparison of enhancement methods in terms of PSNR, SSIM, UCIQE, and UIQM.
Table 1. Quantitative comparison of enhancement methods in terms of PSNR, SSIM, UCIQE, and UIQM.
MethodsUIEBRUIE
PSNRSSIMUCIQEUIQMPSNRSSIMUCIQEUIQM
UDCP14.1470.6280.2230.35713.1010.6200.1950.328
MLLE17.7560.7170.2531.34014.3220.6390.2591.623
DICAM18.2080.7550.2611.10515.4800.7960.2511.348
FA+Net19.1740.8360.2731.20815.1610.7720.2461.468
DeepWater17.0790.6800.3150.64115.1380.6650.3170.895
PUIE-Net19.0480.8290.8291.08016.8110.8130.2421.559
Ours20.0930.8050.9651.56318.3780.6890.4921.611
Table 2. Performance comparison of underwater image enhancement performance under different module configurations.
Table 2. Performance comparison of underwater image enhancement performance under different module configurations.
Module NameAdding ModulesPSNRSSIMUCIQEUIQM
Methods 1-15.3440.4920.5231.361
Methods 2+DCR17.8260.6630.7011.511
Methods 3+CBAM16.0870.6180.6901.448
Methods 4DCR + CBAM19.2360.7470.7291.587
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, K.; Zhang, Y.; Yuan, D.; Feng, X. Underwater Image Enhancement Using Dynamic Color Correction and Lightweight Attention-Embedded SRResNet. J. Mar. Sci. Eng. 2025, 13, 1546. https://doi.org/10.3390/jmse13081546

AMA Style

Zhang K, Zhang Y, Yuan D, Feng X. Underwater Image Enhancement Using Dynamic Color Correction and Lightweight Attention-Embedded SRResNet. Journal of Marine Science and Engineering. 2025; 13(8):1546. https://doi.org/10.3390/jmse13081546

Chicago/Turabian Style

Zhang, Kui, Yingying Zhang, Da Yuan, and Xiandong Feng. 2025. "Underwater Image Enhancement Using Dynamic Color Correction and Lightweight Attention-Embedded SRResNet" Journal of Marine Science and Engineering 13, no. 8: 1546. https://doi.org/10.3390/jmse13081546

APA Style

Zhang, K., Zhang, Y., Yuan, D., & Feng, X. (2025). Underwater Image Enhancement Using Dynamic Color Correction and Lightweight Attention-Embedded SRResNet. Journal of Marine Science and Engineering, 13(8), 1546. https://doi.org/10.3390/jmse13081546

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop