Next Article in Journal
Development and Validation of a Low-Cost Arduino-Based Lee Disc System for Thermal Conductivity Analysis of Sustainable Roofing Materials
Previous Article in Journal
A Feasibility Study of Automated Detection and Classification of Signals in Distributed Acoustic Sensing
Previous Article in Special Issue
Wind Turbine Blade Defect Recognition Method Based on Large-Vision-Model Transfer Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Progressive Multi-Scale Perception Network for Non-Uniformly Blurred Underwater Image Restoration

1
School of Artificial Intelligence, Henan Institute of Science and Technology, Xinxiang 453003, China
2
National and Local Joint Engineering Laboratory of Internet Application Technology on Mine, China University of Mining and Technology, Xuzhou 221116, China
3
School of Information Science and Engineering, Shenyang University of Technology, Shenyang 110870, China
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(17), 5439; https://doi.org/10.3390/s25175439
Submission received: 13 August 2025 / Revised: 28 August 2025 / Accepted: 29 August 2025 / Published: 2 September 2025
(This article belongs to the Special Issue Deep Learning for Perception and Recognition: Method and Applications)

Abstract

Underwater imaging is affected by spatially varying blur caused by water flow turbulence, light scattering, and camera motion, resulting in severe visual quality loss and diminished performance in downstream vision tasks. Although numerous underwater image enhancement methods have been proposed, the issue of addressing non-uniform blur under realistic underwater conditions remains largely underexplored. To bridge this gap, we propose PMSPNet, a Progressive Multi-Scale Perception Network, designed to handle underwater non-uniform blur. The network integrates a Hybrid Interaction Attention Module to enable precise modeling of feature ambiguity directions and regional disparities. In addition, a Progressive Motion-Aware Perception Branch is employed to capture spatial orientation variations in blurred regions, progressively refining the localization of blur-related features. A Progressive Feature Feedback Block is incorporated to enhance reconstruction quality by leveraging iterative feature feedback across scales. To facilitate robust evaluation, we construct the Non-uniform Underwater Blur Benchmark, which comprises diverse real-world blur patterns. Extensive experiments on multiple real-world underwater datasets demonstrate that PMSPNet consistently surpasses state-of-the-art methods, achieving on average 25.51 dB PSNR and an inference speed of 0.01 s, which provides high-quality visual perception and downstream application input from underwater sensors for underwater robots, marine ecological monitoring, and inspection tasks.

1. Introduction

With the advancement of marine science and technology, high-definition underwater imaging has become increasingly vital for underwater applications such as underwater robotic navigation, seabed topographic mapping, and aquatic life monitoring [1,2]. However, underwater imaging is affected by a variety of factors, including light absorption, scattering effects, and relative motion between the camera and the dynamic scene, resulting in image degradation, especially the non-uniform spatial blur distribution. We perform Fourier analysis on underwater images, extract the direction and intensity of localized frequency domain energy, and map them into blur direction and intensity heatmaps to visualize underwater blur patterns. As illustrated in Figure 1, real-world underwater images commonly exhibit non-uniform blur that varies in spatial extent and directional orientation, violating the common assumption of uniform blur kernels. Such degradation obscures critical visual cues, impairing the performance of downstream computer vision tasks. In autonomous underwater systems, the loss of structural information and motion cues due to non-uniform blur can lead to perceptual errors and decision-making biases, thereby posing challenges to the robustness and reliability of underwater operations.
To mitigate underwater image degradation, numerous hardware-based solutions have been developed, including specialized underwater cameras, structured lighting systems, and active imaging techniques such as laser illumination and time-gated imaging. These approaches aim to reduce visibility loss and scattering effects during image acquisition, thereby alleviating certain forms of blur. However, despite their potential to enhance raw image quality, such systems are commonly challenged for deployment in real-world environments due to their high cost, large physical footprint, and sensitivity to the environment. More critically, hardware-based methods do not explicitly address the underwater non-uniform blur arising from camera motion, moving objects, or dynamic water flow. This limitation has spurred increasing interest in algorithmic deblurring techniques, which can restore image sharpness directly from captured data without the need for auxiliary hardware.
Numerous studies have investigated traditional image enhancement techniques to address underwater image degradation. These methods typically leverage handcrafted priors and physical modeling to compensate for light attenuation and scattering. Common strategies include histogram equalization, white balance, gray-world assumptions, and Retinex-based methods for illumination decomposition. Some approaches further adapt atmospheric dehazing techniques, such as the Dark Channel Prior (DCP), to estimate the transmission map of underwater scenes. While these methods are relatively easy to deploy and interpretable, they generally assume static scenes and well-defined image structures, rendering them ineffective in the presence of the spatially non-uniform blur prevalent in real-world underwater environments. As a result, they commonly fail to recover fine image details and maintain structural consistency, limiting their utility in supporting high-level visual tasks.
With the growing application of deep learning (DL) in image restoration, learning-based methods for underwater image enhancement (UIE) have attracted increasing attention. Convolutional Neural Networks (CNNs) and Generative Adversarial Networks (GANs) are employed to correct color distortions, enhance contrast, and improve the visibility of underwater images, owing to their powerful feature extraction and representation learning capabilities. More recently, Transformer-based architectures have been introduced into underwater image enhancement and restoration tasks, demonstrating promising performance due to their global receptive field and long-range dependency modeling. However, most of these methods remain focused primarily on enhancing the visual quality of the image, lacking explicit mechanisms to model underwater image blur. Furthermore, the enhancement performance of the above networks often scales with their architectural complexity, which increases significantly in Transformer-based designs. This imposes substantial computational burdens, making real-time deployment in underwater applications impractical. Consequently, the performance of existing enhancement algorithms is frequently compromised under non-uniform blur conditions, underscoring the urgent need for dedicated deblurring frameworks tailored to the unique challenges of underwater environments.
Focusing on the challenge of non-uniform blur in underwater images, we propose PMSPNet, a Progressive Multi-Scale Perception Network. PMSPNet perceives and models blur features from multiple perspectives, encompassing both local and global receptive fields, contextual semantic information, as well as the direction and intensity of the blur. By integrating these diverse perceptual cues, the network effectively maximizes blur removal while enhancing overall image quality.
The main work of this article is summarized as follows:
  • We propose a Progressive Multi-Scale Perception Network to effectively eliminate non-uniform blur in underwater images, enabling real-time underwater image enhancement.
  • We introduce a Hybrid Interaction Attention Module that extracts and integrates local and global blur features to capture multi-view information and accurately perceive the direction and intensity of underwater blur.
  • We design a Progressive Motion-Aware Perception Branch and a Progressive Feature Feedback Block to enable progressive fine-tuning of features, precise localization of blur, and efficient recovery of reconstruction details.
  • We construct a Non-uniform Underwater Blur Dataset to provide a benchmark for evaluating underwater image deblurring algorithms. Extensive experiments demonstrate that the proposed method outperforms state-of-the-art approaches, validating its robustness and effectiveness.
The remainder of this article is organized as follows. Section 2 introduces related work. Section 3 describes the PMSPNet network. Section 4 presents the analysis and discussion of the experimental results, and Section 5 of this article concludes with a summary of the article and discusses potential future research areas.

2. Related Work

2.1. Hardware-Based Approach

To address the challenges of underwater image degradation, numerous hardware-based methods have been proposed. These approaches employ specialized imaging systems to capture higher-quality data at the point of acquisition, thereby reducing dependence on post-processing algorithms [3]. Some systems utilize auxiliary light sources or polarization filters to suppress scattering and backscatter effects [4,5]. In contrast, others adopt structured light or multi-camera setups to reconstruct clearer underwater scenes [6,7]. Notably, high-speed cameras and inertial measurement units (IMUs) have been leveraged to estimate and compensate for camera motion, thereby mitigating motion blur during image capture [8]. In parallel, the emergence of IoT-enabled sensor networks has driven research on adaptive routing, data reliability, and energy efficiency, which are critical for real-time underwater monitoring and communication systems [9,10,11]. These studies explore strategies such as cross-layer optimization, secure data aggregation, and energy-aware routing to improve the dependability and scalability of sensor-based infrastructures [12,13,14], providing complementary insights into the broader landscape of underwater sensing and communication. Integrating underwater image restoration techniques with efficient IoT-based sensing systems represents a promising direction, enabling both high-quality perception and reliable data delivery in challenging underwater environments. Although such hardware-enhanced systems can yield superior image quality under controlled conditions, they are commonly constrained by practical limitations, including high cost, limited deployment flexibility, and susceptibility to failure in dynamic or harsh underwater environments. Consequently, there is a pressing need for software-based solutions that can adaptively enhance underwater images without relying on specialized equipment, offering greater practicality, scalability, and robustness in real-world applications.

2.2. Traditional Approach

Traditional approaches to underwater image enhancement typically rely on physical priors or hand-crafted models to address degradation caused by absorption, scattering, and turbidity [15]. Zhou et al. employed a modified underwater image formation model incorporating depth estimation and color correction to effectively mitigate the effects of light absorption and scattering [16]. Ma et al. proposed an improved Retinex-based variational model that integrates information entropy smoothing and non-uniform illumination priors, enabling effective handling of uneven lighting in underwater images [17]. Liu et al. introduced an illumination-constrained, structure-preserving Retinex model with adaptive channel compensation and joint estimation of illumination and reflection, demonstrating competitive performance on turbid underwater images in both subjective and objective evaluations [18]. Zhou et al. combine pixel distribution remapping with a Retinex variational model and noise-texture priors, achieving notable improvements in color correction and contrast enhancement [19].
Beyond physics-based models, some methods focus on enhancing images by improving contrast, correcting color, or reducing haze based on prior knowledge or statistical assumptions without explicitly modeling the underwater imaging process. Zhang et al. proposed a principal component fusion method (PCFB) that enhances underwater images by fusing contrast-enhanced foregrounds and dehazed backgrounds using principal component analysis [20]. Zhang et al. introduced a multi-channel adaptive fusion approach that addresses color distortion and contrast loss through adaptive channel correction and dual-branch enhancement [21]. To accommodate images across different color gamuts, Zhang et al. proposed the RAG-IMF method, which integrates global–local color correction and multi-channel fusion in both RGB and LAB color spaces, thereby extending color gamut and improving quality metrics [22]. Similarly, Jha et al. developed the CBLA method, which performs RGB-based color correction and LAB-based contrast and naturalness restoration for effective underwater enhancement [23]. To preserve fine details, recent works have also explored multi-scale representations and frequency-domain techniques, leading to more robust enhancement across a variety of underwater scenes [24,25].
While the above approaches offer strong interpretability, their performance tends to degrade in highly dynamic underwater environments, where the underlying model assumptions are frequently violated. As a result, they are generally less effective in handling complex, non-uniform degradations such as motion blur and spatially varying illumination.

2.3. Data-Driven Approach

With the rapid advancement of deep learning, data-driven approaches have emerged as a powerful alternative for UIE [26]. Unlike traditional methods, data-driven techniques can implicitly model complex nonlinear degradations and adapt to diverse underwater conditions through training on large-scale datasets. Xue et al. addressed the limitations of conventional color spaces by introducing a learnable Underwater Scenes Orient (USO) color space and a Scene-Adapted Semantic-Aggregated Degradation-Decoupling (S2D2) framework [27]. Park et al. proposed a lightweight enhancement framework based on an adaptive standardization and normalization network, which effectively corrects distorted feature distributions and improves image contrast and brightness, all while maintaining low computational complexity [28].
In parallel, Generative Adversarial Networks (GANs) [29,30], particularly diffusion models [31,32], have further improved the capacity of neural networks to learn complex mappings between degraded underwater images and their high-quality counterparts. These models have demonstrated significant improvements in addressing color distortion, motion blur, and low contrast [33,34]. The recent success of Transformers in computer vision has further accelerated progress in underwater image enhancement [35,36]. Their powerful global modeling capabilities and context-aware mechanisms significantly improve detail restoration and color correction, resulting in more natural and visually appealing enhancement outcomes. Yang et al. introduced a progressive aggregation framework that utilizes a feature-prompted Transformer, combining global-local attention with multi-scale feature aggregation, to enhance detail preservation, color fidelity, and blur removal [37]. Huang et al. proposed an underwater enhancement network incorporating a cross-wise Transformer module and a feature supplementation strategy to capture inter-stage dependencies and compensate for feature loss [38]. Moreover, techniques involving multi-domain feature extraction, physics-guided priors, and unsupervised learning have been actively explored to improve the perceptual and generalization abilities of enhancement networks [39,40,41,42].
Despite their promising results, data-driven methods also face notable limitations. Convolutional Neural Networks (CNNs) struggle to capture long-range dependencies due to their localized and static receptive fields. Although Transformers effectively address this issue, their high computational cost limits their practicality, especially for deployment on resource-constrained edge devices. GAN-based models, while exhibiting strong feature generation capabilities, often suffer from limited generalization across diverse underwater scenes. Furthermore, the aforementioned methods require large volumes of training data, yet datasets specifically addressing underwater blur remain scarce. Therefore, the models are prone to domain bias, making them less effective in handling complex and variable non-uniform underwater blurred images.

3. Methods

To address the challenges, we propose PMSPNet, a Progressive Multi-Scale Perception Network, as illustrated in Figure 2. The network incorporates a Hybrid Interaction Attention Module (HIAM) to capture local details and global contextual dependencies, enabling an initial coarse perception of image blur features. Building upon this, we introduce the Progressive Motion-Aware Perception Branch (PMAB) and the Progressive Feature Feedback Block (PFFB), which can incrementally guide and refine the network’s perceptual capability, thereby achieving accurate localization and representation of non-uniform blur regions. Furthermore, to facilitate more realistic and comprehensive training and evaluation, we construct a dedicated Non-uniform Underwater Blur Dataset (N2UD), which encompasses a wide range of blur patterns encountered in real-world underwater environments.

3.1. N2UD

To address the scarcity of underwater blur datasets, we constructed a specialized dataset by filtering an existing publicly available underwater image dataset. Specifically, we combined quantitative blur metrics with manual visual inspection to identify and select images exhibiting non-uniform blur. The resulting collection forms a new benchmark, termed the Non-uniform Underwater Blur Dataset (N2UD), which serves as a representative testbed for evaluating underwater deblurring algorithms under complex and realistic degradation conditions. As illustrated in Figure 3, N2UD includes underwater images affected by non-uniform blur across various resolutions and scenes, ensuring diversity and practical relevance.
The quantitative blur metrics of which can be expressed as
isBlurred = 1 δ U + δ C + δ G + δ H > 1 0 otherwise ,
where δ U denotes the underwater image sharpness measure, δ C represents the global contrast level, δ G is the mean image gradient, and δ H corresponds to the high-frequency energy. When IsBlurred equals 1, it signifies that the image exhibits non-uniform blurring. The underwater image sharpness measure δ U is defined as
δ U = 1 1 k 1 k 2 i = 1 k 1 j = 1 k 2 20 · log I max ( i j ) I min ( i j ) + ϵ < U t 0 otherwise ,
where the image is converted to the YUV color space and then partitioned into k 1   ×   k 2 non-overlapping blocks; I max ( i j ) and I min ( i j ) represent the maximum and minimum luminance values, respectively, for the ( i , j ) th block; ϵ is a small constant ( 10 6 ) to avoid division by zero; and U t is the corresponding threshold, set to 2.0. The δ C is expressed as
δ C = 1 σ ( I LAB ( l ) ) 255 < C t 0 otherwise ,
where I LAB ( l ) is the LAB color space of the input image I RGB , σ ( · ) is the standard deviation, and C t is the corresponding threshold, set to 0.15. The δ G is expressed as
δ G = 1 1 N x = 1 W y = 1 H M ( x , y ) < G t 0 otherwise ,
where N is the total number of pixels in the image; W ,   H are the width and height of the image, respectively; and G t is the corresponding threshold, set to 15. The M ( · ) is the gradient magnitude calculation for each pixel point, which is expressed as
M ( x , y ) = G x ( x , y ) 2 + G y ( x , y ) 2 ,
where G x ( · , · ) and G y ( · , · ) respectively perform Sobel filtering on the horizontal and vertical gradients of the grayscale image I gray of the original image I RGB . The δ H is expressed as
δ H = 1 x = 1 H y = 1 W | F HF ( x , y ) | x = 1 H y = 1 W | F ( x , y ) | < H t 0 otherwise ,
where H t is the corresponding threshold, set to 0.2. F HF ( · ) is expressed as
F HF ( x , y ) = F ( x , y ) · Mask ( x , y ) ,
where F ( x , y ) and Mask ( x , y ) are expressed as
F ( x , y ) = FS ( u = 0 H 1 v = 0 W 1 I gray ( u , v ) · e 2 ß i ( x u H + y v W ) ) ,
Mask ( x , y ) = 0 ( x c u ) 2 + ( y c v ) 2 r 1 otherwise ,
where FS ( · ) is a frequency shift that centers the zero-frequency (DC) component, and Mask ( x , y ) { 0 , 1 } H × W is a circular low-pass suppression mask centered at ( c u , c v ) = ( H 2 , W 2 ) , with radius r = min ( H , W ) 4 .
Considering that existing blur metrics may not fully capture the spatial variations of blur, we conducted manual visual inspections of each candidate image to refine the dataset. Based on evaluations by multiple experts specializing in visual perception, we retained only those images exhibiting noticeable non-uniform blur. This process ensures that the dataset accurately reflects the complex and spatially varying blur characteristics commonly observed in real-world underwater environments.

3.2. Hybrid Interaction Attention Module

To enhance the network’s capacity for deep feature extraction and precise localization of blurred regions in underwater scenes, we propose a Hybrid Interaction Attention Module (HIAM), as shown in Figure 4. Unlike traditional hybrid methods [43,44], HIAM, a cross-scale dual-channel design, utilizes an interactive cross-attention fusion dual-attention strategy for adaptive weighting control. This hybrid design enables collaborative modeling of localized blur features and global semantic structures, achieving coarse estimation of blurred feature extraction and localization.
Specifically, given an input feature I R C × H × W , the process is represented as
F a = Conv 3 × 3 ( L a ( I , S a ) + G a ( Q , K , I ) ) ,
where Conv 3 × 3 ( · ) indicates 3 × 3 convolution operation. The calculation process for other characters is as follows.
Firstly, we obtain Q , K , and V as follows:
Q , K , V = DSC ( Conv 1 × 1 ( Conv 3 × 3 ( I , I , I ) ) ) ,
where Conv 1 × 1 ( · ) indicates 1 × 1 convolution operation, and DSC ( · ) indicates depthwise separable convolution. Subsequently, Q , K , and V are obtained by
Q = ( P m ( DSC q ( Q ) ) + P a ( DSC q ( Q ) ) ) W Q ,
K = ( P m ( DSC k ( K ) ) + P a ( DSC k ( K ) ) ) W K ,
V = ( P m ( DSC v ( I ) ) + P a ( DSC v ( I ) ) ) W V ,
where P m ( · ) and P a ( · ) are adaptive global max pooling and average pooling, and W Q ,   W K ,   W V R C × d are trainable matrices. S a can be given by the following:
S a = σ sig ( σ soft ( Q ( K ) T d ) V ) ,
where σ sig ( · ) and σ soft ( · ) are Sigmoid and Softmax activation functions, and d is the dimension of the key vector. Thus, L a ( I , S a ) can be expressed as
L a ( I , S a ) = Conv 3 × 3 ( Q ( K ) T d ( V S a ) ) ,
where ⊙ is a pointwise multiplication operation. G a ( Q , K , I ) can be expressed as
G a ( Q , K , I ) = S a I + I .

3.3. Progressive Motion-Aware Perception Branch

To further enhance the modeling of aware features in underwater non-uniform blur scenarios, we introduce a Progressive Motion-Aware Perception Branch (PMPB), as depicted in the lower part of Figure 4. This module progressively refines the localization of non-uniform blur features in a coarse-to-fine manner, thereby strengthening the network’s multi-level perception of structural details.
Specifically, a two-dimensional Butterworth filter is applied to the input image to perform frequency adjustment. This operation amplifies the response of blurred regions across varying frequency components, improving the network’s sensitivity to related features [45,46]. The filtering process is defined as
I b = L 2 R ( Concat ( H hf ( L ) , A , B ) ) ,
where L 2 R ( · ) converts the image from LAB to RGB color space; Concat ( · ) is a channel splicing operation; and L = I LAB ( l ) , A = I LAB ( a ) , and B = I LAB ( b ) are the luminance and color channels in the LAB space, respectively. H hf ( · ) represents high-frequency extraction operations, expressed as
H hf ( L ) = ( L ˜ spatial ) ,
where ( · ) denotes the real part of a complex number. L ˜ spatial can be expressed as
L ˜ spatial = F 1 ( L ˜ s ) ,
where F 1 · is the inverse fast Fourier transform (IFFT). L ˜ s can be expressed as
L ˜ s = IS ( L ˜ ) ,
L ˜ = L ^ · ( 1 H bw ) ,
where IS ( · ) indicates that the centralized spectrum diagram will be restored to its original layout. L ^ C B × 1 × H × W can be expressed as
L ^ = FS ( F ( L ) ) ,
where F ( · ) is the fast Fourier transform (FFT).
The Butterworth filter H bw is constructed as
H bw ( u , v ) = 1 1 + ( D 0 ( u H w ) 2 + ( v W 2 ) 2 + ϵ ) 2 n ,
where D 0 is the cutoff frequency, n is the filter order, and ϵ is the anti-decimation constant, set to 10 6 .
Subsequently, the frequency domain enhanced image I b is subjected to initial feature extraction, which is denoted as
I i = σ re ( Conv 1 × 1 ( I b ) + Conv 3 × 3 ( σ re ( Conv 3 × 3 ( I b ) ) ) ) ,
where σ re ( · ) is the ReLU activation function.
Finally, we introduce Deformable Convolution [47] to adaptively adjust the sampling position for more accurate perception and feature extraction of the non-uniform blur region, which is expressed as
I o p 0 = k = 1 K w k · ( q N ( p 0 + p k + p k ) G ( q , p 0 + p k + p k ) · I i q ) ,
where I o p 0 is the value of the output feature map at position p 0 ; w k is the weight of the k th position of the convolution kernel; N is a bilinear interpolating neighborhood; G ( q , · ) is the interpolated weight, satisfying q G ( q , · ) = 1 ; p k is the k th standard sampling offset in the convolution kernel; p k is the learnable offset of the k th position; and I i q is the value of the input feature I i at position q.
The fine features are fed into the backbone to provide it with feature guidance, which is expressed as
F a = I o + Conv 3 × 3 ( Concat ( F a , I o ) ) .

3.4. Progressive Feature Feedback Block

To achieve lossless extraction of hierarchical semantic features during reconstruction, we propose a Progressive Feature Feedback Block (PFFB), as illustrated in Figure 4. This module facilitates inter-layer information interaction by constructing a hierarchical feedback pathway, allowing recovered features from subsequent stages to guide and refine the representation learning in previous stages. The structure of the PFFB is represented as
F f = σ re ( DSC 1 ( F att + F m ) ) ,
where F att can be expressed as
F att = σ soft ( MLP 1 , 2 ( P m ( F m ) + P m ( F m ) ) ) DSC 2 ( F up + F cu ) ,
where F m = DSC 3 ( F up + F cu ) , MLP 1 , 2 ( · ) is a multilayer perceptron, and F up and F cu are the previous and current level recovery features in the decoder, respectively.

3.5. Loss Function

To increase visual quality and perceptual fidelity of the restored image, we design a composite loss function that jointly enforces constraints on color accuracy, structural consistency, frequency response, and perceptual realism. This multi-objective formulation ensures that the restored images align closely with ground truth data and exhibit enhanced perceptual quality. The overall loss is defined as
L t = L charb + λ fft · L fft + λ lab · L lab + λ lch · L lch + L vgg + L color ,
where L charb , L fft , L lab , L lch , L vgg , and L color are charbonnier loss, fast fourier transform loss, LAB color loss, LCH color loss, perceptual loss, and color constancy loss, respectively. λ fft , λ lab , and λ lch are the corresponding weighting coefficients, set to 1 × 10 1 , 1 × 10 6 , and 1 × 10 2 , respectively. L charb is defined as
L charb ( x , y ) = 1 N i = 1 N ( x i y i ) 2 + ϵ 2 ,
where x and y are the predicted image and ground truth, and N is the total number of pixels in the image. L fft is defined as
L fft ( x , y ) = F ( x ) F ( y ) 1 ,
where | | · | | 1 is L1 loss. L lab is defined as
L lab ( x , y ) = L l ( x , y ) + α · L ab ( x , y ) ,
where α is the weighting coefficient, set to 1. L l ( · ) and L ab ( · ) are the brightness and color channel loss in the LAB channel, defined as
L l ( x , y ) = 1 N i = 1 N | L x ( i ) L y ( i ) | ,
L ab ( x , y ) = 1 N i = 1 N k = 1 N P y ( i , k ) · log ( P y ( i , k ) P x ( i , k ) + ϵ ) 1 N i = 1 N k = 1 N P y ( i , k ) · log ( P x ( i , k ) + ϵ ) ,
where P x ( i , k ) and P y ( i , k ) are the softmax weights for the kth bin at the ith pixel position of images x and y. L lch is defined as
L lch = L ll + L lc + L lh ,
where L ll and L lc represent brightness and chromaticity L1 loss (reference Equation (34)), and L lh represents hue distribution loss (reference Equation (35)) in LCH color space.
In addition, L vgg utilizes the multi-scale features of the VGG19 network to measure the differences between the enhanced image and the reference image in the high-level semantic space. L color achieves color balance consistency by minimizing the pairwise Euclidean distances between the R, G, and B channels.

4. Experiments

4.1. Datasets

We train and evaluate the proposed network using the constructed N2UD dataset and compare its performance against state-of-the-art methods on N2UD and three publicly available underwater image enhancement benchmarks: EUVP [48], LSUI [49], and UIEB [50]. Additionally, to assess the practical benefits of the enhanced images in real-world scenarios, we conduct downstream task evaluations on the DUO dataset [51], demonstrating the model’s effectiveness in supporting higher-level underwater vision tasks.

4.1.1. N2UD

The N2UD dataset comprises 3201 real underwater image pairs, each consisting of a non-uniformly blurred image and its corresponding reference. Among these, 2246 pairs were used for training, 624 for testing, and 322 for validation. The dataset consists of several publicly available underwater image enhancement datasets. Using a combination of quantitative blur metrics with manual visual inspection, we screened high-quality non-uniform blur image pairs. Specifically, the dataset includes 555 images from the EUVP dataset, 2190 from the LSUI dataset, and 465 from the UIEB dataset. The selected images span a range of resolutions, from 256 × 256 to 1280 × 720 , and diverse underwater environments and imaging conditions. The dataset is available at https://github.com/UI2025/N2UD, accessed on 28 August 2025.

4.1.2. EUVP

The EUVP dataset is designed to support the enhancement of color and structural details in real-world underwater images. It employs multiple underwater sensors (GoPros, Aqua AUV’s uEye cameras, low-light USB cameras, and Trident ROV’s HD camera) to capture underwater images and conduct marine exploration and human-robot cooperative experiments in different locations under various visibility conditions. It contains both paired and unpaired collections of approximately 20,000 images, covering a broad spectrum of underwater environments, lighting conditions, and viewpoints. The dataset includes both real and synthetically degraded underwater images at various resolutions, making it suitable for training and evaluating deep learning-based underwater image enhancement models. The dataset is available at https://irvlab.cs.umn.edu/resources/euvp-dataset, accessed on 17 May 2025.

4.1.3. LSUI

The LSUI dataset, filtered and screened based on various real underwater image data, is a large-scale collection of 4279 real-world underwater images captured across multiple scenes, including shallow waters, deep sea, coral reefs, and shipwrecks. It encompasses a wide variety of conditions in terms of lighting, water quality, and color degradation. LSUI emphasizes real data distributions and high scene complexity, making it well-suited for studying naturally occurring underwater degradations. The dataset is available at https://lintaopeng.github.io/_pages/UIE%20Project%20Page.html, accessed on 17 May 2025.

4.1.4. UIEB

The UIEB dataset collects and processes underwater images on different platforms and includes private underwater shooting videos. It is one of the most widely used benchmarks in underwater image enhancement. It includes 890 real-world underwater images of varying resolutions, most of which have been enhanced by domain experts to form a high-quality paired dataset. It is commonly used for both subjective and objective evaluation of enhancement algorithms. In addition, UIEB includes a challenging subset, Challenging-60, comprising 60 challenging unpaired samples used to test algorithmic robustness under extreme conditions. The dataset is available at https://li-chongyi.github.io/proj_benchmark.html, accessed on 17 May 2025.

4.1.5. DUO

The DUO dataset is a high-quality benchmark designed for underwater object detection (UOD) tasks. These data were obtained by collecting underwater images through underwater photography and labeling them, and they were used to study the perception decisions and sensor sensitivity issues of underwater robots. It contains 7782 images captured from diverse underwater environments, including shallow and deep water, turbid regions, and varied lighting conditions. The dataset features a wide array of target types, including starfish, sea urchins, and fish, along with substantial variation in scene structure, imaging quality, and target scale. All images are provided at a uniform resolution of 1920 × 1080 , with 6671 images allocated for training and 1111 for testing. In this work, we utilize the trained enhancement model to augment the DUO dataset and perform downstream detection tasks, thereby validating the positive influence of our method on higher-level vision applications. The dataset is available at https://github.com/chongweiliu/DUO, accessed on 18 May 2025.

4.2. Experimental Configuration

4.2.1. Implementation Details

The proposed network was implemented using the PyTorch 2.0.1 framework and trained on a Linux-based system equipped with a GeForce RTX 3090 GPU (24 GB), 250 GB of memory, and 80 Intel(R) Xeon(R) Gold 5218R processors (2.10 GHz). The training was performed using the Adam optimizer with an initial learning rate of 1 × 10 4 . A cosine learning rate scheduling strategy with a warm-up phase was employed, in which the learning rate was gradually increased during the first 10 epochs, after which it followed a cosine decay schedule. The network was trained for 200 epochs with a batch size of 8. To improve the model’s robustness and generalization capability, standard data augmentation techniques such as random horizontal and vertical flipping were applied during training. To prevent overfitting and ensure stable convergence, an early stopping mechanism was employed, in which training was terminated if the validation loss failed to improve for 10 consecutive epochs. Furthermore, random seeds were fixed across all experiments to guarantee reproducibility of the reported results.

4.2.2. Evaluation Metrics

To comprehensively assess the performance of the proposed method in restoring underwater images affected by non-uniform blur, we employed a diverse set of evaluation metrics covering full-reference, no-reference, perceptual, and sharpness-based aspects. For full-reference evaluation, we used the Peak Signal-to-Noise Ratio (PSNR), the Structural Similarity Index (SSIM), and the Feature Similarity Index (FSIM) to quantify pixel-level fidelity and structural consistency between the restored images and their ground-truth counterparts. To evaluate perceptual quality, we adopted LPIPS [52], which measures perceptual similarity using deep feature representations and aligns better with human visual perception.
For no-reference evaluation, we employed underwater-specific quality metrics, including UIQM [53] and UCIQE [54], to evaluate key visual attributes such as color fidelity, contrast, and clarity. Additionally, we introduced NIQE [55], a natural image quality evaluator, and URanker [56], a human preference-based perceptual scoring model, to further assess image naturalness and subjective quality. To evaluate the preservation of fine details and edge sharpness, we incorporated high-frequency sharpness metrics, including Laplacian variance, Tenengrad, and Brenner gradient, each capturing edge clarity and detail recovery from different computational perspectives.
The integration of these complementary metrics enables a comprehensive and objective evaluation of image enhancement performance, ensuring the model’s robustness and practical applicability in real-world underwater scenarios and downstream tasks.

4.3. Performance Comparison

4.3.1. N2UD

To comprehensively evaluate the effectiveness of the proposed method, we conducted extensive experiments on the constructed N2UD dataset using both full-reference and no-reference image quality assessment metrics. The results are presented in Table 1 and Table 2.
As shown in Table 1, traditional image processing methods exhibit generally poor performance across full-reference metrics. Specifically, their PSNR values typically fall below 18 dB, and LPIPS scores often exceed 0.3, indicating significant limitations in perceptual quality and visual fidelity. These methods struggle to adaptively model complex blur degradation, which severely limits their scalability in real-world applications. Furthermore, traditional methods are often computationally intensive and inefficient. Some require up to 1.24 s to process a single image, rendering them unsuitable for time-sensitive tasks. In contrast, deep learning-based UIE methods demonstrate more robust performance. GAN-based approaches offer modest improvements in some metrics. CycleGAN achieves relatively high SSIM ( 0.83 ± 0.11 ) and LPIPS ( 0.24 ± 0.01 ) compared to traditional methods. However, it exhibits considerable PSNR variance ( ± 4.72 ), highlighting a lack of consistency and reliability in generative adversarial models. U-Shape, leveraging a multi-scale architecture, performs well across all full-reference metrics but still falls short of the performance achieved by PMSPNet. Among recently proposed Transformer-based methods, HistFormer suffers from mode collapse issues, leading to degraded performance across multiple metrics. Additionally, it incurs a high computational cost (44.42 GFLOPs) and slower inference speeds. PhaseFormer, in contrast, shows a more balanced performance. However, its relatively low SSIM indicates limitations in preserving structural integrity. PUIE-Net achieves favorable results across most metrics, yet its complexity is a concern, with computational overhead reaching 150.69 GFLOPs and inference time extending to 0.13 s per image. SGUIE performs well in FSIM and LPIPS, reflecting strengths in perceptual quality. Nonetheless, its overall effectiveness remains inferior when compared to PMSPNet. PMSPNet achieves superior performance in both objective metrics and computational efficiency, confirming its robustness and practical applicability for real-world underwater deblurring tasks.
As shown in the no-reference image quality evaluation results in Table 2, although some methods perform reasonably well on individual metrics, a noticeable gap remains in overall quality. Notably, PMSPNet consistently leads across multiple perceptual quality metrics, such as UIQM and UCIQE, indicating that its enhanced images exhibit more natural color, clarity, and contrast, aligning well with human visual preferences. Interestingly, some traditional methods achieve higher scores on sharpness-based metrics, such as Tenengrad and Brenner. This contrasts with their overall inferior performance on full-reference and other no-reference metrics. The root cause of this “contradiction” is that many current reference-free metrics capture only specific local attributes, failing to reflect comprehensive image quality or perceptual realism.
The vision comparisons in Figure 5 further support this conclusion. While traditional methods can enhance local edge sharpness, they often suffer from over-sharpening, resulting in structural distortions, unnatural textures, or even pattern collapse, leading to a visually unrealistic “pseudo-enhancement” effect. In contrast, deep learning-based methods leverage strong feature representation and adaptive learning to restore degraded details more comprehensively. By capturing complex, non-uniform blur patterns across multiple scales and dimensions, PMSPNet achieves accurate regional restoration while preserving structural clarity and avoiding artifacts. This results in improved naturalness, semantic consistency, and stronger generalization in real-world conditions.
To further validate model robustness, Figure 6 presents the statistical distribution of three representative full-reference metrics. Traditional methods exhibit low and tightly clustered scores, indicating limited and inconsistent enhancement. Deep learning-based methods perform better overall. PMSPNet shows the highest median and maximum values, especially exceeding 35 dB in PSNR, and maintains stable, high distributions in SSIM and FSIM. The low dispersion of PMSPNet scores reflects its adaptability across various scenes, attributed to its hybrid interaction attention and progressive motion-aware modules, which jointly enhance spatially non-uniform blur modeling and semantic detail recovery.

4.3.2. EUVP

To further assess the generalization capability of the proposed network, we evaluated pre-trained models on the EUVP dataset. As shown in Table 3, PMSPNet achieves the best performance across nearly all full-reference metrics, including PSNR (25.81), SSIM (0.85), and FSIM (0.94), and the lowest LPIPS (0.21), demonstrating superior fidelity and perceptual quality. In no-reference perceptual assessments, PMSPNet remains competitive, achieving a UIQM of 3.09, a UCIQE of 0.36, and a URanker of 1.16, consistently outperforming some existing methods.
As illustrated in Figure 7, visual comparisons further substantiate PMSPNet’s advantages. Compared to prior methods, PMSPNet produces images with sharper edges, more natural color reproduction, and enhanced contrast. Traditional methods such as HLRP, ACDC, and MMLE suffer from color shifts, amplified noise, or loss of structural detail. Although deep learning methods like CycleGAN and U-Shape show improved enhancement, they still exhibit color deviations or insufficient deblurring in complex scenes. In contrast, PMSPNet effectively restores clarity and color fidelity even under severe degradation, delivering results that are perceptually closest to the ground truth.

4.3.3. LSUI

Table 4 presents the evaluation results on the LSUI dataset. PMSPNet achieves the highest scores across key full-reference metrics, along with the lowest LPIPS, indicating superior fidelity and perceptual quality in the restored images. In the no-reference evaluation, PMSPNet also ranks first in UIQM and second in UCIQE, further demonstrating its effectiveness in enhancing visual quality. In contrast, traditional methods and other deep learning-based approaches show reduced performance across most metrics due to their limited capacity to manage complex color degradation and structural blurring prevalent in underwater environments.
The visual results in Figure 8 further highlight the advantages of PMSPNet. Traditional algorithms often leave residual blur or introduce significant color distortions. While learning-based models such as U-Shape and CCL-Net show improved results, they still suffer from over-smoothing and color shifts, particularly in fine details. In comparison, PMSPNet effectively recovers key textures, preserves natural color tones, and enhances contrast, producing visually more realistic and aesthetically pleasing results.

4.3.4. UIEB

The UIEB dataset presents a broad spectrum of underwater imaging challenges, including severe color distortion, low contrast, and complex lighting conditions. As shown in Table 5, PMSPNet outperforms most compared methods across full-reference and no-reference metrics. Specifically, it achieves a PSNR of 22.43 dB, SSIM of 0.86, and a low LPIPS of 0.21, reflecting strong reconstruction accuracy and perceptual similarity. While methods like CCL-Net and SGUIE yield competitive results, their higher variance in metrics such as URanker and UIQM suggests inconsistent performance. In contrast, PMSPNet delivers high-quality results, demonstrating robustness across varying underwater degradation types.
As illustrated in Figure 9, PMSPNet excels at restoring fine structures and maintaining natural color tones, even in severely degraded scenarios. Compared to methods like UIR-PolyKernel and HistFormer, which often produce over-smoothed textures or color shifts, PMSPNet effectively reconstructs edges and preserves the texture of marine organisms. It also avoids common artifacts, such as excessive blue saturation or overly enhanced contrast, frequently observed in other approaches. This balance across metrics highlights the visual fidelity and realism of PMSPNet, making it both quantitatively superior and perceptually compelling.
In summary, the above experiments demonstrate that PMSPNet can effectively remove underwater non-uniform blur, generating clearer structures, more natural colors, and fewer artifacts and exhibiting strong generalization across diverse underwater scenes. Compared to both traditional and deep learning-based baselines, PMSPNet achieves an optimal trade-off between detail preservation and deblurring, especially under challenging conditions, validating the effectiveness and robustness of the proposed approach.

4.4. Ablation Study

To validate the effectiveness of the proposed components in PMSPNet, we conduct comprehensive ablation experiments by removing key modules and evaluating the impact of different loss function combinations. Furthermore, we assess the scalability of the proposed algorithm in downstream tasks.

4.4.1. Butterworth Filter

We introduce a Butterworth filter in the input stage of the PMPB to improve the frequency correspondence and mitigate the aliasing phenomenon. As shown in Table 6, although removing the module results in a slight increase in PSNR by 0.47, all other evaluation metrics exhibit a decline. In particular, SSIM drops to 0.88, UIQM decreases to 3.08, and UCIQE falls to 0.39. These results underscore the module’s effectiveness in enhancing structural clarity and perceptual quality as perceived by the human visual system. As illustrated in Figure 10b, the absence of this module leads to noticeable blurring and color artifacts, particularly around object edges and in background regions. This further confirms the module’s critical role in improving the overall structural integrity of the image and enhancing the network’s sensitivity to spatially non-uniform blur.

4.4.2. Deformable Convolution

The high dynamics of deformable convolution enable precise localization and directional modeling of non-uniform blur, allowing the network to effectively capture regionally continuous blur patterns and apply accurate weighting to key features. As shown in Table 6, the removal of deformable convolution results in a substantial decline across all evaluation metrics, with PSNR dropping sharply to 17.18, indicating a significant reduction in reconstruction quality. Furthermore, the network’s overall stability is noticeably compromised. As illustrated in Figure 10c, the network without this module struggles to preserve object boundaries and spatial consistency, and the residual blur remains largely unaddressed. These results demonstrate that deformable convolution significantly enhances the network’s capacity to identify critical image features and improves the modeling of local structures, which is essential for effective UIE under non-uniform degradation conditions.

4.4.3. Progressive Motion-Aware Perception Branch

The PMPB integrates a Butterworth filter and deformable convolution to progressively capture the spatial distribution of blur and guide feature extraction in the backbone, enabling a coarse-to-fine refinement of blur localization. As shown in Table 6, while the removal of PMPB results in a slight increase in PSNR, it reduces the overall stability of the network. This observation also validates the complementary roles of HIAM and PMPB: HIAM facilitates multi-directional feature perception and coarse localization of blur, whereas PMPB further refines this localization, enhancing the network’s generalization and robustness. Additionally, the observed decline in metrics such as SSIM confirms PMPB’s contribution to structural clarity and perceptual quality. As illustrated in Figure 10d, the absence of PMPB increases color distortion and reduces image sharpness, particularly in fine structures. These results underscore the importance of PMPB in preserving texture details and scale-aware representations, achieving a critical balance between global semantic understanding and local detail restoration.

4.4.4. Progressive Feature Feedback Block

The PFFB hierarchically integrates multi-level features by facilitating cross-layer feedback, enabling the fusion of spatial details with deep semantic representations. As shown in Table 6, the removal of this module leads to increased information loss during image reconstruction, which diminishes the network’s adaptive reconstruction capability, resulting in a consistent, albeit modest, degradation across all evaluation metrics. Figure 10e reveals noticeable artifacts and diminished smoothness, particularly along background contours, indicating the loss of critical structural details. These results demonstrate that PFFB plays a vital role in regulating cross-layer information flow, guiding adaptive feature reweighting, and reducing reconstruction-induced information loss. Consequently, it contributes significantly to improving structural fidelity and perceptual consistency in enhanced underwater images.

4.4.5. Loss Function

We investigated the impact of different loss components on the overall performance of the proposed network. As illustrated in Table 7, removing most auxiliary losses leads to a notable decline across multiple evaluation metrics. Among these, the FFT, LAB, and LCH loss functions contribute most significantly to performance improvement, enhancing objective image quality and perceptual fidelity. In particular, the inclusion of the LAB loss enhances network stability, underscoring the importance of multi-domain color perception. While the VGG and Color losses did not yield substantial improvements in numerical metrics, they contributed positively to visual quality by enhancing hue perception and overall appearance. Figure 10f–k demonstrates that models trained without specific loss terms suffer from excessive smoothing or unnatural color shifts. In contrast, the complete loss configuration consistently generates underwater images with sharper edges, more balanced color tones, and enhanced visual appeal, validating the effectiveness of each component.

4.4.6. Blur Perceptual Localization

To further investigate the contribution of components to blur perception, we visualize the heatmaps generated under various ablation settings in Figure 11, using the second column of Figure 1 as a representative example. When the Butterworth filter is removed, the heatmap exhibits higher noise levels, indicating a weakened ability to extract key features and accurately capture blur-related information. The removal of deformable convolution and the PFFB markedly impairs the network’s capacity for accurately identifying and localizing blurred regions. These components are essential for enabling coarse-to-fine blur localization and iterative attention modulation. Eliminating the PMPB module causes the network’s attention to drift while demonstrating the capability of HIAM in achieving coarse localization of blur features.
From the perspective of loss functions, removing the FFT loss impairs the network’s ability to capture frequency-domain information, thereby diminishing its capacity to identify and localize blur across multiple spatial domains. VGG, LAB, and LCH losses contribute to guiding the network’s attention toward perceptually and semantically important regions, facilitating more accurate blur localization. While Color loss has a relatively smaller effect on network performance, it plays a crucial role in suppressing over-sharpening, thereby enhancing visual naturalness.
In contrast, the proposed model exhibits the most concentrated and semantically accurate activation responses, highlighting blurred regions. These results validate the effectiveness of our architectural design and multi-component loss in improving the network’s capability for blur perception and structural fidelity restoration in underwater environments.

4.4.7. Downstream Task Evaluation

To further validate the practical benefits of the proposed method, we conducted evaluations on downstream tasks, including classic feature extraction and underwater object detection (UOD).
As illustrated in Figure 12, we applied SIFT keypoint detection and Canny edge detection to assess the influence of enhancement on low-level visual features. The first row shows the SIFT results, while the second row displays the Canny edges. Traditional methods often fail to preserve sufficient structural detail, introducing noise and blurring in non-salient regions, which results in some key edges being obscured or missing. In contrast, images enhanced by PMSPNet exhibit clear contours and rich edge information, closely matching the ground truth, demonstrating that PMSPNet effectively enhances geometric clarity, facilitating better feature localization.
To assess high-level perception, we applied YOLOv12 [71] for underwater object detection. A YOLOv12 model pre-trained on the original DUO dataset was used to detect objects in images enhanced by different methods. As shown in Figure 13, traditional enhancement approaches often introduce significant color distortions, which impair detection performance. Deep learning-based baselines exhibit limited performance in small object detection, primarily due to underperformance in detail preservation. In contrast, PMSPNet-enhanced images have more accurate and stable detection outcomes, achieving higher confidence scores and better alignment with the original detection results. Notably, because YOLOv12 was trained on the original images, detection performance is best in raw images.
The above results demonstrate that PMSPNet not only improves perceptual image quality but also preserves semantic fidelity, making it well-suited for integration into real-world underwater robotic systems and visual monitoring pipelines.

4.4.8. Limitation Analysis

Although PMSPNet demonstrates superior performance in non-uniform underwater image deblurring, several limitations remain. Due to the incorporation of multi-scale perception and progressive feedback mechanisms, the model introduces additional computational overhead, which may restrict its real-time deployment on extremely resource-constrained underwater robotic platforms. While the proposed N2UD dataset covers diverse non-uniform blur patterns, it still cannot fully represent the wide spectrum of degradations in real-world underwater environments, such as extreme turbidity, lighting fluctuations, or dynamic background interference. While we validated the universality of PMSPNet on downstream tasks such as detection, edge detection, and keypoint localization, comprehensive evaluations on segmentation tasks remain limited. In addition, this study primarily focuses on algorithm-level comparisons. Integrating advanced hardware-oriented solutions, such as novel imaging sensors or optical acquisition systems, represents another promising research direction.
Future work will focus on addressing these issues by optimizing the network components to improve their flexibility, enabling lightweight and real-time deployment. Moreover, we will expand the dataset with more challenging real-world scenarios and extend the validation to segmentation tasks.

5. Conclusions

This article proposes PMSPNet, a Progressive Multi-Scale Perception Network designed for the challenging task of non-uniform underwater image deblurring. PMSPNet incorporates a Hybrid Interaction Attention Module (HIAM) to effectively capture fine-grained visual textures and long-range contextual dependencies, enabling robust modeling of feature ambiguity and spatial disparity. Furthermore, the Progressive Motion-Aware Perception Branch (PMPB) is introduced to explicitly represent spatial orientation variations and progressively refine the localization of blur-affected regions. In addition, the Progressive Feature Feedback Block (PFFB) enhances feature reconstruction by leveraging multi-level information in a feedback manner, improving feature restoration quality. To enable reliable evaluation, we construct the N2UD dataset, which contains diverse non-uniform blur patterns representative of real-world underwater environments. Extensive experiments on real-world datasets validate the superiority of our method in terms of both quantitative metrics and visual quality. While achieving an inference speed of 0.01 s, it reached the highest PSNR of 25.51 dB and SSIM of 0.92 on the N2UD dataset. PMSPNet also demonstrates clear advantages in downstream tasks such as edge detection and object recognition, which enhances the visual reliability of underwater perception systems, facilitating downstream tasks such as object detection, recognition, and navigation in robotic platforms. Future efforts will focus on extending PMSPNet for real-time deployment in underwater robotic systems. Moreover, extending PMSPNet to handle real-time underwater video restoration and multimodal fusion with sonar or depth data remains a promising direction. In addition, diffusion-based generative models, despite their current limitations in computational efficiency, offer great potential for modeling complex blur patterns and realistic underwater degradations. Integrating the strengths of diffusion models into underwater deblurring frameworks will be an important avenue for our future research.

Author Contributions

Conceptualization, D.K.; methodology, D.K. and Y.Z.; software, D.K. and Y.Z.; validation, X.Z. and Y.W. (Yanyan Wang); formal analysis, D.K. and Y.Z.; investigation, D.K. and Y.W. (Yanyan Wang); resources, D.K. and X.Z.; data curation, Y.Z. and Y.W. (Yanqiang Wang); writing—original draft preparation, D.K. and Y.Z.; writing—review and editing, D.K. and Y.Z.; visualization, X.Z. and Y.W. (Yanqiang Wang); supervision, X.Z.; project administration, D.K. and Y.Z.; funding acquisition, D.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Project of Henan Province, grant numbers 242102211025, 252102211059, and 252102210114.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found at https://github.com/UI2025/N2UD (N2UD), accessed on 28 August 2025; https://irvlab.cs.umn.edu/resources/euvp-dataset (EUVP), accessed on 17 May 2025; https://lintaopeng.github.io/_pages/UIE%20Project%20Page.html (LSUI), accessed on 17 May 2025; https://li-chongyi.github.io/proj_benchmark.html (UIEB), accessed on 17 May 2025; and https://github.com/chongweiliu/DUO (DUO), accessed on 18 May 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, J.; Liu, C.; Long, B.; Zhang, D.; Jiang, Q.; Muhammad, G. Degradation-Decoupling Vision Enhancement for Intelligent Underwater Robot Vision Perception System. IEEE Internet Things J. 2025, 12, 17880–17895. [Google Scholar] [CrossRef]
  2. Sun, L.; Wang, Y.; Hui, X.; Ma, X.; Bai, X.; Tan, M. Underwater Robots and Key Technologies for Operation Control. Cyborg Bionic Syst. 2024, 5, 0089. [Google Scholar] [CrossRef]
  3. González-Sabbagh, S.P.; Robles-Kelly, A. A Survey on Underwater Computer Vision. ACM Comput. Surv. 2023, 55, 1–39. [Google Scholar] [CrossRef]
  4. Shen, L.; Reda, M.; Zhang, X.; Zhao, Y.; Kong, S.G. Polarization-Driven Solution for Mitigating Scattering and Uneven Illumination in Underwater Imagery. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–15. [Google Scholar] [CrossRef]
  5. Li, Y.; Chen, Y.; Zhang, J.; Li, Y.; Fu, X. An Underwater Image Restoration Method With Polarization Imaging Optimization Model for Poor Visible Conditions. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 3924–3939. [Google Scholar] [CrossRef]
  6. Cai, W.; Zhu, J.; Zhang, M. Multi-modality object detection with sonar and underwater camera via object-shadow feature generation and saliency information. Expert Syst. Appl. 2025, 287, 128021. [Google Scholar] [CrossRef]
  7. Wu, H.; Liu, Z.; Li, C.; Wang, H.; Zhai, Y.; Dong, L. A laser field synchronous scanning imaging system for underwater long-range detection. Opt. Laser Technol. 2024, 175, 110849. [Google Scholar] [CrossRef]
  8. Xu, S.; Zhang, K.; Wang, S. AQUA-SLAM: Tightly Coupled Underwater Acoustic-Visual-Inertial SLAM With Sensor Calibration. IEEE Trans. Robot. 2025, 41, 2785–2803. [Google Scholar] [CrossRef]
  9. He, Y.; Han, G.; Hou, Y.; Lin, C. Environment-Tolerant Trust Opportunity Routing Based on Reinforcement Learning for Internet of Underwater Things. IEEE Trans. Mob. Comput. 2025, 24, 6348–6360. [Google Scholar] [CrossRef]
  10. Xu, C.; Song, S.; Liu, J.; Pan, M.; Xu, G.; Cui, J.H. Joint Power Control and Multipath Routing for Internet of Underwater Things in Varying Environments. IEEE Internet Things J. 2025, 12, 15197–15210. [Google Scholar] [CrossRef]
  11. Jiang, B.; Feng, J.; Cui, X.; Wang, J.; Liu, Y.; Song, H. Security and Reliability of InternSecurity and Reliability of Internet of Underwater Things: Architecture, Challenges, and Opportunities. ACM Comput. Surv. 2024, 57, 1–37. [Google Scholar] [CrossRef]
  12. Draz, U.; Ali, T.; Yasin, S.; Chaudary, M.H.; Yasin, I.; Ayaz, M.; Aggoune, E.H.M. Hybrid Underwater Localization Communication Framework for Blockchain-Enabled IoT Underwater Acoustic Sensor Network. IEEE Internet Things J. 2025, 12, 16858–16885. [Google Scholar] [CrossRef]
  13. Rupa, C.; Varshitha, G.S.; Divya, D.; Gadekallu, T.R.; Srivastava, G. A Novel and Robust Authentication Protocol for Secure Underwater Communication Systems. IEEE Internet Things J. 2025. [Google Scholar] [CrossRef]
  14. Zhu, R.; Boukerche, A.; Yang, Q. An Efficient Secure and Adaptive Routing Protocol Based on GMM-HMM-LSTM for Internet of Underwater Things. IEEE Internet Things J. 2024, 11, 16491–16504. [Google Scholar] [CrossRef]
  15. Chandrasekar, A.; Sreenivas, M.; Biswas, S. PhISH-Net: Physics Inspired System for High Resolution Underwater Image Enhancement. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 1506–1516. [Google Scholar]
  16. Zhou, J.; Yang, T.; Ren, W.; Zhang, D.; Zhang, W. Underwater image restoration via depth map and illumination estimation based on a single image. Opt. Express 2021, 29, 29864–29886. [Google Scholar] [CrossRef]
  17. Ma, H.; Huang, J.; Shen, C.; Jiang, Z. Retinex-inspired underwater image enhancement with information entropy smoothing and non-uniform illumination priors. Pattern Recognit. 2025, 162, 111411. [Google Scholar] [CrossRef]
  18. Liu, S.; Zheng, Y.; Li, J.; Lu, H.; An, D.; Shen, Z.; Wang, Z. Turbid Underwater Image Enhancement with Illumination-Constrained and Structure-Preserved Retinex Model. IEEE Trans. Circuits Syst. Video Technol. 2025. [Google Scholar] [CrossRef]
  19. Zhou, J.; Wang, S.; Lin, Z.; Jiang, Q.; Sohel, F. A Pixel Distribution Remapping and Multi-Prior Retinex Variational Model for Underwater Image Enhancement. IEEE Trans. Multimed. 2024, 26, 7838–7849. [Google Scholar] [CrossRef]
  20. Zhang, W.; Liu, Q.; Feng, Y.; Cai, L.; Zhuang, P. Underwater Image Enhancement via Principal Component Fusion of Foreground and Background. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 10930–10943. [Google Scholar] [CrossRef]
  21. Qiang, H.; Zhong, Y.; Zhu, Y.; Zhong, X.; Xiao, Q.; Dian, S. Underwater Image Enhancement Based on Multichannel Adaptive Compensation. IEEE Trans. Instrum. Meas. 2024, 73, 1–10. [Google Scholar] [CrossRef]
  22. Zhang, T.; Su, H.; Fan, B.; Yang, N.; Zhong, S.; Yin, J. Underwater Image Enhancement Based on Red Channel Correction and Improved Multiscale Fusion. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–20. [Google Scholar] [CrossRef]
  23. Jha, M.; Bhandari, A.K. CBLA: Color-Balanced Locally Adjustable Underwater Image Enhancement. IEEE Trans. Instrum. Meas. 2024, 73, 1–11. [Google Scholar] [CrossRef]
  24. Wang, Y.; Hu, S.; Yin, S.; Deng, Z.; Yang, Y.H. A multi-level wavelet-based underwater image enhancement network with color compensation prior. Expert Syst. Appl. 2024, 242, 122710. [Google Scholar] [CrossRef]
  25. Zhang, W.; Liu, Q.; Lu, H.; Wang, J.; Liang, J. Underwater Image Enhancement via Wavelet Decomposition Fusion of Advantage Contrast. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 7807–7820. [Google Scholar] [CrossRef]
  26. Kong, D.; Zhang, Y.; Zhao, X.; Wang, Y.; Cai, L. MUFFNet: Lightweight dynamic underwater image enhancement network based on multi-scale frequency. Front. Mar. Sci. 2025, 12, 1541265. [Google Scholar] [CrossRef]
  27. Xue, X.; Yuan, J.; Ma, T.; Ma, L.; Jia, Q.; Zhou, J.; Wang, Y. Degradation-Decoupled and semantic-aggregated cross-space fusion for underwater image enhancement. Inf. Fusion 2025, 118, 102927. [Google Scholar] [CrossRef]
  28. Park, C.W.; Eom, I.K. Underwater image enhancement using adaptive standardization and normalization networks. Eng. Appl. Artif. Intell. 2024, 127, 107445. [Google Scholar] [CrossRef]
  29. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  30. Kong, D.; Mao, J.; Zhang, Y.; Zhao, X.; Wang, Y.; Wang, S. Dual-Domain Adaptive Synergy GAN for Enhancing Low-Light Underwater Images. J. Mar. Sci. Eng. 2025, 13, 1092. [Google Scholar] [CrossRef]
  31. Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. In Proceedings of the Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 6840–6851. [Google Scholar]
  32. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-Resolution Image Synthesis with Latent Diffusion Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 10684–10695. [Google Scholar]
  33. Li, W.; Wu, X.; Fan, S.; Wei, S.; Gowing, G. INGC-GAN: An Implicit Neural-Guided Cycle Generative Approach for Perceptual-Friendly Underwater Image Enhancement. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 10084–10098. [Google Scholar] [CrossRef] [PubMed]
  34. Qing, Y.; Liu, S.; Wang, H.; Wang, Y. DiffUIE: Learning Latent Global Priors in Diffusion Models for Underwater Image Enhancement. IEEE Trans. Multimed. 2025, 27, 2516–2529. [Google Scholar] [CrossRef]
  35. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  36. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar] [CrossRef]
  37. Yang, J.; Zhu, S.; Liang, H.; Bai, S.; Jiang, F.; Hussain, A. PAFPT: Progressive aggregator with feature prompted transformer for underwater image enhancement. Expert Syst. Appl. 2025, 262, 125539. [Google Scholar] [CrossRef]
  38. Huang, Z.; Wang, X.; Xu, C.; Li, J.; Feng, L. Underwater variable zoom: Depth-guided perception network for underwater image enhancement. Expert Syst. Appl. 2025, 259, 125350. [Google Scholar] [CrossRef]
  39. Song, J.; Xu, H.; Jiang, G.; Yu, M.; Chen, Y.; Luo, T.; Song, Y. Frequency domain-based latent diffusion model for underwater image enhancement. Pattern Recognit. 2025, 160, 111198. [Google Scholar] [CrossRef]
  40. Yin, J.; Wang, Y.; Guan, B.; Zeng, X.; Guo, L. Unsupervised Underwater Image Enhancement Based on Disentangled Representations via Double-Order Contrastive Loss. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–15. [Google Scholar] [CrossRef]
  41. Liu, X.; Jiang, Y.; Wang, Y.; Liu, T.; Wang, J. MDA-Net: A Multidistribution Aware Network for Underwater Image Enhancement. IEEE Trans. Geosci. Remote Sens. 2025, 63, 1–13. [Google Scholar] [CrossRef]
  42. Liang, D.; Chu, J.; Cui, Y.; Zhai, Z.; Wang, D. NPT-UL: An Underwater Image Enhancement Framework Based on Nonphysical Transformation and Unsupervised Learning. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–19. [Google Scholar] [CrossRef]
  43. Guo, X.; Chen, X.; Wang, S.; Pun, C.M. Underwater Image Restoration Through a Prior Guided Hybrid Sense Approach and Extensive Benchmark Analysis. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 4784–4800. [Google Scholar] [CrossRef]
  44. Ji, X.; Chen, S.; Hao, L.Y.; Zhou, J.; Chen, L. FBDPN: CNN-Transformer hybrid feature boosting and differential pyramid network for underwater object detection. Expert Syst. Appl. 2024, 256, 124978. [Google Scholar] [CrossRef]
  45. Shalini, J.; Ashok Kumar, L. An explainable artificial intelligence driven fall system for sensor data analysis enhanced by butterworth filtering. Eng. Appl. Artif. Intell. 2025, 158, 111364. [Google Scholar] [CrossRef]
  46. Kurinjimalar, R.; Pradeep, J.; Harikrishnan, M. Underwater Image Enhancement Using Gaussian Pyramid, Laplacian Pyramid and Contrast Limited Adaptive Histogram Equalization. In Proceedings of the 2024 IEEE 3rd World Conference on Applied Intelligence and Computing (AIC), Gwalior, India, 27–28 July 2024; pp. 729–734. [Google Scholar] [CrossRef]
  47. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  48. Islam, M.J.; Xia, Y.; Sattar, J. Fast Underwater Image Enhancement for Improved Visual Perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
  49. Peng, L.; Zhu, C.; Bian, L. U-Shape Transformer for Underwater Image Enhancement. IEEE Trans. Image Process. 2023, 32, 3066–3079. [Google Scholar] [CrossRef]
  50. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An Underwater Image Enhancement Benchmark Dataset and Beyond. IEEE Trans. Image Process. 2020, 29, 4376–4389. [Google Scholar] [CrossRef] [PubMed]
  51. Liu, C.; Li, H.; Wang, S.; Zhu, M.; Wang, D.; Fan, X.; Wang, Z. A Dataset and Benchmark of Underwater Object Detection for Robot Picking. In Proceedings of the 2021 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar] [CrossRef]
  52. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  53. Panetta, K.; Gao, C.; Agaian, S. Human-Visual-System-Inspired Underwater Image Quality Measures. IEEE J. Ocean. Eng. 2016, 41, 541–551. [Google Scholar] [CrossRef]
  54. Yang, M.; Sowmya, A. An Underwater Color Image Quality Evaluation Metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef] [PubMed]
  55. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2013, 20, 209–212. [Google Scholar] [CrossRef]
  56. Guo, C.; Wu, R.; Jin, X.; Han, L.; Zhang, W.; Chai, Z.; Li, C. Underwater Ranker: Learn Which Is Better and How to Be Better. Proc. AAAI Conf. Artif. Intell. 2023, 37, 702–709. [Google Scholar] [CrossRef]
  57. Zhang, W.; Zhou, L.; Zhuang, P.; Li, G.; Pan, X.; Zhao, W.; Li, C. Underwater Image Enhancement via Weighted Wavelet Visual Perception Fusion. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 2469–2483. [Google Scholar] [CrossRef]
  58. An, S.; Xu, L.; Deng, Z.; Zhang, H. HFM: A hybrid fusion method for underwater image enhancement. Eng. Appl. Artif. Intell. 2024, 127, 107219. [Google Scholar] [CrossRef]
  59. Zhuang, P.; Wu, J.; Porikli, F.; Li, C. Underwater Image Enhancement with Hyper-Laplacian Reflectance Priors. IEEE Trans. Image Process. 2022, 31, 5442–5455. [Google Scholar] [CrossRef]
  60. Zhang, W.; Wang, Y.; Li, C. Underwater Image Enhancement by Attenuated Color Channel Correction and Detail Preserved Contrast Enhancement. IEEE J. Ocean. Eng. 2022, 47, 718–735. [Google Scholar] [CrossRef]
  61. Zhang, W.; Zhuang, P.; Sun, H.H.; Li, G.; Kwong, S.; Li, C. Underwater Image Enhancement via Minimal Color Loss and Locally Adaptive Contrast Enhancement. IEEE Trans. Image Process. 2022, 31, 3997–4010. [Google Scholar] [CrossRef]
  62. Zhang, W.; Jin, S.; Zhuang, P.; Liang, Z.; Li, C. Underwater Image Enhancement via Piecewise Color Correction and Dual Prior Optimized Contrast Enhancement. IEEE Signal Process. Lett. 2023, 30, 229–233. [Google Scholar] [CrossRef]
  63. Yuan, J.; Cai, Z.; Cao, W. TEBCF: Real-World Underwater Image Texture Enhancement Model Based on Blurriness and Color Fusion. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  64. Peng, Y.T.; Chen, Y.R.; Chen, G.R.; Liao, C.J. Histoformer: Histogram-Based Transformer for Efficient Underwater Image Enhancement. IEEE J. Ocean. Eng. 2025, 50, 164–177. [Google Scholar] [CrossRef]
  65. Khan, M.R.; Negi, A.; Kulkarni, A.; Phutke, S.S.; Vipparthi, S.K.; Murala, S. Phaseformer: Phase-Based Attention Mechanism for Underwater Image Restoration and Beyond. In Proceedings of the 2025 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Tucson, AZ, USA, 28 February–4 March 2025; pp. 9618–9629. [Google Scholar] [CrossRef]
  66. Guo, X.; Dong, Y.; Chen, X.; Chen, W.; Li, Z.; Zheng, F.; Pun, C.M. Underwater Image Restoration via Polymorphic Large Kernel CNNs. In Proceedings of the ICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India, 6–11 April 2025; pp. 1–5. [Google Scholar] [CrossRef]
  67. Liu, Y.; Jiang, Q.; Wang, X.; Luo, T.; Zhou, J. Underwater Image Enhancement with Cascaded Contrastive Learning. IEEE Trans. Multimed. 2025, 27, 1512–1525. [Google Scholar] [CrossRef]
  68. Fu, Z.; Wang, W.; Huang, Y.; Ding, X.; Ma, K.K. Uncertainty Inspired Underwater Image Enhancement. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Proceedings, Part XVIII; Springer: Berlin/Heidelberg, Germany, 2022; pp. 465–482. [Google Scholar] [CrossRef]
  69. Fu, Z.; Lin, H.; Yang, Y.; Chai, S.; Sun, L.; Huang, Y.; Ding, X. Unsupervised Underwater Image Restoration: From a Homology Perspective. Proc. AAAI Conf. Artif. Intell. 2022, 36, 643–651. [Google Scholar] [CrossRef]
  70. Qi, Q.; Li, K.; Zheng, H.; Gao, X.; Hou, G.; Sun, K. SGUIE-Net: Semantic Attention Guided Underwater Image Enhancement with Multi-Scale Perception. IEEE Trans. Image Process. 2022, 31, 6816–6830. [Google Scholar] [CrossRef] [PubMed]
  71. Tian, Y.; Ye, Q.; Doermann, D. YOLOv12: Attention-Centric Real-Time Object Detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
Figure 1. Illustration of non-uniform blur in underwater images. (i) The raw underwater images. (ii) The estimated blur trajectories, where the arrow directions and positions represent the motion blur vectors. (iii) The corresponding blur intensity heatmaps, where warmer colors indicate stronger blur magnitude.
Figure 1. Illustration of non-uniform blur in underwater images. (i) The raw underwater images. (ii) The estimated blur trajectories, where the arrow directions and positions represent the motion blur vectors. (iii) The corresponding blur intensity heatmaps, where warmer colors indicate stronger blur magnitude.
Sensors 25 05439 g001
Figure 2. PMSPNet network architecture. Core components include the Hybrid Interaction Attention Module (HIAM), Progressive Motion-Aware Perception Branch (PMPB), and Progressive Feature Feedback Block (PFFB). The data come from the Non-uniform Underwater Blur Dataset (N2UD).
Figure 2. PMSPNet network architecture. Core components include the Hybrid Interaction Attention Module (HIAM), Progressive Motion-Aware Perception Branch (PMPB), and Progressive Feature Feedback Block (PFFB). The data come from the Non-uniform Underwater Blur Dataset (N2UD).
Sensors 25 05439 g002
Figure 3. The N2UD dataset, which contains various underwater environments and different degrees of non-uniform blur.
Figure 3. The N2UD dataset, which contains various underwater environments and different degrees of non-uniform blur.
Sensors 25 05439 g003
Figure 4. PMSPNet network components, including HIAM, PMPB, and PFFB.
Figure 4. PMSPNet network components, including HIAM, PMPB, and PFFB.
Sensors 25 05439 g004
Figure 5. Visual comparison on the N2UD dataset, with the PSNR value of the image shown in the upper-right corner of the image. (a) Raw images. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Figure 5. Visual comparison on the N2UD dataset, with the PSNR value of the image shown in the upper-right corner of the image. (a) Raw images. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Sensors 25 05439 g005
Figure 6. Box plot comparison of PSNR, SSIM, and FSIM on the N2UD dataset.
Figure 6. Box plot comparison of PSNR, SSIM, and FSIM on the N2UD dataset.
Sensors 25 05439 g006
Figure 7. Visual comparison on the EUVP dataset, with the PSNR value of the image shown in the upper-right corner of the image. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Figure 7. Visual comparison on the EUVP dataset, with the PSNR value of the image shown in the upper-right corner of the image. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Sensors 25 05439 g007
Figure 8. Visual comparison on the LSUI dataset, with the PSNR value of the image shown in the upper-right corner of the image. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Figure 8. Visual comparison on the LSUI dataset, with the PSNR value of the image shown in the upper-right corner of the image. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Sensors 25 05439 g008
Figure 9. Visual comparison on the UIEB dataset, with the PSNR value of the image shown in the upper-right corner of the image. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Figure 9. Visual comparison on the UIEB dataset, with the PSNR value of the image shown in the upper-right corner of the image. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Sensors 25 05439 g009
Figure 10. Visual comparison of networks using different components and loss functions. PSNR is shown in the top-right, and the bottom-left presents a zoomed-in view of the red-boxed region. (a) Raw image. (b) w/o Butterworth. (c) w/o Deformable Convolution. (d) w/o PFFB. (e) w/o PMPB. (f) w/o FFT loss. (g) w/o FFT and LAB loss. (h) w/o FFT, LAB and LCH loss. (i) w/o FFT, LAB, LCH, and VGG loss. (j) w/o FFT, LAB, LCH, VGG, and Color loss. (k) PMSPNet. (l) Ground Truth.
Figure 10. Visual comparison of networks using different components and loss functions. PSNR is shown in the top-right, and the bottom-left presents a zoomed-in view of the red-boxed region. (a) Raw image. (b) w/o Butterworth. (c) w/o Deformable Convolution. (d) w/o PFFB. (e) w/o PMPB. (f) w/o FFT loss. (g) w/o FFT and LAB loss. (h) w/o FFT, LAB and LCH loss. (i) w/o FFT, LAB, LCH, and VGG loss. (j) w/o FFT, LAB, LCH, VGG, and Color loss. (k) PMSPNet. (l) Ground Truth.
Sensors 25 05439 g010
Figure 11. Visualization of blur localization under different components and loss functions. Each group of three images corresponds to the enhanced outputs at three different scales. (a) w/o Butterworth. (b) w/o Deformable Convolution. (c) w/o PFFB. (d) w/o PMPB. (e) w/o FFT loss. (f) w/o FFT and LAB loss. (g) w/o FFT, LAB, and LCH loss. (h) w/o FFT, LAB, LCH, and VGG loss. (i) w/o FFT, LAB, LCH, VGG, and Color loss. (j) PMSPNet.
Figure 11. Visualization of blur localization under different components and loss functions. Each group of three images corresponds to the enhanced outputs at three different scales. (a) w/o Butterworth. (b) w/o Deformable Convolution. (c) w/o PFFB. (d) w/o PMPB. (e) w/o FFT loss. (f) w/o FFT and LAB loss. (g) w/o FFT, LAB, and LCH loss. (h) w/o FFT, LAB, LCH, and VGG loss. (i) w/o FFT, LAB, LCH, VGG, and Color loss. (j) PMSPNet.
Sensors 25 05439 g011
Figure 12. Visual comparison of SIFT keypoint and Canny edge detection results. The first row presents the visualization of SIFT keypoints, while the second row displays the corresponding Canny edge detection results. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Figure 12. Visual comparison of SIFT keypoint and Canny edge detection results. The first row presents the visualization of SIFT keypoints, while the second row displays the corresponding Canny edge detection results. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Ground Truth.
Sensors 25 05439 g012
Figure 13. Comparison of YOLOv12 detection effects on enhanced underwater images from the DUO dataset. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Detection effect of raw image.
Figure 13. Comparison of YOLOv12 detection effects on enhanced underwater images from the DUO dataset. (a) Raw image. (b) WFAC [25]. (c) WWPF [57]. (d) HFM [58]. (e) HLRP [59]. (f) ACDC [60]. (g) MMLE [61]. (h) PCDE [62]. (i) TEBCF [63]. (j) CycleGAN [29]. (k) U-Shape [49]. (l) FUnIE-GAN [48]. (m) Histoformer [64]. (n) Phaseformer [65]. (o) UIR-PolyKernel [66]. (p) CCL-Net [67]. (q) PUIE-Net [68]. (r) USUIR [69]. (s) SGUIE [70]. (t) PMSPNet. (u) Detection effect of raw image.
Sensors 25 05439 g013
Table 1. Comparison of performance on the N2UD dataset, where evaluation includes full-reference image quality metrics and resource consumption metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
Table 1. Comparison of performance on the N2UD dataset, where evaluation includes full-reference image quality metrics and resource consumption metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
MethodsPSNR ↑SSIM ↑FSIM ↑LPIPS ↓ Params (M) ↓FLOPs (G) ↓Time (s) ↓
WFAC [25]15.73 ± 3.210.66 ± 0.140.77 ± 0.090.35 ± 0.10--0.38
WWPF [57]17.48 ± 3.59±0.73 ± 0.130.82 ± 0.070.28 ± 0.11--0.26
HFM [58]17.31 ± 3.180.76 ± 0.110.88 ± 0.060.31 ± 0.14--0.53
HLRP [59]12.80 ± 1.970.22 ± 0.070.64 ± 0.050.51 ± 0.08--0.01
ACDC [60]16.63 ± 2.900.70 ± 0.120.82 ± 0.070.34 ± 0.11--0.22
MMLE [61]17.80 ± 3.730.73 ± 0.120.82 ± 0.070.29 ± 0.11--0.08
PCDE [62]15.40 ± 2.960.62 ± 0.140.75 ± 0.080.39 ± 0.11--0.29
TEBCF [63]17.90 ± 2.300.69 ± 0.120.80 ± 0.080.30 ± 0.09--1.24
CycleGAN [29]23.91 ± 4.720.83 ± 0.110.91 ± 0.050.24 ± 0.0122.7699.3640.03
U-Shape [49]24.32 ± 3.870.83 ± 0.110.92 ± 0.040.22 ± 0.0631.5926.100.05
FUnIE-GAN [48]21.42 ± 3.490.79 ± 0.090.90 ± 0.040.28 ± 0.063.5926.720.06
Histoformer [64]13.96 ± 2.250.30 ± 0.140.64 ± 0.070.72 ± 0.0725.7144.420.03
Phaseformer [65]24.13 ± 3.170.64 ± 0.110.93 ± 0.040.21 ± 0.091.7814.120.03
UIR-PolyKernel [66]22.19 ± 4.620.84 ± 0.090.92 ± 0.040.25 ± 0.091.8913.680.01
CCL-Net [67]23.81 ± 5.390.83 ± 0.200.91 ± 0.100.23 ± 0.160.5537.360.06
PUIE-Net [68]24.40 ± 3.780.91 ± 0.080.96 ± 0.040.17 ± 0.080.83150.690.13
USUIR [69]18.79 ± 2.860.79 ± 0.100.89 ± 0.040.32 ± 0.080.2314.880.01
SGUIE [70]24.13 ± 4.500.87 ± 0.090.93 ± 0.040.20 ± 0.0818.6320.160.02
PMSPNet25.51 ± 3.980.92 ± 0.090.95 ± 0.040.19 ± 0.084.4426.770.01
Table 2. Comparison of performance on the N2UD dataset, where evaluation is based on no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
Table 2. Comparison of performance on the N2UD dataset, where evaluation is based on no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
MethodsUIQM ↑UCIQE ↑NIQE ↓URANKER ↑Laplacian ↑Tenengrad ↑Brenner ↑
WFAC [25]3.16 ± 0.330.42 ± 0.026.02 ± 3.462.49 ± 0.890.11 ± 0.180.53 ± 0.232044.20 ± 1559.13
WWPF [57]2.85 ± 0.400.44 ± 0.045.43 ± 2.202.50 ± 0.7410.05 ± 0.080.44 ± 0.171291.93 ± 911.55
HFM [58]2.91 ± 0.370.47 ± 0.035.49 ± 2.022.35 ± 0.810.02 ± 0.060.31 ± 0.14662.62 ± 639.15
HLRP [59]2.62 ± 0.670.49 ± 0.075.78 ± 2.251.69 ± 0.960.03 ± 0.030.39 ± 0.091016.97 ± 489.98
ACDC [60]3.34 ± 0.200.38 ± 0.035.24 ± 1.782.68 ± 0.760.04 ± 0.080.42 ± 0.15990.85 ± 790.32
MMLE [61]2.77 ± 0.410.44 ± 0.045.56 ± 2.422.53 ± 0.860.07 ± 0.100.46 ± 0.191504.57 ± 1072.266
PCDE [62]2.66 ± 0.680.46 ± 0.035.50 ± 1.982.59 ± 0.740.06 ± 0.080.50 ± 0.171671.21 ± 995.43
TEBCF [63]3.00 ± 0.350.45 ± 0.025.60 ± 1.752.42 ± 0.690.049 ± 0.070.46 ± 0.141274.67 ± 699.73
CycleGAN [29]3.19 ± 0.420.41 ± 0.064.62 ± 1.301.50 ± 0.810.01 ± 0.010.26 ± 0.10448.48 ± 291.55
U-Shape [49]3.10 ± 0.460.38 ± 0.055.05 ± 1.371.35 ± 0.740.01 ± 0.010.22 ± 0.09313.03 ± 202.13
FUnIE-GAN [48]3.10 ± 0.430.43 ± 0.054.17 ± 0.901.80 ± 0.780.02 ± 0.020.30 ± 0.14607.95 ± 504.67
Histoformer [64]3.09 ± 0.270.31 ± 0.0412.00 ± 3.090.74 ± 0.540.01 ± 0.010.12 ± 0.0499.17 ± 76.09
Phaseformer [65]2.77 ± 0.390.43 ± 0.067.75 ± 6.611.26 ± 0.790.02 ± 0.030.24 ± 0.10399.73 ± 353.68
UIR-PolyKernel [66]2.91 ± 0.620.38 ± 0.075.09 ± 1.381.18 ± 0.990.01 ± 0.020.24 ± 0.13435.62 ± 447.07
CCL-Net [67]3.03 ± 0.480.41 ± 0.065.56 ± 2.051.46 ± 0.760.02 ± 0.030.26 ± 0.11488.43 ± 408.52
PUIE-Net [68]3.00 ± 0.510.40 ± 0.065.55 ± 1.851.42 ± 0.850.02 ± 0.020.25 ± 0.11430.14 ± 375.36
USUIR [69]2.96 ± 0.300.46 ± 0.034.80 ± 1.151.51 ± 0.820.01 ± 0.010.29 ± 0.11528.35 ± 372.24
SGUIE [70]2.96 ± 0.560.38 ± 0.075.46 ± 1.521.31 ± 0.900.01 ± 0.020.22 ± 0.11350.12 ± 326.33
PMSPNet3.46 ± 0.410.46 ± 0.065.30 ± 1.611.73 ± 0.790.01 ± 0.010.22 ± 0.09319.80 ± 254.67
Table 3. Comparison of performance on the EUVP dataset, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
Table 3. Comparison of performance on the EUVP dataset, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
MethodsPSNR ↑ SSIM ↑FSIM ↑LPIPS ↓UIQM ↑UCIQE ↑URANKER ↑
WFAC [25]13.24 ± 2.430.54 ± 0.100.68 ± 0.070.45 ± 0.102.81 ± 0.260.43 ± 0.022.97 ± 0.90
WWPF [57]14.62 ± 2.830.60 ± 0.100.75 ± 0.050.39 ± 0.092.68 ± 0.330.45 ± 0.042.16 ± 0.81
HFM [58]15.13 ± 2.740.66 ± 0.100.82 ± 0.040.44 ± 0.112.93 ± 0.250.49 ± 0.032.44 ± 0.99
HLRP [59]11.41 ± 1.860.17 ± 0.060.61 ± 0.040.60 ± 0.062.65 ± 0.590.50 ± 0.062.09 ± 0.99
ACDC [60]14.42 ± 2.610.60 ± 0.100.75 ± 0.070.46 ± 0.093.34 ± 0.150.38 ± 0.032.98 ± 0.81
MMLE [61]14.12 ± 2.690.59 ± 0.090.72 ± 0.060.41 ± 0.102.56 ± 0.310.45 ± 0.042.78 ± 0.98
PCDE [62]13.55 ± 2.480.52 ± 0.130.68 ± 0.080.47 ± 0.112.37 ± 0.560.47 ± 0.022.97 ± 0.80
TEBCF [63]17.07 ± 2.550.68 ± 0.090.79 ± 0.070.35 ± 0.072.82 ± 0.360.45 ± 0.032.59 ± 0.81
CycleGAN [29]22.68 ± 3.520.79 ± 0.070.89 ± 0.040.29 ± 0.063.11 ± 0.500.40 ± 0.061.30 ± 0.85
U-Shape [49]24.92 ± 3.780.83 ± 0.070.93 ± 0.020.23 ± 0.052.97 ± 0.620.38 ± 0.051.18 ± 0.84
FUnIE-GAN [48]24.06 ± 2.600.79 ± 0.050.90 ± 0.020.27 ± 0.042.88 ± 0.570.41 ± 0.051.36 ± 0.82
Histoformer [64]14.82 ± 2.880.33 ± 0.140.65 ± 0.070.71 ± 0.083.12 ± 0.230.30 ± 0.040.80 ± 0.50
Phaseformer [65]23.58 ± 2.640.61 ± 0.070.91 ± 0.020.27 ± 0.062.63 ± 0.510.40 ± 0.050.98 ± 0.81
UIR-PolyKernel [66]24.92 ± 3.860.87 ± 0.050.93 ± 0.020.22 ± 0.042.85 ± 0.740.40 ± 0.061.37 ± 0.90
CCL-Net [67]24.55 ± 3.170.84 ± 0.070.93 ± 0.020.23 ± 0.042.97 ± 0.600.38 ± 0.061.34 ± 0.80
PUIE-Net [68]24.71 ± 2.700.85 ± 0.060.93 ± 0.020.20 ± 0.042.97 ± 0.610.37 ± 0.061.12 ± 0.76
USUIR [69]17.53 ± 2.470.73 ± 0.080.86 ± 0.030.35 ± 0.082.82 ± 0.220.47 ± 0.041.78 ± 0.89
SGUIE [70]25.48 ± 3.230.84 ± 0.060.92 ± 0.030.24 ± 0.052.83 ± 0.710.38 ± 0.061.12 ± 0.87
PMSPNet25.81 ± 3.220.85 ± 0.070.94 ± 0.020.21 ± 0.043.09 ± 0.460.36 ± 0.051.16 ± 0.78
Table 4. Comparison of performance on the LSUI dataset, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
Table 4. Comparison of performance on the LSUI dataset, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
MethodsPSNR ↑SSIM ↑FSIM ↑LPIPS ↓UIQM ↑UCIQE ↑URANKER ↑
WFAC [25]15.35 ± 3.090.61 ± 0.140.73 ± 0.090.37 ± 0.082.76 ± 0.450.43 ± 0.022.55 ± 0.84
WWPF [57]17.51 ± 3.440.70 ± 0.120.81 ± 0.060.30 ± 0.082.76 ± 0.410.45 ± 0.052.59 ± 0.65
HFM [58]17.63 ± 2.930.74 ± 0.110.87 ± 0.060.33 ± 0.112.80 ± 0.310.47 ± 0.032.41 ± 0.73
HLRP [59]13.04 ± 1.870.22 ± 0.080.64 ± 0.040.55 ± 0.052.80 ± 0.580.47 ± 0.091.65 ± 0.88
ACDC [60]16.96 ± 2.630.71 ± 0.120.82 ± 0.060.33 ± 0.103.34 ± 0.180.38 ± 0.032.63 ± 0.72
MMLE [61]17.59 ± 3.150.69 ± 0.110.79 ± 0.060.31 ± 0.082.55 ± 0.440.45 ± 0.042.59 ± 0.80
PCDE [62]15.25 ± 2.300.59 ± 0.110.73 ± 0.070.40 ± 0.092.32 ± 0.570.47 ± 0.032.75 ± 0.69
TEBCF [63]17.95 ± 2.050.68 ± 0.120.80 ± 0.080.31 ± 0.082.93 ± 0.330.45 ± 0.032.58 ± 0.64
CycleGAN [29]24.93 ± 4.320.85 ± 0.110.92 ± 0.050.23 ± 0.093.21 ± 0.370.42 ± 0.061.63 ± 0.71
U-Shape [49]24.94 ± 3.540.84 ± 0.110.92 ± 0.040.22 ± 0.073.10 ± 0.410.39 ± 0.051.38 ± 0.61
FUnIE-GAN [48]21.47 ± 3.320.80 ± 0.100.90 ± 0.040.28 ± 0.063.09 ± 0.380.43 ± 0.051.84 ± 0.64
Histoformer [64]14.05 ± 2.040.31 ± 0.140.65 ± 0.070.72 ± 0.073.07 ± 0.280.31 ± 0.040.70 ± 0.56
Phaseformer [65]24.64 ± 3.140.64 ± 0.110.93 ± 0.040.20 ± 0.092.79 ± 0.330.44 ± 0.061.29 ± 0.70
UIR-PolyKernel [66]22.22 ± 4.200.84 ± 0.090.92 ± 0.040.26 ± 0.092.91 ± 0.540.38 ± 0.071.16 ± 0.90
CCL-Net [67]25.15 ± 4.450.87 ± 0.150.94 ± 0.080.19 ± 0.123.02 ± 0.430.41 ± 0.061.52 ± 0.67
PUIE-Net [68]26.21 ± 3.630.90 ± 0.090.95 ± 0.040.18 ± 0.083.06 ± 0.440.39 ± 0.061.39 ± 0.65
USUIR [69]18.86 ± 2.740.80 ± 0.110.89 ± 0.040.33 ± 0.082.96 ± 0.290.45 ± 0.041.44 ± 0.75
SGUIE [70]24.59 ± 4.400.87 ± 0.100.93 ± 0.040.19 ± 0.082.96 ± 0.490.39 ± 0.071.37 ± 0.78
PMSPNet26.46 ± 3.910.92 ± 0.090.96 ± 0.030.13 ± 0.073.31 ± 0.380.45 ± 0.061.51 ± 0.70
Table 5. Comparison of performance on the UIEB dataset, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
Table 5. Comparison of performance on the UIEB dataset, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better, while ↓ indicates that a lower value is better.
MethodsPSNR ↑ SSIM ↑FSIM ↑LPIPS ↓UIQM ↑UCIQE ↑URANKER ↑
WFAC [25]15.39 ± 2.250.66 ± 0.130.74 ± 0.100.35 ± 0.112.79 ± 0.520.43 ± 0.022.39 ± 0.79
WWPF [57]17.52 ± 2.710.76 ± 0.100.84 ± 0.080.25 ± 0.112.75 ± 0.540.44 ± 0.042.54 ± 0.89
HFM [58]17.76 ± 3.440.79 ± 0.110.89 ± 0.070.26 ± 0.142.93 ± 0.520.47 ± 0.032.33 ± 0.98
HLRP [59]13.33 ± 1.580.19 ± 0.070.64 ± 0.040.55 ± 0.073.10 ± 0.670.43 ± 0.091.46 ± 0.88
ACDC [60]17.70 ± 3.200.78 ± 0.100.86 ± 0.080.27 ± 0.123.39 ± 0.350.38 ± 0.022.57 ± 0.85
MMLE [61]17.35 ± 2.910.73 ± 0.110.80 ± 0.080.29 ± 0.112.46 ± 0.570.45 ± 0.042.56 ± 0.96
PCDE [62]15.20 ± 3.670.61 ± 0.190.75 ± 0.120.38 ± 0.132.27 ± 0.940.44 ± 0.022.69 ± 0.81
TEBCF [63]17.68 ± 2.510.76 ± 0.130.84 ± 0.110.25 ± 0.102.84 ± 0.380.46 ± 0.032.60 ± 0.87
CycleGAN [29]19.44 ± 4.280.77 ± 0.100.88 ± 0.060.27 ± 0.093.19 ± 0.500.40 ± 0.061.15 ± 1.02
U-Shape [49]20.72 ± 3.590.81 ± 0.100.89 ± 0.060.21 ± 0.073.25 ± 0.430.37 ± 0.051.39 ± 1.09
FUnIE-GAN [48]18.02 ± 2.100.76 ± 0.070.88 ± 0.040.29 ± 0.083.42 ± 0.210.43 ± 0.052.09 ± 1.06
Histoformer [64]12.50 ± 1.620.23 ± 0.130.59 ± 0.090.73 ± 0.053.15 ± 0.220.32 ± 0.040.80 ± 0.47
Phaseformer [65]22.41 ± 3.280.68 ± 0.150.93 ± 0.040.16 ± 0.092.82 ± 0.470.43 ± 0.051.48 ± 1.03
UIR-PolyKernel [66]17.72 ± 4.060.80 ± 0.100.90 ± 0.050.24 ± 0.112.96 ± 0.780.38 ± 0.071.07 ± 1.44
CCL-Net [67]16.93 ± 5.990.64 ± 0.310.79 ± 0.170.38 ± 0.273.16 ± 0.520.41 ± 0.061.32 ± 0.97
PUIE-Net [68]22.43 ± 3.950.90 ± 0.070.94 ± 0.050.13 ± 0.073.09 ± 0.490.39 ± 0.071.45 ± 1.27
USUIR [69]19.95 ± 3.410.82 ± 0.090.91 ± 0.060.24 ± 0.103.18 ± 0.350.46 ± 0.031.56 ± 0.98
SGUIE [70]20.42 ± 4.470.86 ± 0.090.92 ± 0.050.19 ± 0.113.10 ± 0.630.37 ± 0.071.25 ± 1.35
PMSPNet22.43 ± 3.370.86 ± 0.080.92 ± 0.050.21 ± 0.103.27 ± 0.390.39 ± 0.051.53 ± 1.08
Table 6. Comparison of performance using different components, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better. ✓ indicates the model is evaluated w/the corresponding module, while ✗ denotes the model is evaluated w/o the corresponding module.
Table 6. Comparison of performance using different components, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better. ✓ indicates the model is evaluated w/the corresponding module, while ✗ denotes the model is evaluated w/o the corresponding module.
ButterWorthDeformablePMPBPFFBPSNR ↑SSIM ↑FSIM ↑ UIQM ↑ UCIQE ↑
25.98 ± 4.200.88 ± 0.090.94 ± 0.043.08 ± 0.400.39 ± 0.06
17.18 ± 5.140.73 ± 0.130.87 ± 0.062.31 ± 0.840.41 ± 0.08
25.99 ± 4.290.89 ± 0.090.94 ± 0.043.09 ± 0.390.40 ± 0.06
24.78 ± 4.070.87 ± 0.090.94 ± 0.043.08 ± 0.410.39 ± 0.06
25.51 ± 3.980.92 ± 0.090.95 ± 0.043.46 ± 0.410.46 ± 0.06
Table 7. Comparison of performance using different loss functions, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better. ✓ indicates the model is evaluated w/the corresponding module, while ✗ denotes the model is evaluated w/o the corresponding module.
Table 7. Comparison of performance using different loss functions, where evaluation includes full-reference and no-reference image quality metrics. All results are reported in the format of mean ± standard deviation. ↑ indicates that a higher value is better. ✓ indicates the model is evaluated w/the corresponding module, while ✗ denotes the model is evaluated w/o the corresponding module.
CharbonnierFFTLABLCHVGGColorPSNR ↑SSIM ↑FSIM ↑UIQM ↑UCIQE ↑
24.00 ± 3.980.86 ± 0.090.91 ± 0.042.90 ± 0.450.41 ± 0.07
21.27 ± 4.310.83 ± 0.090.91 ± 0.042.95 ± 0.410.35 ± 0.06
21.43 ± 4.210.83 ± 0.090.91 ± 0.043.08 ± 0.370.35 ± 0.06
22.53 ± 4.360.84 ± 0.090.92 ± 0.043.08 ± 0.370.37 ± 0.06
24.06 ± 3.840.85 ± 0.090.92 ± 0.043.08 ± 0.380.39 ± 0.06
25.51 ± 3.800.92 ± 0.090.95 ± 0.043.46 ± 0.410.46 ± 0.06
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kong, D.; Zhang, Y.; Zhao, X.; Wang, Y.; Wang, Y. Progressive Multi-Scale Perception Network for Non-Uniformly Blurred Underwater Image Restoration. Sensors 2025, 25, 5439. https://doi.org/10.3390/s25175439

AMA Style

Kong D, Zhang Y, Zhao X, Wang Y, Wang Y. Progressive Multi-Scale Perception Network for Non-Uniformly Blurred Underwater Image Restoration. Sensors. 2025; 25(17):5439. https://doi.org/10.3390/s25175439

Chicago/Turabian Style

Kong, Dechuan, Yandi Zhang, Xiaohu Zhao, Yanyan Wang, and Yanqiang Wang. 2025. "Progressive Multi-Scale Perception Network for Non-Uniformly Blurred Underwater Image Restoration" Sensors 25, no. 17: 5439. https://doi.org/10.3390/s25175439

APA Style

Kong, D., Zhang, Y., Zhao, X., Wang, Y., & Wang, Y. (2025). Progressive Multi-Scale Perception Network for Non-Uniformly Blurred Underwater Image Restoration. Sensors, 25(17), 5439. https://doi.org/10.3390/s25175439

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop