Next Article in Journal
A Survey on Privacy Preservation Techniques in IoT Systems
Previous Article in Journal
High-Precision Waveform Stacking Location Method for Microseismic Events Based on S-Transform
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Impact of Underwater Image Enhancement on Feature Matching

by
Jason M. Summers
1,*,
Mark W. Jones
1,* and
Catherine Seale
2
1
Department of Computer Science, Swansea University, Swansea SA1 8EN, UK
2
Beam, Bristol BS1 6BX, UK
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(22), 6966; https://doi.org/10.3390/s25226966
Submission received: 14 October 2025 / Revised: 7 November 2025 / Accepted: 12 November 2025 / Published: 14 November 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

We introduce local matching stability and furthest matchable frame as quantitative measures for evaluating the success of underwater image enhancement. This enhancement process addresses visual degradation caused by light absorption, scattering, marine growth, and debris. Enhanced imagery plays a critical role in downstream tasks such as path detection and autonomous navigation for underwater vehicles, relying on robust feature extraction and frame matching. To assess the impact of enhancement techniques on frame-matching performance, we propose a novel evaluation framework tailored to underwater environments. Through metric-based analysis, we identify strengths and limitations of existing approaches and pinpoint gaps in their assessment of real-world applicability. By incorporating a practical matching strategy, our framework offers a robust, context-aware benchmark for comparing enhancement methods. Finally, we demonstrate how visual improvements affect the performance of a complete real-world algorithm—Simultaneous Localization and Mapping (SLAM)—reinforcing the framework’s relevance to operational underwater scenarios.

1. Introduction

In recent years, there have been large developments of offshore wind farms (OWFs). These, and other offshore structures and assets, are in greater need of inspection and monitoring compared to structures on land because structures in an underwater environment may be subject to biological growth, turbulent currents, corrosion and physical degradation, all of which may contribute to an increase in fatigue. Checks and maintenance for these structures present many challenges due to the depth and length of the inspections, and so, many surveys are conducted with remotely operated underwater vehicles (ROUVs or ROVs) or autonomous underwater vehicles (AUVs). These vehicles record hours of footage using a range of sensors, particularly optical. This footage is analysed to detect structural damage, either through environment mapping or by identifying specific regions of interest within the footage. Due to the length of the footage, it is useful to move towards the introduction of automatic analysis algorithms.
Current image enhancement research typically evaluates performance using known or synthetically generated ground truths, along with metrics such as PSNR, SSIM, and LPIPS. However, little attention has been given to its impact on downstream tasks like feature matching, camera motion estimation, and trajectory reconstruction. In this work, we address this gap by evaluating enhancement methods through their influence on these downstream applications.

1.1. Challenges of Underwater Data

Image enhancement is used to address several challenges presented by underwater images:
  • Challenge 1—Occlusion and Illumination:
Inconsistencies in visibility can arise from the varying depths, capture angles, sedimentation and properties of the water, which can greatly affect a model’s ability to classify elements of an underwater image [1]. Objects such as bubbles, debris and biological life can often block regions of the footage, potentially causing the inspector to miss key parts of a structure. Light degradation can also be an issue; the Lambert–Beer empirical law states that decay in the intensity of light depends on the properties of the medium through which the light travels; so, the water itself can alter the colour and illumination of objects [2,3]. For every 10 metres of depth underwater, the light available is halved [4]. Visibility can also vary with weather, and storms can create turbulent water, creating a more complex medium that amplifies the issues above. Light degradation and absorption are not constant for all wavelengths of light, longer wavelengths like red and orange are more easily absorbed in contrast to shorter wavelengths like blue and green. Light from the surface is therefore altered with depth, where footage often has a blue-green tint.
  • Challenge 2—Noise and Distractions:
Underwater scenes frequently encounter various sources of noise. Biological debris, known as marine snow, can cause challenges with backlighting, while live fish and bubbles are prominent elements that can affect deep vision models significantly. Additionally, video and image capture may introduce blurring, video noise, and lens distortion. These factors collectively pose barriers to effective image analysis tasks, including feature or object identification, for both humans and vision models. For vision models being used by the inspector, artifacts can negatively affect the contrast of images, or lead to poor embedding spaces or inaccurate similarity scores in the case of deep models. This challenge is particularly pronounced in video inspections, where ROV footage often spans multiple hours.
  • Challenge 3—Non-stable Video Capture:
Turbulent water causes the ROV to have unpredictable movement and video capture, hindering the temporal value that video could provide. The stability of footage is an important factor in accessing the trajectory of objects in the footage [5], which, in turn, can affect the feature and object detection capabilities of ML models [6,7].
  • Objectives:
We set out to answer the question of how we measure enhancement with regard to real-world applications using the following objectives:
  • Identify and analyse the current approaches for image enhancement in the underwater domain.
  • Examine the current metrics used to quantify and compare underwater image enhancements.
  • Propose new measures of feature matching consistency over multiple subsequent frames and furthest frame matching to evaluate enhancement.
  • Incorporate these into a new framework for evaluating the real-world applicability of underwater image enhancement.
Next, we review underwater image enhancement (UIE), including how to improve an image, and how to assess the quality of the improvement. We report classical approaches (Section 1.2), followed by a look into deep learning-based methods (Section 1.3) and cover the methods of quality assessment and image evaluation (Section 1.4).

1.2. Classical and Physically Based Approaches for UIE

Despite the rise in interest in deep learning in recent years, we found that non-deep learning-based methods still play a significant role in underwater image enhancement.
Popular methods are histogram equalisation (HE), modifying channel priors, and wavelet transforms. These methods share a common approach of spectral augmentation, in that they manipulate regions across the light spectrum to improve contrast. However, while histogram equalization generally provides enhancement across the entire spectrum, wavelet transforms and channel priors focus on specific regions with more discrimination. Histogram manipulation and equalisation was a focus of many publications [8,9,10,11,12,13,14,15]. A basic form of this is global HE, which applies this method to an entire image using a transformation function derived from the image cumulative distribution function (CDF). Histogram equalisation methods are often described in this field as having limited effectiveness in underwater scenes due to the nature of light degradation [13,16]. Consequently, HE techniques have evolved to address domain specific image characteristics, including resulting from the physical properties of water.
The most common improvement in the literature is to perform HE at different scales to limit the influence of noise. Such methods include Contrast Limited Adaptive Histogram Equalization (CLAHE) [17], an important variant designed to enhance contrast in noisy images. Although much of its early use was in medical imagery, it has garnered interest in the underwater imagery domain, being used as a baseline for multiple experiments in the literature [14,18]. It works by performing HE on tiles of an image while limiting the contrast to prevent over-amplification of noise before being interpolated to form the enhanced image.
Bai et al. [8] adopt another multi-scale approach, applying both global Histogram Equalization (GHE) and local Histogram Equalization (LHE) to address different aspects of enhancement. They then adopt a fusion strategy inspired by [19] to combine the luminance, saliency, and exposure weight maps generated from the equalized components into the final enhanced image.
The idea of using fusion to combine various representations is a popular one across both classical and deep methods [8,20,21,22,23,24,25,26,27]. Ancuti et al. [19] present a multi-scale fusion pipeline that breaks the image into Laplacian and Gaussian pyramids, which are blended at each level. Saliency mapping is considered to improve the visibility of objects in the scene. The outcome demonstrates an effective enhancement technique that offers advantages for industrial computer vision. Severely hazy and unevenly illuminated scenes, including industrial scenes, all sourced from a large selection of cameras, were among those considered during their evaluation.

1.3. Deep-Learning-Based Methods for UIE

Deep model training continues to be one of the biggest areas of research in this field, growing rapidly. Some relatively older designs continue to remain relevant but with updated features and techniques being incorporated over time.
The challenge of image enhancement, using deep learning, primarily revolves around the use of image-to-image models. These models typically comprise an encoder, responsible for generating a new representation of the input image, and a decoder, tasked with reconstructing or enhancing the image based on this new representation. These are broadly referred to as encoder–decoder models but many variations exist.
Convolutional Neural Networks (CNNs) are widely adopted for feature extraction, using convolving filters or kernels to capture spatial information within image data. Since their inception, CNNs have been prominent in much of the computer vision field, and remain so for underwater image enhancement, being heavily used in many architectures in the literature [24,28,29,30,31]. Attention mechanisms are a more recent addition to the machine learning arsenal, with their efficacy notably demonstrated in 2017 by Vaswani et al. [32]. In the context of images, a spatial attention mechanism can be employed to prioritize and focus on different spatial regions in an image, according to their relevance to the task at hand. Their full potential is still being realised but we can already see examples in the literature that utilise this mechanism for underwater enhancement [33,34,35,36].
Concerning overall architectural frameworks, it is clear from the literature that generative adversarial networks (GANs) had a major impact and influence on the field. Although GANs are adept at generating realistic data, their general functionality does not inherently facilitate image enhancement tasks. CycleGAN [37] is a design that learns mappings between two different domains to facilitate unpaired image-to-image translation tasks and utilises encoder–decoder feature extraction much like an autoencoder. Multiple studies [24,29,33,35,38,39,40,41,42] build on the idea of GANs by using the discriminator to critique the reconstruction capabilities of an autoencoder. Namely, it will discriminate between undistorted images from the source dataset and distorted images that have been enhanced via the autoencoder.
FUnIE-GAN [43] is a successful example of the utilisation of a multitude of methods. They present a convolutional encoder–decoder system that uses an adversarial loss generated by a discriminator for perceptual image enhancement. The design of the encoder–decoder is a U-Net [44], where skip-connections are used between the mirrored encoder–decoder layers. Alongside a new dataset, they also formulate an ensemble of loss functions to access perpetual image quality on numerous levels. Namely, they use L 1 loss to access global similarity, a VGG-19 backed content loss function [45,46,47], and, finally, an adversarial loss.

1.4. Evaluating Performance and SLAM

Objective quantitative metrics are required for performance evaluation, rather than depending on visual inspection of the enhancement results. Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM) dominate the literature. Other general image quality metrics included Natural Image Quality Evaluator (NIQE) [48], used by papers such as [28,35,49,50,51], Visual Information Fidelity (VIF) [52], used in [28,49], and Perceptual Image Quality Evaluator (PIQE) [53], used in [28,35,50]. There are also metrics focused specifically on underwater image quality, such as Underwater Image Quality Measure (UIQM) [54], which saw use in [41,49,51,55], and Underwater Colour Image Quality Evaluation (UCIQE) [56], which was used in [41,51,55,57]. These metrics deliver a quick and objective analysis for the quality of an image but lack real-world implications, particularly the effect of model-based image enhancement on downstream tasks.
The process of feature matching is crucial to mapping out environments and structures using optical data, an important procedure for industrial surveys, particularly in underwater environments where structural damage is common, and inspections are needed frequently. While some studies, such as those by Ancuti et al. [19] and Yan et al. [57], have tested enhanced data using local feature matching between images, they have been limited to example pairs of images, rather than testing across multiple frames of a video. This ignores the importance of frame-by-frame relationships and does not represent how the enhancement impacts video-based tasks. There is a lack of literature that conducts a comprehensive comparative evaluation in this fashion and this is something we address in this work. Hidalgo and Bräunl [58] examine the behaviours of feature detection, and matching, in frames sourced from varied noisy underwater ROV videos. They compare SIFT [59], SURF [60], ORB [61], BRISK [62], and AKAZE [63] feature detection methods by recording the average features found by each detector, and the average number of inliers using nearest neighbours and homography between two consecutive frames. In the number of features detected and the number of inliers matched, both ORB and BRISK feature detectors performed the best, with SIFT performance following close behind. They additionally apply two image enhancement algorithms, a fusion filter [64], and a backscatter removal filter [65] to each dataset. They found that the enhancements improved the number of detected features across the board but resulted in very low improvements in inliers for all detectors except AKAZE. We evaluate the impact of further state-of-the-art underwater image enhancement on feature detection over longer duration.
SLAM: Feature matching is a core component of Simultaneous Localization And Mapping (SLAM) for mapping 3D points and path of capture source such as a ROV in an environment. Zhang et al. [18] utilised sequential frame matching and SLAM performance as an empirical assessment of a model. They test CLAHE, Median Filtering (MF), and Dark Channel Prior (DCP) using ORB-SLAM 2, an implementation of SLAM that identifies ORB features [61]. They provide both a practical metric for comparing models and guidance for optimising SLAM performance. This approach warrants further exploration and refinement. Hence, in this paper, we address the problem on two levels: first, by measuring the impact of visual enhancement on sequential frame matching ability, and second, by observing its effect on the complete pipeline, in our case, ORB-SLAM 3.
In conclusion, while existing underwater image enhancement methods focus primarily on visual quality metrics, they overlook the downstream impact on real-world tasks such as frame matching and navigation which are part of algorithms such as SLAM. To bridge this gap, we introduce the following innovations:
  • Local Matching Stability: a measure of how consistently features are matched across enhanced frames.
  • Furthest Matchable Frame: a metric capturing the temporal reach of reliable feature matching post-enhancement.
  • Context-Aware Evaluation Framework: tailored to underwater environments and grounded in well-known and practical matching strategies.
  • SLAM-Integrated Benchmarking: demonstrating how enhancement affects performance in an industrial context.

2. Materials and Methods

To find the impact of visual enhancement on downstream tasks, we apply a range of enhancement methods from the literature, including deep model methods, on underwater videos and measure their impact on feature matching.

2.1. Frame Matching

Frame matching involves identifying and correlating features between two images using feature descriptors. Location differences of matched features can indicate movement of the camera or a scene object. By comparing multiple matched features with similar displacements, the camera’s relative movement can be deduced geometrically using a homography model. A higher number of features that consistently align with the model increases confidence in the estimation of the camera’s motion. Matched features whose displacements do not align with the model, either due to the object motion, general noise, or occlusion, are ignored. This is commonly achieved using the Random Sample Consensus (RANSAC) algorithm, which iteratively fits the model to subsets of matches and identifies the largest set of inliers, rejecting outliers as inconsistent matches. The sensitivity of this filter is defined by a threshold for the maximum permissible distance (in pixels) between the actual position of a feature and the predicted position by the model during tuning.
In order to test if an enhancement better facilitates frame matching in a video, we employ two functions. One tests how a given frame in the video matches with the next n consecutive frames, and the second will find the furthest frame that can still be matched given a set of threshold parameters. The threshold parameters for this study were chosen after experimentation, balancing the number of matched features to avoid both over and under-fitting. The parameters provide a suitable filter for rigorous testing while still enabling successful matches that visibly show aligned, often parallel, feature trace lines during testing, as seen in Figure 1.

2.1.1. Feature Extraction

The main approaches for extracting features are ORB [61], KAZE [66], AKAZE [63], and BRISK [62], being available in the OpenCV library [67]. SIFT [59] and SURF [60] are also important types of feature extraction but have limited availability, particularly SURF, within OpenCV. Superpoint [68] is a keypoint detection system built using a self-supervised CNN framework, and is part of a modern wave of new deep model designs for feature detection and homography [69].
We examine the feature methods SIFT, ORB, BRISK, KAZE, AKAZE, and SuperPoint, omitting SURF due to licensing issues within OpenCV. In order to find sufficient features but maintain an acceptable processing time, we limit the number of extracted features for ORB and SIFT to 1000, the number of octaves to sample for BRISK features to 4, and set the threshold for KAZE and AKAZE to 0.001 .

2.1.2. Local Matching Stability

First, we apply a feature-finding technique on every frame in the chosen video (see Figure 2). After extracting all features for all frames, we iterate frame by frame and attempt to match features with the subject f r a m e with those from f r a m e + 1 , then f r a m e + 2 , and up to f r a m e + n for the chosen n, recording the camera homography data, namely, the average reprojection errors for those features, inlier number, and inlier percentage. We record the matching performance over the n = 10 frames from the subject frame (see Figure 3).

2.1.3. Furthest Matchable Frame Metric

In order to compare where enhancement has affected long-term feature matching, we use a similar process to our local matching function, but continue as far as possible given a threshold on the RANSAC filtered the matching results. The parameters used for this threshold are as follows: RANSAC Threshold: 10.0 , Inlier Ratio: 0.3 , Max Reprojection Error: 20.0 . Threshold values were determined through extensive testing on diverse underwater scenes. The selected configuration consistently offered the best balance between feature detection robustness and reliable frame-to-frame matching, maintaining tracking stability even under low visibility and dynamic lighting conditions.
The inlier ratio refers to the minimum proportion of points that must be classified as inliers for the model to be considered valid. We found that this typically low value of 30 % was appropriate given the noisiness of underwater data. The max reprojection error measures the distance (in pixels) between the actual and model predicted positions of a feature. However, it is applied to the results of the final model rather than during the RANSAC fitting process. This value is typically closer to or lower than the RANSAC threshold to ensure a stricter more robust model fit. However, due to radial lens distortion, a more lenient reprojection error threshold was used. Frames satisfying these values will yield correctly tracked camera motion with high certainty.

2.1.4. Procedure

The complete procedure for the feature-matching is shown in Algorithm 1. A sequential search is used because we seek continuous matching success; a binary search could be more efficient but would jump past intermediate failures and identify matches that appear later due to loops or repeated scene content, which would misrepresent how long features remain trackable.
Algorithm 1: The Complete Feature-Matching-Based Evaluation Framework
Sensors 25 06966 i001

2.1.5. ORB-SLAM 3

As discussed in Section 1.4, visual SLAM is an important procedure in underwater surveys and a key example of frame matching that is used in industry tasks. ORB-SLAM 3 is one implementation that has been frequently used as a benchmark [18,70] and is the state of the art for SLAM implementations using ORB based features. We finalise our testing by observing the performance of SLAM after visual enhancement, using consistent tracking indicators and loop closures as measures.

2.2. Test Datasets

Our first test video is a publicly available video with loop closures and Inertial Measurement Unit (IMU) data, captured and used in a study by Joshi et al. [71]. This video contains an exploration of a natural cave system. The second test video is sourced from a two-hour inspection of a wind turbine base. This footage was captured using a ROV as it traversed near a variety of pipe structures, including anodes. The footage exhibits many of the highlighted noise forms and additional challenges, such as segments of dark and featureless backgrounds. As a result of these additional challenges, selecting a candidate segment for this test proved to be a difficult task in itself, as segments of quick movement and reduced features prevented consistent completion of the ORB-SLAM 3 process. The desired video should remain challenging, be discriminating amongst techniques and still allow a complete SLAM run to be achievable. We chose a five-minute segment that starts near the seabed and continues up the wind turbine structure, referring to this video as ‘Seabed’. Camera set-up and parameters are in the original work for ‘Cave’ [71]. The ‘Seabed’ video was captured by ROVCO’s SUBSLAM X1 camera featuring a stereo pair of 4 K cameras in a pressure-resistant enclosure (tested to 6700 m). See Figure 4 for representative frames from the two videos.

2.3. Test Models

We test models developed by Ancuti et al. [19], FUnIE GAN [43], WaveNet [49,72], WaterNet [31], and, finally, UVENet [73]. This selection includes a variety of approaches, ranging from an important milestone [19] to the state of the art [73], and covers both classical approaches [19,72] and deep models [31,43,49,73].
Each enhancement method is individually applied to each frame of the videos to create the test video set. UVENet [73] highlighted a challenge found among other recent models, particularly those with temporal elements. These models can struggle to process long, high-definition videos. This limitation stems in part from their architectural design and from the limited sizes, both in length and definition, of their training clips. In order to not encounter memory issues while maximising quality, the ‘cave’ video was divided into three parts and individually enhanced. Even so, subsequent upscaling was needed to obtain a definition matching the original.

3. Results

3.1. Classic Enhancement Validation

An initial analysis was conducted using classic metrics, namely, PSNR and SSIM, to provide a baseline for our results and ensure the enhancement pipeline was functioning correctly.
These values (Table 1), particularly the SSIM scores, are comparable to those from the original research papers, and overall suggest successful enhancements by each of those methods on these challenging new datasets. Based on these values, Waternet [31] performs the best.

3.2. Frame Matching Benchmark

Each video is preprocessed by the appropriate feature detection and description algorithm (AKAZE, BRISK etc.). For example, ORB took approximately 25 min to process 6030 frames on an Intel Core i9-12900k with 128 GB, corresponding to a throughput of approximately four frames per second. The entire feature matching evaluation (Algorithm 1), took 210 s at a frame rate of approximately 28 fps.
After running our measures discussed in Section 2.1, we have, for each subject frame, a furthest matchable frame (FMF) value, and the number of inliers, inlier percentage and the average reprojection error for the next ten frames after each subject frame.

3.2.1. Inlier Decay

Figure 5 shows the mean number of inliers for the next ten frames across the whole video for each enhancement method using the ORB detector. The decay curve demonstrates that the next frame (1) has the highest number of inlier matches, and each subsequent frame has fewer matches with the subject frame compared to the previous frame. The video enhanced by WaterNet, WaveNet and the unaltered original video perform consistently the best, with WaveNet and the original video being almost tied in Figure 5. The video enhanced by Demir and Kaplan [72] using a sharpening-smoothing image filter with a CLAHE performs the poorest in facilitating inlier detection.
Similar results are seen in the Seabed dataset (Figure 6), where WaterNet and the original video have almost identical results and perform the best. Demir and Kaplan [72] is also the lowest performer. The results of the two datasets are consistent.

3.2.2. Furthest Matchable Frame

The FMF serves as a metric to evaluate how effectively a video enables optimal and consistent feature detection, showing that objects and features have the sustained clarity needed for frame matching for as long as possible. We have two approaches to evaluate the FMF values recorded for each frame. The first is the average furthest matched frame value across the whole video and is presented in Table 2 and Table 3.
We can see in both datasets that videos enhanced by classical methods performed poorly, with Demir and Kaplan [72] coming last in every detector except SuperPoint, and Ancuti et al. [19] coming in the bottom three apart from a few notable exceptions. The unaltered original video for both datasets performed consistently well for all detector types, often coming second. WaveNet, UVENet, and FUnIE-GAN showed varied results, often in the middle rankings, with the one exception of SuperPoint features, where UVENet showed a good improvement over the original video. Only WaterNet showed consistently improved frame matching over all feature types, while all others were consistently outperformed by the original video.
Figure 7 and Figure 8 underscore much of our previous findings. Demir and Kaplan [72] starts to plateau at the lowest values for frames ahead for both datasets, demonstrating an inability to facilitate significantly far matches compared to all other methods. Ancuti et al. [19], UVENet, and WaveNet demonstrate inconsistent results between the datasets. Importantly, WaterNet again performs consistently well on both datasets and is the only method to show a marked improvement in the original video frames.

3.3. ORB-SLAM Findings

Finally, we tested the impact of enhanced videos against the original in a full SLAM pipeline. We ran each enhanced or original video from both datasets through ORB-SLAM 3 and recorded tracking status, camera trajectory, loop closure logs, and 3D point cloud data. We found that although ORB-SLAM 3 was able to maintain good tracking for most of the original video, all enhanced videos resulted in significant tracking loss with poor or no recovery once tracking was lost. As a result, each enhanced video showed evidence of trajectory drift and no video, except the original, established any loop closures. Comparing the number of points mapped with the number of frames successfully tracked showed a more even result, with Ancuti et al. [19] doing the best in both datasets, and the original video still doing better than many of the enhanced videos, as well as having the more complete final 3D cloud. These results are presented in Appendix B.

4. Discussion and Conclusions

Although the SSIM and PSNR scores indicate that the selected enhancement methods perform without worsening the quality of the original test videos, our findings reveal that the impact of these enhancements are not positive. Testing six distinct approaches, including classical and deep-learning-based methods, we found a wide range of frame-matching abilities. In the majority of cases the videos produced after enhancement were equal or worse in performance compared to the unaltered original video. Demir and Kaplan [72] stood out as performing the worst in our evaluation, while WaterNet [31] was the best and the only method that reliably facilitated better frame-matching results, although only marginally. All enhancement methods tested were detrimental to SLAM’s tracking accuracy, and loop closure detection, leading to drift and incomplete map generation.
Evaluating enhancement performance using full SLAM is computationally prohibitive, particularly for large-scale video data due to significant memory demands. Our central insight is that frame matching—a core component of SLAM—can serve as an effective surrogate metric. This approach offers substantial computational efficiency while still yielding reliable and robust results for cross-method comparison.
Our experiments indicate strong correspondence between frame-matching metrics and the visual trajectories produced by full SLAM, yet the computational overhead is markedly reduced. Consequently, our evaluation framework provides a lightweight, scalable solution for benchmarking underwater image enhancement methods in the context of downstream applications. It offers a practical mechanism to assess whether new enhancement techniques can meaningfully support tasks such as SLAM-based localisation and navigation.
However, there are opportunities to deepen this analysis by incorporating more detailed information on SLAM or other relevant algorithms. Doing so may shed light not only on what causes negative impacts from visual enhancements, but also on why these impacts occur. Key point persistence information provided by ORB-SLAM 3 could be utilised in the future to measure feature clarity consistency on a frame-by-frame basis, forming another metric for enhancement performance. Additionally, there are other enhancement styles yet to be explored, such as temporal models, which fully utilise the information inherent in video data. There are also other industrial algorithms like 3D reconstruction, which have unique challenges not fully addressed in this work.
This framework opens opportunities for future work through the potential use of local matching stability as a component of a loss function in a deep enhancement model. This, managed by a perceptual loss, could ensure that learned improvements promote real and easily matched features across frames, promoting more effective and accurate mapping, although the computation would need to be sped up to enable use for training loss versus validation loss.

Author Contributions

Conceptualization, J.M.S., M.W.J. and C.S.; methodology, J.M.S., M.W.J. and C.S.; software, J.M.S. and M.W.J.; validation, J.M.S. and M.W.J.; formal analysis, J.M.S. and M.W.J.; investigation, J.M.S., M.W.J. and C.S.; resources, M.W.J. and C.S.; data curation, J.M.S., M.W.J. and C.S.; writing—original draft preparation, J.M.S. and M.W.J.; writing—review and editing, J.M.S., M.W.J. and C.S.; visualization, J.M.S. and M.W.J.; supervision, M.W.J. and C.S.; project administration, M.W.J. and C.S.; funding acquisition, M.W.J. and C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research and APC was supported by the Engineering and Physical Sciences Research Council (EPSRC) [grant number EP/S021892/1].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The ‘Cave’ video dataset presented in this study is openly available by Joshi et al. [71]. The ‘Seabed’ video dataset is not readily available because of commercial sensitivity. Code to generate, evaluate and visualise the measures will be made available shortly after publication.

Conflicts of Interest

Author Catherine Seale is employed by the company Beam. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Appendix A. Additional Data

Table A1 and Table A2 provide specific values for Figure 5 and Figure 6.
Table A1. Average number of inliers for different frame offsets across various enhancement methods for the Cave video.
Table A1. Average number of inliers for different frame offsets across various enhancement methods for the Cave video.
Average Number of Inliers for Frame Offset:
Enhancement Method12345678910
Ancuti et al. 2017 [19]481.55434.28393.00355.03319.56287.79259.67233.78211.19190.07
Demir et al. 2023 [72]389.42344.26308.34276.01245.87219.35195.76173.81154.44136.74
FUnIE-GAN [43]424.67378.53339.95304.80274.28245.92221.17198.70178.44160.05
Original495.88450.94408.73369.22333.09300.13271.48245.50221.57199.82
WaterNet [31]517.97470.54427.50386.80349.24315.69285.87259.46235.03213.18
WaveNet [49]530.53465.93415.48373.71335.45302.53273.40246.36222.63201.85
Table A2. Average number of inliers for different frame offsets across various enhancement methods for the Seabed video.
Table A2. Average number of inliers for different frame offsets across various enhancement methods for the Seabed video.
Average Number of Inliers for Frame Offset:
Enhancement Method12345678910
Ancuti et al. 2017 [19]406.48370.15348.37330.85313.80297.83281.35266.19250.64235.75
Demir et al. 2023 [72]328.03294.69275.08257.64240.95225.57210.74195.76180.22166.76
FUnIE-GAN [43]369.97340.06321.03304.23287.30270.58254.42238.81222.39206.70
Original442.21406.62383.94364.30344.26325.53307.27289.50271.79254.33
WaterNet [31]440.29405.84383.18364.23344.62325.70307.73289.87272.23255.46
WaveNet [49]399.12354.27326.05303.41282.19262.95244.83226.72208.81192.47
For example, 495.88 for the original unenhanced video frames in column 1 of Table A1 means that, on average, there will be 495.88 inliers found between frame n and frame n + 1 throughout the whole (original unenhanced) video. As the offset increases from one to ten, fewer inliers will be found due to camera movement and marine snow and other artifacts in the case of underwater images.
This is a useful and vital measure, because at many-hours-long and 30 frames per second of high-definition footage, SLAM on underwater surveys is highly computationally expensive, and therefore, a way to accelerate the method is to reduce the number of frames. This experiment demonstrates that even three frames per second reduces accuracy and SLAM’s ability to track features.

Appendix B. ORB-SLAM 3 Results Extended

Appendix B.1. Motivation

SLAM on underwater imagery is challenging due to the poor visual quality. Therefore, it would seem natural to use enhancement methods on the video frames to improve videos prior to SLAM processing. This was the motivation for this work where SLAM performance is analysed in this section. But it was found that enhancement methods degraded the performance of SLAM over using the original video. There are no available ground truth videos with positional data in an underwater ocean environment (i.e., underwater video with ground truth position, e.g., collected from GPS, in open sea conditions). Therefore, surrogates need to be used to analyse SLAM performance, along with qualitative assessment of paths. We also developed the idea of using feature matching (as the key component of SLAM) as a quantitative measure of enhancement performance, and thus this appears in the main part of this paper, where we can rely on quantitative results rather than the qualitative discussion in this section.

Appendix B.2. Tracking Ability

We first compared the SLAM tracking ability in the original video with the videos produced from the frames enhanced by each method. We took the average tracking status across five runs of the original and each enhanced video. The tracking status is one of the pieces of data that ORB-SLAM 3 produces per frame, which we log during its operation and belongs to one of three statuses, initialising tracking, OK tracking, and lost/no tracking. The most consistent to register good tracking was the unaltered original video. For both the Cave and Seabed videos, all three deep-learning-based enhancement models managed to regain tracking, while the classical methods, on average, did not regain tracking. We saw variation in the number of frames needed to initialise. For the Cave video, Ancuti et al.’s enhancements proved to be the fastest at initialising, taking an average of 59 frames to start tracking, compared to an average of 370.6 frames for all other cave videos. However, in the case of the seabed video, the classical methods, including Ancuti et al. take the longest by far to initialise.
Table A3. The average number of frames to initiate tracking.
Table A3. The average number of frames to initiate tracking.
Frames to Initialise
Enhancement MethodCave (↓)Seabed (↓)
Original video36558
FUnIE-GAN [43]38198
Ancuti et al. 2017 [19]59270
Demir et al. 2023 [72]380235
Waternet [31]35789
WaveNet [49]37074
The poor performance on the Cave video prompted us to test how well each enhanced video would be initialised and tracked within a problematic area indicated by where all enhanced videos lost tracking. We performed this test starting from frame 3500. Despite this area being problematic for SLAM when in the context of the whole video, starting from this region resulted in improved initialisation and better tracking from all enhancement methods. However, the original video remained the only video that resulted in a complete survey, suggesting the enhancements still degraded the performance of SLAM tracking overall.

Appendix B.3. Trajectory and Loop Closures

The poor tracking performances for the enhanced videos are made more apparent by considering the recorded trajectories from SLAM. Figure A1 and Figure A2 show the average camera trajectory created from each video over five runs.
Figure A1. A top–down view of the average path for ‘Cave’ video.
Figure A1. A top–down view of the average path for ‘Cave’ video.
Sensors 25 06966 g0a1
Figure A2. A side view of the average path for ‘Seabed’ video.
Figure A2. A side view of the average path for ‘Seabed’ video.
Sensors 25 06966 g0a2
It is again clear that the original video results in the more complete SLAM survey, showing a clear, sustained path that largely matches the path visually seen in the video. A ground truth is unavailable for comparison as creating a ground truth for underwater imagery is extremely problematic; therefore, expert judgment on the path derived from SLAM, compared to the motion observed in the video is the best qualitative measure that we have. In the case of the enhanced cave videos, we see that no sustained paths were found by SLAM, and therefore there is a substantial difference between the original video and enhanced videos. More interestingly we find that in the case of the Seabed video, where tracking was better, there are sustained paths from all deep-learning enhancements but with clear divergences between them. Given the encouraging tracking performance for the original video, the drift in the enhanced video surveys suggests again that the enhancements are detrimental to SLAM.
Figure A3 shows the recorded camera trajectory when SLAM is initiated in the previously problematic region of the video at frame 3500. The result is a set of more sustained trajectories from all enhancement methods, but again there is strong evidence of drift. The figure shows that all enhancement methods resulted in half a loop before tracking was lost and then regained at the familiar beginning of the loop, preventing a full loop closure. Reinforcing this analysis are the loop closure event logs created by ORB-SLAM 3.
Figure A3. A top–down view of the average path for the ‘Cave’ video started at frame 3500.
Figure A3. A top–down view of the average path for the ‘Cave’ video started at frame 3500.
Sensors 25 06966 g0a3
We found that no enhancement method resulted in the capture of a single loop closure in five runs. Only the original video resulted in any loop closure detection, with one event being identified in three of five runs, and in all five when the video started at frame 3500. Despite being unable to complete a loop closure, Figure A3 shows that ORB-SLAM 3 was able to relocalise itself after losing track. This suggests that these enhanced videos may be capable of facilitating loop closures, but as the enhancements lead to inconsistent tracking, the likelihood of loop closure detection is reduced.

References

  1. Jalal, A.; Salman, A.; Mian, A.; Shortis, M.; Shafait, F. Fish detection and species classification in underwater environments using deep learning with temporal information. Ecol. Inform. 2020, 57, 101088. [Google Scholar] [CrossRef]
  2. Raveendran, S.; Patil, M.D.; Birajdar, G.K. Underwater image enhancement: A comprehensive review, recent trends, challenges and applications. Artif. Intell. Rev. 2021, 54, 5413–5467. [Google Scholar] [CrossRef]
  3. Schettini, R.; Corchs, S. Underwater image processing: State of the art of restoration and image enhancement methods. EURASIP J. Adv. Signal Process. 2010, 2010, 746052. [Google Scholar] [CrossRef]
  4. Zhang, W.; Dong, L.; Pan, X.; Zou, P.; Qin, L.; Xu, W. A survey of restoration and enhancement for underwater images. IEEE Access 2019, 7, 182259–182279. [Google Scholar] [CrossRef]
  5. Shadrin, S.S.; Varlamov, O.O.; Ivanov, A.M. Experimental autonomous road vehicle with logical artificial intelligence. J. Adv. Transp. 2017, 2017, 2492765. [Google Scholar] [CrossRef]
  6. Shruthi, P.; Resmi, R. Path Planning For Autonomous Car. In Proceedings of the 2019 2nd International Conference on Intelligent Computing, Instrumentation and Control Technologies (ICICICT), Kannur, India, 5–6 July 2019; IEEE: Piscataway, NJ, USA, 2019; Volume 1, pp. 1387–1390. [Google Scholar] [CrossRef]
  7. Ess, A.; Leibe, B.; Schindler, K.; Van Gool, L. Moving obstacle detection in highly dynamic scenes. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 56–63. [Google Scholar] [CrossRef]
  8. Bai, L.; Zhang, W.; Pan, X.; Zhao, C. Underwater image enhancement based on global and local equalization of histogram and dual-image multi-scale fusion. IEEE Access 2020, 8, 128973–128990. [Google Scholar] [CrossRef]
  9. Li, X.; Hou, G.; Tan, L.; Liu, W. A hybrid framework for underwater image enhancement. IEEE Access 2020, 8, 197448–197462. [Google Scholar] [CrossRef]
  10. Luo, W.; Duan, S.; Zheng, J. Underwater image restoration and enhancement based on a fusion algorithm with color balance, contrast optimization, and histogram stretching. IEEE Access 2021, 9, 31792–31804. [Google Scholar] [CrossRef]
  11. Park, T.H.; Eom, I.K. Sand-dust image enhancement using successive color balance with coincident chromatic histogram. IEEE Access 2021, 9, 19749–19760. [Google Scholar] [CrossRef]
  12. Jin, S.; Qu, P.; Zheng, Y.; Zhao, W.; Zhang, W. Color Correction and Local Contrast Enhancement for Underwater Image Enhancement. IEEE Access 2022, 10, 119193–119205. [Google Scholar] [CrossRef]
  13. Peng, Y.T.; Chen, Y.R.; Chen, Z.; Wang, J.H.; Huang, S.C. Underwater image enhancement based on histogram-equalization approximation using physics-based dichromatic modeling. Sensors 2022, 22, 2168. [Google Scholar] [CrossRef]
  14. Xiang, D.; Wang, H.; Zhai, C.; He, D. Research on Histogram Equalization Algorithm Based on Optimized Adaptive Quadruple Segmentation and Cropping of Underwater Image (AQSCHE). IEEE Access 2023, 11, 69356–69365. [Google Scholar] [CrossRef]
  15. Wang, H.; Frery, A.C.; Li, M.; Ren, P. Underwater image enhancement via histogram similarity-oriented color compensation complemented by multiple attribute adjustment. Intell. Mar. Technol. Syst. 2023, 1, 12. [Google Scholar] [CrossRef]
  16. Song, W.; Wang, Y.; Huang, D.; Liotta, A.; Perra, C. Enhancement of underwater images with statistical model of background light and optimization of transmission map. IEEE Trans. Broadcast. 2020, 66, 153–169. [Google Scholar] [CrossRef]
  17. Reza, A.M. Realization of the contrast limited adaptive histogram equalization (CLAHE) for real-time image enhancement. J. VLSI Signal Process. Syst. Signal Image Video Technol. 2004, 38, 35–44. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Zhou, L.; Li, H.; Zhu, J.; Du, W. Marine Application Evaluation of Monocular SLAM for Underwater Robots. Sensors 2022, 22, 4657. [Google Scholar] [CrossRef]
  19. Ancuti, C.O.; Ancuti, C.; De Vleeschouwer, C.; Bekaert, P. Color balance and fusion for underwater image enhancement. IEEE Trans. Image Process. 2017, 27, 379–393. [Google Scholar] [CrossRef] [PubMed]
  20. Lu, H.; Li, Y.; Nakashima, S.; Kim, H.; Serikawa, S. Underwater image super-resolution by descattering and fusion. IEEE Access 2017, 5, 670–679. [Google Scholar] [CrossRef]
  21. Li, C.; Tang, S.; Yan, J.; Zhou, T. Low-light image enhancement via pair of complementary gamma functions by fusion. IEEE Access 2020, 8, 169887–169896. [Google Scholar] [CrossRef]
  22. Lin, S.; Li, Z.; Zheng, F.; Zhao, Q.; Li, S. Underwater image enhancement based on adaptive color correction and improved retinex algorithm. IEEE Access 2023, 11, 27620–27630. [Google Scholar] [CrossRef]
  23. Qian, J.; Kong, B.; Yang, J. Underwater image clarification based on double-opponency light estimation and red channel prior. IEEE Access 2023, 11, 64383–64396. [Google Scholar] [CrossRef]
  24. Liu, X.; Gao, Z.; Chen, B.M. MLFcGAN: Multilevel feature fusion-based conditional GAN for underwater image color correction. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1488–1492. [Google Scholar] [CrossRef]
  25. Yang, H.H.; Huang, K.C.; Chen, W.T. LAFFNet: A lightweight adaptive feature fusion network for underwater image enhancement. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 685–692. [Google Scholar] [CrossRef]
  26. Tian, Y.; Xu, Y.; Zhou, J. Underwater image enhancement method based on feature fusion neural network. IEEE Access 2022, 10, 107536–107548. [Google Scholar] [CrossRef]
  27. Jiang, S.; Wang, W.; Su, Z.; Wang, S. Automatic Detection of Surface Defects on Underwater Pile-Pier of Bridges Based on Image Fusion and Deep Learning. Struct. Control Health Monit. 2023, 2023, 8429099. [Google Scholar] [CrossRef]
  28. Zhou, F.; Sun, X.; Dong, J.; Zhu, X.X. SurroundNet: Towards effective low-light image enhancement. Pattern Recognit. 2023, 141, 109602. [Google Scholar] [CrossRef]
  29. Jiang, X.; Yu, H.; Zhang, Y.; Pan, M.; Li, Z.; Liu, J.; Lv, S. An Underwater Image Enhancement Method for a Preprocessing Framework Based on Generative Adversarial Network. Sensors 2023, 23, 5774. [Google Scholar] [CrossRef]
  30. Wang, Y.; Guo, J.; Gao, H.; Yue, H. UIEC^2-Net: CNN-based underwater image enhancement using two color space. Signal Process. Image Commun. 2021, 96, 116250. [Google Scholar] [CrossRef]
  31. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An underwater image enhancement benchmark dataset and beyond. IEEE Trans. Image Process. 2019, 29, 4376–4389. [Google Scholar] [CrossRef]
  32. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar] [CrossRef]
  33. Wang, J.; Li, P.; Deng, J.; Du, Y.; Zhuang, J.; Liang, P.; Liu, P. CA-GAN: Class-condition attention GAN for underwater image enhancement. IEEE Access 2020, 8, 130719–130728. [Google Scholar] [CrossRef]
  34. Qi, Q.; Li, K.; Zheng, H.; Gao, X.; Hou, G.; Sun, K. SGUIE-Net: Semantic attention guided underwater image enhancement with multi-scale perception. IEEE Trans. Image Process. 2022, 31, 6816–6830. [Google Scholar] [CrossRef]
  35. Siddiqua, M.; Belhaouari, S.B.; Akhter, N.; Zameer, A.; Khurshid, J. MACGAN: An all-in-one image restoration under adverse conditions using multidomain attention-based conditional GAN. IEEE Access 2023, 11, 70482–70502. [Google Scholar] [CrossRef]
  36. Zhong, Y.; Wang, J.; Lu, Q. SCAUIE-Net: Underwater image enhancement method based on spatial and channel attention. IEEE Access 2023, 11, 72172–72185. [Google Scholar] [CrossRef]
  37. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar] [CrossRef]
  38. Panetta, K.; Kezebou, L.; Oludare, V.; Agaian, S. Comprehensive underwater object tracking benchmark dataset and underwater image enhancement with GAN. IEEE J. Ocean. Eng. 2021, 47, 59–75. [Google Scholar] [CrossRef]
  39. Xu, B.; Zhou, D.; Li, W. Image enhancement algorithm based on GAN neural network. IEEE Access 2022, 10, 36766–36777. [Google Scholar] [CrossRef]
  40. Cong, R.; Yang, W.; Zhang, W.; Li, C.; Guo, C.L.; Huang, Q.; Kwong, S. PUGAN: Physical model-guided underwater image enhancement using GAN with dual-discriminators. IEEE Trans. Image Process. 2023, 32, 4472–4485. [Google Scholar] [CrossRef]
  41. Qian, J.; Li, H.; Zhang, B.; Lin, S.; Xing, X. DRGAN: Dense Residual Generative Adversarial Network for Image Enhancement in an Underwater Autonomous Driving Device. Sensors 2023, 23, 8297. [Google Scholar] [CrossRef]
  42. Wang, H.; Zhang, W.; Xu, Y.; Li, H.; Ren, P. WaterCycleDiffusion: Visual–textual fusion empowered underwater image enhancement. Inf. Fusion 2026, 127, 103693. [Google Scholar] [CrossRef]
  43. Islam, M.J.; Xia, Y.; Sattar, J. Fast underwater image enhancement for improved visual perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
  44. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the MICCAI 2015 Proceedings, Part III 18. Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  45. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar] [CrossRef]
  46. Ignatov, A.; Kobyshev, N.; Timofte, R.; Vanhoey, K.; Van Gool, L. Dslr-quality photos on mobile devices with deep convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3277–3285. [Google Scholar] [CrossRef]
  47. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the Computer Vision–ECCV 2016 Proceedings, Part II 14. Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 694–711. [Google Scholar] [CrossRef]
  48. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  49. Sharma, P.; Bisht, I.; Sur, A. Wavelength-based attributed deep neural network for underwater image restoration. ACM Trans. Multimed. Comput. Commun. Appl. 2023, 19, 2. [Google Scholar] [CrossRef]
  50. Saleem, A.; Paheding, S.; Rawashdeh, N.; Awad, A.; Kaur, N. A non-reference evaluation of underwater image enhancement methods using a new underwater image dataset. IEEE Access 2023, 11, 10412–10428. [Google Scholar] [CrossRef]
  51. Wang, Z.; Shen, L.; Xu, M.; Yu, M.; Wang, K.; Lin, Y. Domain adaptation for underwater image enhancement. IEEE Trans. Image Process. 2023, 32, 1442–1457. [Google Scholar] [CrossRef]
  52. Sheikh, H.R.; Bovik, A.C. Image information and visual quality. IEEE Trans. Image Process. 2006, 15, 430–444. [Google Scholar] [CrossRef] [PubMed]
  53. Venkatanath, N.; Praneeth, D.; Bh, M.C.; Channappayya, S.S.; Medasani, S.S. Blind image quality evaluation using perception based features. In Proceedings of the 2015 National Conference on Communications (NCC), Mumbai, India, 27 February–1 March 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–6. [Google Scholar] [CrossRef]
  54. Panetta, K.; Gao, C.; Agaian, S. Human-visual-system-inspired underwater image quality measures. IEEE J. Ocean. Eng. 2015, 41, 541–551. [Google Scholar] [CrossRef]
  55. Han, Y.; Huang, L.; Hong, Z.; Cao, S.; Zhang, Y.; Wang, J. Deep supervised residual dense network for underwater image enhancement. Sensors 2021, 21, 3289. [Google Scholar] [CrossRef]
  56. Yang, M.; Sowmya, A. An underwater color image quality evaluation metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef] [PubMed]
  57. Yan, S.; Chen, X.; Wu, Z.; Tan, M.; Yu, J. HybrUR: A hybrid physical-neural solution for unsupervised underwater image restoration. IEEE Trans. Image Process. 2023, 32, 5004–5016. [Google Scholar] [CrossRef] [PubMed]
  58. Hidalgo, F.; Bräunl, T. Evaluation of several feature detectors/extractors on underwater images towards vSLAM. Sensors 2020, 20, 4343. [Google Scholar] [CrossRef]
  59. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  60. Bay, H.; Tuytelaars, T.; Van Gool, L. SURF: Speeded up robust features. In Proceedings of the Computer Vision–ECCV 2006: 9th European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; Proceedings, Part I 9. Springer: Berlin/Heidelberg, Germany, 2006; pp. 404–417. [Google Scholar] [CrossRef]
  61. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar] [CrossRef]
  62. Leutenegger, S.; Chli, M.; Siegwart, R.Y. BRISK: Binary robust invariant scalable keypoints. In Proceedings of the 2011 IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2548–2555. [Google Scholar] [CrossRef]
  63. Alcantarilla, P.F.; Nuevo, J.; Bartoli, A. Fast Explicit Diffusion for Accelerated Features in Nonlinear Scale Spaces. In Proceedings of the British Machine Vision Conference, Bristol, UK, 9–13 September 2013; BMVA Press: Sheffield, UK, 2013; pp. 13.1–13.11. [Google Scholar] [CrossRef]
  64. Ancuti, C.; Ancuti, C.O.; Haber, T.; Bekaert, P. Enhancing underwater images and videos by fusion. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 81–88. [Google Scholar] [CrossRef]
  65. Zhang, H. Removing Backscatter to Enhance the Visibility of Underwater Object. Master’s Thesis, Nanyang Technological University, Singapore, 2016. [Google Scholar]
  66. Alcantarilla, P.F.; Bartoli, A.; Davison, A.J. KAZE features. In Proceedings of the Computer Vision–ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Proceedings, Part VI 12. Springer: Berlin/Heidelberg, Germany, 2012; pp. 214–227. [Google Scholar] [CrossRef]
  67. Bradski, G. The OpenCV Library. Dr. Dobb’s J. Softw. Tools 2000, 25, 120–123. [Google Scholar]
  68. DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superpoint: Self-supervised interest point detection and description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 224–236. [Google Scholar] [CrossRef]
  69. Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar] [CrossRef]
  70. Fu, Q.; Yu, H.; Wang, X.; Yang, Z.; He, Y.; Zhang, H.; Mian, A. Fast ORB-SLAM without keypoint descriptors. IEEE Trans. Image Process. 2021, 31, 1433–1446. [Google Scholar] [CrossRef] [PubMed]
  71. Joshi, B.; Rahman, S.; Kalaitzakis, M.; Cain, B.; Johnson, J.; Xanthidis, M.; Karapetyan, N.; Hernandez, A.; Li, A.Q.; Vitzilaios, N.; et al. Experimental comparison of open source visual-inertial-based state estimation algorithms in the underwater domain. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 7227–7233. [Google Scholar] [CrossRef]
  72. Demir, Y.; Kaplan, N.H. Low-light image enhancement based on sharpening-smoothing image filter. Digit. Signal Process. 2023, 138, 104054. [Google Scholar] [CrossRef]
  73. Du, D.; Li, E.; Si, L.; Xu, F.; Niu, J. End-to-end underwater video enhancement: Dataset and model. arXiv 2024, arXiv:2403.11506. [Google Scholar] [CrossRef]
Figure 1. An example of matched features showing ideal feature tracing from two close frames in the ‘Seabed’ video.
Figure 1. An example of matched features showing ideal feature tracing from two close frames in the ‘Seabed’ video.
Sensors 25 06966 g001
Figure 2. The initial stage of the pipeline, frames are processed with an enhancement model. The chosen feature detector is then used to obtain all features for all frames.
Figure 2. The initial stage of the pipeline, frames are processed with an enhancement model. The chosen feature detector is then used to obtain all features for all frames.
Sensors 25 06966 g002
Figure 3. Feature matching process for two frames. The local matching stability and furthest matchable frame measures are computed using information about the matches.
Figure 3. Feature matching process for two frames. The local matching stability and furthest matchable frame measures are computed using information about the matches.
Sensors 25 06966 g003
Figure 4. Sample frames from ‘Cave’ [71] and ‘Seabed’ videos.
Figure 4. Sample frames from ‘Cave’ [71] and ‘Seabed’ videos.
Sensors 25 06966 g004
Figure 5. Average number of inliers across subsequent frames offset (1–10) from the subject frame for ORB detector in the Cave dataset.
Figure 5. Average number of inliers across subsequent frames offset (1–10) from the subject frame for ORB detector in the Cave dataset.
Sensors 25 06966 g005
Figure 6. Average number of inliers across subsequent frames for ORB detector in the Seabed dataset. Note that the line for WaterNet (top performer) almost perfectly overlaps the line for the original unenhanced video.
Figure 6. Average number of inliers across subsequent frames for ORB detector in the Seabed dataset. Note that the line for WaterNet (top performer) almost perfectly overlaps the line for the original unenhanced video.
Sensors 25 06966 g006
Figure 7. Cumulative distribution of matching distance for ORB feature extraction in the Cave dataset. Higher lines indicate the enhancement, on average, allows frames to match with frames further ahead in the video than lower lines.
Figure 7. Cumulative distribution of matching distance for ORB feature extraction in the Cave dataset. Higher lines indicate the enhancement, on average, allows frames to match with frames further ahead in the video than lower lines.
Sensors 25 06966 g007
Figure 8. Cumulative distribution of matching distance for ORB feature extraction in the Seabed dataset. Higher lines indicate the enhancement, on average, allows frames to match with frames further ahead in the video than lower lines.
Figure 8. Cumulative distribution of matching distance for ORB feature extraction in the Seabed dataset. Higher lines indicate the enhancement, on average, allows frames to match with frames further ahead in the video than lower lines.
Sensors 25 06966 g008
Table 1. The SSIM and PSNR averages across the Seabed video. Higher is better for both SSIM and PSNR. Methods indicated by † are physically based.
Table 1. The SSIM and PSNR averages across the Seabed video. Higher is better for both SSIM and PSNR. Methods indicated by † are physically based.
MethodSSIM (↑)PSNR (↑)
FUnIE-GAN [43]0.875321.68
Ancuti et al. [19] †0.839424.15
Demir and Kaplan [72] †0.858522.83
WaterNet [31]0.920924.85
WaveNet [49]0.713220.72
UVENet [73]0.628717.46
Table 2. Average furthest matching frame (FMF) values for the ‘Cave’ video. A value of k in the table means that on average for the whole video each frame matches with the k-th frame ahead. Methods indicated by † are physically based.
Table 2. Average furthest matching frame (FMF) values for the ‘Cave’ video. A value of k in the table means that on average for the whole video each frame matches with the k-th frame ahead. Methods indicated by † are physically based.
Detector
Enhancement MethodAKAZE (↑)BRISK (↑)KAZE (↑)ORB (↑)Super Point (↑)
Ancuti et al. 2017 [19] †20.0415.4125.0316.2920.85
Demir et al. 2023 [72] †15.2910.7620.9012.5722.19
FUnIE-GAN [43]23.9819.6627.0914.8521.28
Original24.2622.9327.5617.6820.31
UVENet 2024 [73]20.7822.6726.5915.3522.23
WaterNet [31]24.5323.5829.3019.2520.67
WaveNet [49]23.8622.0127.6317.6721.10
Table 3. Average FMF values for the ‘Seabed’ video.
Table 3. Average FMF values for the ‘Seabed’ video.
Detector
Enhancement MethodAKAZE (↑)BRISK (↑)KAZE (↑)ORB (↑)SIFT (↑)Super Point (↑)
Ancuti et al. 2017 [19]38.5551.5264.5531.2836.5727.74
Demir et al. 2023 [72]29.5341.7543.8418.7820.9527.51
FUnIE-GAN [43]38.4953.0854.8026.4126.5524.52
Original41.3558.4757.1330.9331.4920.89
UVENet [73]39.2055.1956.0929.3324.7928.52
WaterNet [31]42.1261.8060.2133.0432.5121.82
WaveNet [49]36.4953.1249.9120.1723.3822.69
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Summers, J.M.; Jones, M.W.; Seale, C. Impact of Underwater Image Enhancement on Feature Matching. Sensors 2025, 25, 6966. https://doi.org/10.3390/s25226966

AMA Style

Summers JM, Jones MW, Seale C. Impact of Underwater Image Enhancement on Feature Matching. Sensors. 2025; 25(22):6966. https://doi.org/10.3390/s25226966

Chicago/Turabian Style

Summers, Jason M., Mark W. Jones, and Catherine Seale. 2025. "Impact of Underwater Image Enhancement on Feature Matching" Sensors 25, no. 22: 6966. https://doi.org/10.3390/s25226966

APA Style

Summers, J. M., Jones, M. W., & Seale, C. (2025). Impact of Underwater Image Enhancement on Feature Matching. Sensors, 25(22), 6966. https://doi.org/10.3390/s25226966

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop