1. Introduction
A considerable part of the earth is covered with water, with the underwater world consisting of an astounding variety of resources. However, the underwater world is not as friendly as the atmospheric region. In order to explore underwater resources and discover the aquatic world, the use of Remotely Operated Vehicles (ROV), such as underwater robots and submarines with proper underwater cameras may be required. The challenges of acquiring undistorted underwater images, which primarily focus on the object of interest, are well-documented due to the distortions caused by marine organisms, floating objects, marine snow, bacteria and algae present in sea water. These sources of distortion make the captured underwater images less informative, and consequently, have limited applicability for some underwater applications, which require undistorted or less distorted images, such as edge detection, feature point detection, image classification, image stitching, object detection, underwater studies, underwater archaeology, marine ecology, assisting aquatic robots, species recognition, and underwater geology. As such, there is a requirement for underwater image capture to have a high degree of accuracy and quality for proper interpretation of its information.
Underwater images are commonly dominated by blue and green shades, as red light from the visible spectrum is quickly absorbed and loses its strength, even in the first part of the ocean, which is within 10 m depth. Other lights from the visible spectrum, such as orange, yellow, green, and blue, are also absorbed by the water as we go deeper into the ocean.
Figure 1 shows the light absorption property of water and penetration levels of lights [
1], at various depths. Due to these problems, a proper method that is able to restore underwater images is the need of the hour and is required for various studies and scientific research areas. By gaining more information from the images, the underwater images can be used for different underwater applications.
Many restoration algorithms have been proposed in the literature. Reference [
2] provides a review of various underwater image restoration methods that are available in the literature, generally classifying underwater restoration methods into hardware, software, and network-based approaches.
Hardware-based approaches employ a variety of hardware to process underwater images for restoration purposes. These include range-gated imaging techniques [
3], polarisers [
4], imaging using stereo cameras [
5], and remotely operated vehicles [
6]. However, these methods have been shown to suffer from errors caused by calibration of the hardware devices.
Network-based approaches involve the use of deep-learning algorithms to process underwater images. Convolutional neural networks [
7,
8] and generative adversarial networks [
9,
10] are some of the neural networks that have been used for this purpose. However, deep-learning methods commonly require a good dataset with a large number of underwater images, together with ground truth images, which are very difficult to acquire in the case of underwater image processing.
Software-based approaches use the Image Formation Model (IFM) to restore the captured underwater images, by finding the background light and transmission maps. Dark Channel Prior (DCP), as proposed by He et al. [
11], uses IFM in image restoration, by assuming scene points closer to the camera as dark images and vice versa. However, due to the longer wavelength and faster attenuation property of red light, the method fails to estimate the proper results, and always ends up choosing the red channel as the darkest of all channels. Variations of DCP have consequently been proposed for underwater images; using green and blue channels only, [
12,
13,
14], using the inverse of the red channel [
15], and using the maximum intensity prior [
16]. The performances of these methods have been shown to vary depending on different lighting conditions and priors chosen. Instead of estimating the transmission map directly, Peng et al. [
17] use a depth-estimation strategy to restore the underwater images. The proposed method involves the use of depth estimation and transmission map estimation with attenuation coefficient priors, by considering the backscattering effect, and has been proven to show superior results [
18].
Generally, the efficiency of an image-processing algorithm is calculated by comparing its output processed underwater images to other similar algorithms, and by using quality metrics, such as Peak Signal to Noise Ratio (PSNR), Mean Squared Error (MSE), and dedicated underwater performance metrics, including Underwater Color Image Quality Evaluation (UCIQE) [
19] and Underwater Image Quality Measure (UIQM) [
20]. These are more common measures in evaluating image processing methods. However, there is no approach to evaluate an image processing algorithm based on its applicability in real applications. This is very important since the motive of developing an image-processing method is not only to restore or enhance the underwater images, but ultimately, to help improve the efficiency of the real applications. In this paper, the first ever approach to evaluating the efficiency of the proposed algorithm based on its usefulness for underwater applications, has been shown.
The contributions of the paper are: (1) proposing an underwater image restoration method, which estimates depth maps using combinations of a blurriness map, background light neutralization, and red-light intensity. The background light neutralization is estimated using the four-quadrant method, which demands lower computation compared with other methods [
18], and (2) demonstrations of the restored underwater image using the proposed method on different underwater applications to evaluate the efficiency of the algorithm. This represents the first ever demonstration of a developed algorithm on real underwater applications.
The structure of the rest of the paper is as follows:
Section 2 describes the proposed underwater image restoration method that may be used to effectively recover original images from acquired underwater images. Consequently, the usages of the recovered underwater images on different underwater applications are explored in
Section 3.
Section 4 discusses results from the proposed restoration method, as well as its implementation on the selected underwater applications. The last section concludes the paper.
2. Proposed Restoration Method
The image restoration process employs the Image Formation Model (IFM) given in Equation (1), to obtain the original scene from a captured underwater scene, with the process involving estimation of the different parameters of the underwater IFM.
As can be seen, there are two distinct parts of the captured underwater image . describes radiance of the object as it travels in the underwater medium, whilst represents the scattering of background light as it travels towards the camera. Transmission map describes the part of the object radiance that reaches the camera, after considering for absorption and scattering.
Recovering the original object radiance
from the acquired image
at the camera requires knowledge of the background light
as well as the transmission map
, with this information commonly estimated. Taking
and
as the estimated transmission map and background light, respectively, the recovered scene radiance
may be estimated as:
where,
is the spectral attenuation coefficient of the direct signal and
is the estimated depth map of the image.
Figure 2 depicts the flowchart of the proposed underwater restoration method for estimating the recovered scene radiance
from the captured underwater image
. Blurriness estimated image
and background light neutralized image
, are calculated from the input image
, which are then used, together, with the red-light intensity
, to estimate depth
of the underwater image. Subsequently, transmission map
may then be estimated using the estimated depth map
by selecting the appropriate spectral attenuation coefficients. The input image
, estimated background light
, and transmission map
, are then used to find the final scene radiance recovered image
, as per Equation (2).
2.1. Depth Estimation and Background Light Estimation
Blurriness map estimation
is the first process in the restoration process, by estimating the refined blurriness map, through the initial map and rough map of the image [
17]. This is then followed by background light estimation. To determine the background light, the input image
is segmented into four quadrants, and the mean value of the pixels calculated. Equation (4) is then used to estimate background light
,
where
The selected pixel, which constitutes the estimated background light , may not be the brightest of all pixels in the entire input image, as two quadrants with extremes light intensity have been excluded from the selection process. The estimated background light shall be used for the scene radiance recovery using Equation (2).
Background light neutralized image
needs to be estimated to find the depth map of the underwater image. Initially, average light intensity
in the two quadrants, excluding the two extremes, is determined, and taken as average of the underwater image,
where
is the four quadrants
, with
representing the average light intensity in the respective quadrant
. The brightest
and darkest
quadrants are neglected, as they are two extremes of the spectrum. Average light intensity in the remaining two quadrants is then calculated, and taken as average of the underwater image:
This average light intensity
is then used to modify all the pixels of the input image
to retrieve the contrast neutralized image
, as follows:
To denoise the image, discrete wavelet transform (DWT) is applied on the contrast-neutralized image and the gray version of the input image . Inverse discrete wavelet transform (IWDT) is finally applied to retrieve the background light-neutralized image, with approximation and detailed coefficients modified based on the average of approximation coefficients and max rule applied on detailed coefficients.
The blurriness map, background light-neutralized image, and intensity of the red channel can then be used for the depth-estimation process [
18]. The maximum intensity of the red channel, known as red channel map
of the image, is represented by
where
is the intensity of the red channel and
is a square local patch centred at
x. The factors used for estimating depth are passed through a stretching function given by Equation (10) [
18].
where
can either be the red channel map
, blurriness map
or background neutralised image
, to give
,
and
, respectively.
is a stretching function, which accepts vector
as its input.
The final depth estimation can be found by Equation (12),
where
θb and
θa are
and
, respectively; with
(.) giving average of the input and the sigma functions
S(a,v) given as:
2.2. Transmission Map Estimation
The proposed transmission map estimation involves the use of depth estimation in Equation (3). Reference [
17] estimates transmission map using only the direct signal, with the effects of backscattered signals neglected. In contrast, the proposed transmission map estimation involves the use of both direct and backscattered signals in estimating the transmission map, as shown in Equation (14).
where
is the spectral attenuation coefficient of the direct signal and
is the spectral attenuation coefficient of the backscattered signal.
2.2.1. Transmission Map of Direct Signal
The transmission map of direct signal is estimated using spectral attenuation coefficients calculated for red, green and blue channels, together with the calculated depth map
. The transmission map for the red channel can be calculated using,
Restoration results are not sensitive to spectral attenuation coefficient
of the red channel [
17], with values between [0.125,0.20] for oceanic water type I [
21], and hence, spectral coefficient value
of the red channel is set to 0.142.
The transmission map for the green and blue channels due to direct signal can be found by utilizing the transmission and attenuation coefficient of the red channel [
22],
where the linear relationship between the attenuation coefficients of the green, blue, and red channels is given by Equation (17), with values
, and
[
23]. Wavelengths for the red, green and blue light are taken to be 620 nm, 540 nm, and 450 nm, respectively [
17].
where
is the background light estimated using Equation (4) for the respective channel
.
2.2.2. Transmission Map of Backscattered Signal
Comprehensive studies have been conducted on the estimation of spectral attenuation backscattering coefficients, with Mie theory used to predict spectral behavior. Whitmire et al. [
24] use Slow Descent Rate Optical Profiler (Slow DROP), to experimentally calculate backscattering coefficients of particulate matters in five research cruises at five different wavelengths covering the visible spectrum, over a period of three years. The values selected based on the wavelengths of interest are shown in
Table 1.
Total backscattering coefficient
is a summation of pure water backscattering coefficient
and particulate matter backscattering coefficient
,
Transmission map due to backscattered signal may be derived from Equations (3) and (18) as follow,
Spectral attenuation coefficients of the direct
and backscattered
signals may be used to estimate the raw transmission map, using Equation (14). The estimated transmission is then further refined by using a guided filter [
26], instead of soft matting [
11], because of its better refinement properties.
Scene radiance recovery involves the use of the estimated background light and transmission map to form the final scene radiance. The refined transmission map is used in Equation (2) to acquire the final restored image.
3. Different Underwater Applications
There are many applications of underwater images, out of which three of the most common applications have been chosen for evaluation of the proposed underwater image restoration method. The three applications are edge detection, Speeded Up Robust Feature (SURF), and image classification using machine learning (ML). One of the main aims of a restoration method is to reduce blurriness from underwater images; among other things, to facilitate edge detection, which may be performed with the use of Sobel edge-detecting operator. Edge detection is mainly used for obstacle detection by unmanned underwater vehicles. Textured details of underwater images may also be improved by using a restoration method, and the effectiveness of the restoration method towards this objective may be evaluated by considering the number of feature points detected by SURF. For underwater images with coral reefs and fish with a variety of shapes and sizes, SURF is used to detect features of objects, which may be performed with the help of the SURF function in MATLAB software. Finally, image classification is performed to prove the effectiveness of a restoration method in detecting targets. This application is specifically used for underwater pipeline corrosion, marine sea salt detection, subsea terrain classification and mineral exploration.
3.1. Edge Detection
Since underwater images are used in pattern recognition, image decomposition, visual inspection, and also in important processing tasks in computer vision related to underwater segmentation, the output underwater image needs to be clear, with good texture and details. Edge detection is particularly useful in underwater image processing, in order to localize coral reefs and other related tasks. In this paper, the Sobel edge detector is used to detect the number of edges in an image [
27], whereby output from the proposed restoration method is used as input to the Sobel edge detector, in order to ascertain whether the proposed restoration method actually improves features and texture details of objects in an underwater image. A comparison of performances is then made with the input image, in terms of the number of edges detected.
3.2. Speeded Up Robust Feature (SURF)
SURF is a detection algorithm used to detect points of interest in an underwater image. One of the basic tasks of computer vision algorithms is local feature points matching, which forms the basis for underwater studies, such as for the detection of marine animals, and fish species recognition [
28]. The SURF feature matching provided by MATLAB software is used for performance analysis. For the detection of points of interest, SURF uses an integer approximation of the determinant of Hessian blob detector that is determined using a three-integer operation on a precomputed integral image. The feature descriptor is based on the Haar wavelet response. This can also be used underwater to detect and locate objects, reconstruct 3D scenes and extract points of interests.
3.3. Image Classification
Image classification is an important application in image processing. It segregates objects in an underwater image based on the object of interest, which can be useful in various fields, such as for the detection of pipeline corrosion, marine salt, fish detection, detection of ship wrecks, mineral exploration, marine animal detection, pollution monitoring, subsea investigation, and sea floor terrain examination.
In machine learning, the model learns a pattern in a dataset, from which prediction of a given situation of interest can be made. The learning process starts by providing a training dataset, which is fed to a designed model to establish the relationship between dependent and independent variables. Once trained, the pretrained model may then be used to predict the output given a set of test inputs.
Machine-learning methods are generally classified as either supervised, unsupervised, semisupervised or reinforcement learnings. A supervised machine-learning method is used here for the image classification purpose. The method predicts a new set of outputs based on what has been learned from the past training datasets, with the learning process starting with the help of training data that has a set of input and target vectors. In supervised learning, it is assumed that the actual output values are known for each input pattern.
In the process of classification, image features of the input images from the training set are extracted using the Discrete Wavelet Transform (DWT) and Gray Level Co-occurrence Matrix (GLCM), with these features fed into the classification algorithm for training purposes. Two supervised machine-learning algorithms: Support Vector Machines (SVM) and K-Nearest Neighbor (KNN) have been chosen. SVM has evolved as one of the most powerful supervised machine-learning methods in classification problems and linear and nonlinear regressions, whilst KNN is a nonparametric statistical method for classification and regression problems, and has been used for pattern recognition and feature detection [
29].
5. Conclusions
A proposed restoration method for underwater images is given in this paper, which involves depth estimation from blurriness estimation and background light neutralization process as well a transmission map estimation using direct and backscattered signals. Any method that is designed for restoring the image has to provide efficient results not only in terms of quantitative and qualitative approach, but also in practical applications, which serves the whole purpose of developing the method in the first place. For this reason, the proposed underwater image restoration method has been tested on direct underwater applications. The applications are chosen in such a way that the claims of the proposed methods are proven. Edge detection, SURF and image classification have been chosen. The edge detection has been performed using the Sobel edge-detection operator on raw and restored images using the proposed method, and it has been shown that the number of edges detected on the restored images are always higher than on the original raw images, proving that the proposed restoration method is able to reduce blurriness in an image. Similarly, the SURF function from MATLAB has been used on the gray version of raw and restored underwater images. Subsequently, it has been shown that the number of SURF points detected on the restored underwater images increases. This implies the ability of the proposed method to improve texture details of an underwater image. Finally, image classification using a supervised machine-learning approach has been performed on a set of 75 test images before and after restoration, whereby the restored images give better classification results in terms of accuracy when used for image classification using both SVM and KNN.
These results on different underwater imaging applications show that the recovered images from the proposed method provide good and efficient results when compared to unprocessed raw input images. As such, the proposed method can be considered as an important step to recover underwater images before applying them to underwater applications, with an efficient outcome.