Next Article in Journal
Remote Sensing Monitoring of Ecological-Economic Impacts in the Belt and Road Initiatives Mining Project: A Case Study in Sino Iron and Taldybulak Levoberezhny
Next Article in Special Issue
MIMO Radar Sparse Recovery Imaging with Wideband Interference Prediction
Previous Article in Journal
New Archaeological Discoveries Based on Spatial Information Technology and Cultural Analysis: Taking the Study of the Spatial Relationship between Ancient Chinese Capitals and the Natural Environment as an Example
Previous Article in Special Issue
Unambiguous ISAR Imaging Method for Complex Maneuvering Group Targets
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bubble Plume Target Detection Method of Multibeam Water Column Images Based on Bags of Visual Word Features

1
College of Civil Engineering, Anhui Jianzhu University, Hefei 230601, China
2
School of Resources and Environmental Engineering, Anhui University, Hefei 230601, China
3
Engineering Center for Geographic Information of Anhui Province, Anhui University, Hefei 230601, China
4
School of Geodesy and Geomatics, Wuhan University, Wuhan 430079, China
5
Institute of Marine Science and Technology, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(14), 3296; https://doi.org/10.3390/rs14143296
Submission received: 31 May 2022 / Revised: 27 June 2022 / Accepted: 5 July 2022 / Published: 8 July 2022
(This article belongs to the Special Issue Radar and Sonar Imaging and Processing Ⅲ)

Abstract

:
Bubble plumes, as main manifestations of seabed gas leakage, play an important role in the exploration of natural gas hydrate and other resources. Multibeam water column images have been widely used in detecting bubble plume targets in recent years because they can wholly record water column and seabed backscatter strengths. However, strong noises in multibeam water column images cause many issues in target detection, and traditional target detection methods are mainly used in optical images and are less efficient for noise-affected sonar images. To improve the detection accuracy of bubble plume targets in water column images, this study proposes a target detection method based on the bag of visual words (BOVW) features and support vector machine (SVM) classifier. First, the characteristics of bubble plume targets in water column images are analyzed, with the conclusion that the BOVW features can well express the gray scale, texture, and shape characteristics of bubble plumes. Second, the BOVW features are constructed following steps of point description extraction, description clustering, and feature encoding. Third, the quadratic SVM classifier is used for the recognition of target images. Finally, a procedure of bubble plume target detection in water column images is described. In the experiment using the measured data in the Strait of Georgia, the proposed method achieved 98.6% recognition accuracy of bubble plume targets in validation sets, and 91.7% correct detection rate of the targets in water column images. By comparison with other methods, the experimental results prove the validity and accuracy of the proposed method, and show potential applications of our method in the exploration and research on ocean resources.

Graphical Abstract

1. Introduction

As one of the most widely used ocean survey equipment, multibeam sonars can acquire the full coverage of underwater topographic survey and simultaneously construct seabed backscatter strength images. In recent years, recording and processing the whole water column backscatter data using multibeam sonars have become possible, with the rapid improvement of computer processing power and storage capacity [1].
Multibeam water column images have been used to investigate marine biological resources (fish/fish school [2,3], krill [4], and seabed eelgrass [5,6]), study ocean physical changes such as internal waves [7,8], find underwater shipwrecks [9,10], and detect gas clouds [11] and bubble plumes [12,13,14,15] from seabed seeps. Innangi et al. [16] proposed a detection method of fish schools using multibeam water column images and obtained 3D shapes of various fish schools. Hughes Clarke et al. [9,17,18] studied the imaging principle, noise interference, and other basic knowledge of multibeam water column imaging, which laid a foundation for subsequent research. They also used water column image to estimate the minimum depth of a shipwreck (mast peak), which could provide a safety guarantee for shipping. Marques [19] used multibeam water column images to recognize and precisely locate spherical suspended targets and monitor seabed landslide activities. Gardner et al. [20] found a bubble plume with a depth of 1830 m and a height of 1400 m by using water column images collected from Kongsberg EM302 multibeam sonar in the coastal area of northern California. Weber et al. [21] discovered a bubble plume with 1100 m height using water column images from Kongsberg EM302 in the Gulf of Mexico. Various applications, especially in marine resource exploration and environment research of multibeam water column images have been proven by these studies.
Bubble plumes are the main manifestations of seabed gas leakage. The leaked gases could be the natural gas (methane), meaning the potential storage of natural gas hydrates. The gases also could be the mixed gases produced by decomposition of underwater dead plants and animals. Therefore, bubble plumes play important roles in the exploration of important resources (like natural gas hydrate) and the research of underwater environment [22,23,24,25]. How to accurately and efficiently recognize and detect bubble plumes in multibeam water column images is a hot research issue at present [26].
To recognize targets in sonar images, classifying the special target features has been proven as a feasible solution. Many studies have focused on detection and recognition of sonar image targets using classifiers of image features. Dobeck et al. [27] achieved automatic mine target detection and recognition in sonar images by using a k-nearest neighbor classifier of 45 extracted features, such as shape and intensity of mine targets. Tang et al. [28] studied the multi-channel texture classification algorithm by applying wavelet packet and Fourier transform on side-scan sonar images. The experimental results showed that the feature extraction method is vital for image classification. Reed et al. [29] achieved the recognition of sand slope targets in side-scan sonar images based on texture features, standard classifier, and Markov random field. Rhinelander et al. [30] combined feature extraction, edge filter, median filter, and support vector machine (SVM) classifier to achieve target detection and recognition in side-scan sonar images. Song et al. [31] studied the sonar image target segmentation method based on Markov random field and extreme learning machine. Wang [32] studied a target detection algorithm in side-scan sonar images using feature extraction method based on a neutrosophic set and diffusion maps. Moreover, a combination of various feature extraction methods (gray-level co-occurrence matrix, local binary pattern, Gabor, Tamura, multifractal spectrum, and others) and various classifiers (AdaBoost, back-propagation neural network, convolutional neural network, and others) have been applied on various targets in different sonar images.
In conclusion, numerous target feature extraction methods and classifiers have been used for target recognition and detection in sonar images. For various target recognition in different types of sonar images, the optimal features and classifiers are quite different. For the bubble plume target in multibeam water column, the optimal features and classifiers have yet to be studied. Zhao et al. [33] used Haar-like and local binary pattern feature and AdaBoost classifier to recognize the gas plume targets in multibeam water column images. This feature is suitable for vertical and strong-strength bubble plume targets, but recognition accuracy of inclined and weak-strength bubble plume target still needs to be improved.
In recent years, deep learning methods have been widely applied in recognition and detection of sonar image targets [34,35]. Deep learning methods usually need a large number of samples and labels to ensure model accuracy and generality. However, due to the particularity of sonar data collection and the difficulty in finding underwater targets, it is difficult to guarantee the sufficient number of sonar image target samples, which brings problems to the accuracy target recognition [36,37,38]. Another feasible solution is automatic sample augmentation using deep learning methods (like generative adversarial networks) in the case of small or even zero samples [39,40,41]. However, differences remain between the augmented and real samples, and the accuracy improvement of target recognition and detection using augmented samples still needs to be studied.
Therefore, in this article, we analyze the characteristics of bubble plume, select an applicable feature extraction algorithm and reasonable classifier, and finally realize the automatic accurate recognition and detection of bubble plume targets in multibeam water column images.

2. Theories and Methods

2.1. Basic Principle of Target Detection in Multibeam Water Column Images

The multibeam echo sounder system contains multiple sub-systems, including global navigation satellite system (GNSS), motion reference unit (MRU), gyro-magnetic compass, the bathymetric sonar and other auxiliary sensors [42,43]. Multibeam sonars emit sound waves through the projector arrays and then receive the backscatter strengths at multiple angles through the hydrophone arrays. By continuously recording the backscatter strengths from the transducer to the seabed and even under the seabed, we can obtain the observation of the whole water column, the seabed and surface sediments. The bubble plumes are some of the most important targets in the water column image because they could indicate the possible storage of natural gas hydrates or decomposition of underwater dead plants and animals. The bubble plumes usually were leaked from the seabed, rise to a certain height, and then dissipate. When the sound waves travel through these two different propagation medias (i.e., gas and water), the backscatter strengths from the bubbles are stronger than those from surrounding water, as shown in Figure 1.
In the multibeam water column images, the bubble plume targets are different from the background noises (Figure 1B). Some characteristics of the bubble plume targets in the water column images can be analyzed and concluded as follows:
  • The bubble plume targets have much brighter grayscale features than the water column image background because the backscatter strengths from the gas bubbles are much stronger than those from the surrounding water;
  • The bubble plume targets have special shape features. These targets are generally plume-like or ribbon-like shapes, which are obviously different from other water column targets (such as fish and fish school);
  • The bubble plumes also have special orientation features. The bubble plumes usually range from the seabed to a certain height, and are usually approximately perpendicular to the seabed, but may be bent and oblique due to the ocean current effects. Due to side-lobe effects, bubble plumes in the minimum slant range are easier to detect;
  • Special texture features exist around bubble plume targets. Due to the changes of two different propagation media, the texture features of the bubble plume and surrounding water are quite different.
These feature differences between the bubble plumes and background noises are the basis for the target recognition and detection method in a multibeam water column image.
As analyzed, the bubble plume targets have special grayscale, shape, orientation, and texture features. In the following sections, the bag of visual words (BOVW) features of bubble plume targets are extracted and used to represent these features, and the SVM classifier is used for binary classification of the BOVM features. Then, the recognition and detection procedures of bubble plume targets in the water column image are introduced in detail.

2.2. BOVW Features of Multibeam Water Column Images

The BOVW features were originally used for text classification by calculating the frequencies of important words in the text. The BOVW feature applies the same idea into image classification [44]. The images do not contain any concrete words; therefore, visual vocabulary needs to be constructed using feature point extraction methods, such as speeded up robust features (SURF) (Figure 2B), then the visual vocabulary is established by clustering these SURF descriptors (Figure 2C), and finally, the BOVW features can be encoded by calculating the frequencies of each visual words in the visual vocabulary (Figure 2D). These steps are shown in Figure 2 as:
  • Step 1. All the SURF descriptors (64-dimensional eigenvectors) of interest points from sample images are extracted by SURF detection or uniform-grid-point selection;
  • Step 2. Due to the excessive number SURFs extracted, the SURFs of all sample images were clustered using k-means to obtain k category;
  • Step 3. For any image, the SURF points can be extracted using the interest point selection (Step 1) and calculated as 64-dimensional eigenvectors, then these SURF descriptions are classified into each category (Step 3);
  • Step 4. The occurrence frequencies of all the SURF clustering categories are calculated to form the k-dimensional BOVW feature vector.

2.2.1. SURF Descriptors Extraction

The visual vocabulary of BOVW features need to be constructed by clustering the extracted point features from all different categories of sample images. SURF, as a scale invariant feature, can be used as both detectors and descriptors of interest points at any scale [45]. In multibeam water column image, SURF is also suitable for the detection and description of bubble plume target feature points [46].
The SURF method uses the Hessian matrix H p , σ to detect and describe the interest points p(x, y) at scale σ as
H p , σ = L x x p , σ L x y p , σ L y x p , σ L y y p , σ ,
where L is the convolution of the Gaussian second-order derivative with the image.
As shown in Figure 3, detected SURF points from the target images and background noises are significantly different in feature scale, orientation and distribution, which proves the effectiveness of SURF points as the visual words of water column images. However, affected by different kinds of noises, detected SURF points from different water column images could vary greatly under the same Hessian threshold. If SURF detection is used to find interest points, some features of the target and background images could be ignored. Therefore, we use the uniform-grid-point selection to ensure that all the features in the target and background image can be extracted (Figure 2B).
The SURF descriptors are used to describe the uniform-grid selected points in the water column images. Before feature description, the main orientation of the descriptor is set as upright because no rotation is applied in the water column images. The steps for extracting SURF descriptors are listed as follows:
  • Select the regions around key points and divide them into 4 × 4 small regions;
  • Calculate Four features (dx, dy, |dx| and |dy|) of the sampling point corresponding to the Haar response in each small region;
  • Construct and normalize the 64-dimensional (4 × 4 × 4) eigenvector of each key point.
Therefore, all the SURF descriptors of the uniform-grid selected points in the image can be described using the normalized 64-dimensional eigenvectors (Figure 2B). The grid size would determine the number of extracted points, which would be discussed in the experimental section.

2.2.2. Clustering SURF Descriptors

The visual vocabulary of water column images is constructed based on SURF descriptors of uniform-grid selected points. For sample sets containing many images, all of the SURF descriptors (i.e., the 64-dimensional eigenvector) of each image need to be extracted and summarized to obtain the special BOVW features that can describe the bubble plume targets. Moreover, these extracted SURF descriptors need to be clustered to limit the BOVW feature dimension. k-means, as an efficient clustering method, is suitable for clustering these SURF descriptors. The steps are the following:
  • The k-mean++ algorithm is used to initialize k centroids c(j) (j = 1, …, k);
  • The Euclidean distance Lj between each SURF 64-dimensional eigenvector v and the centroids c(j) is calculated as
L j = i = 1 64 v i c j i 2 , j = 1 , , k .
3.
Each eigenvector is assigned to the nearest centroid and the k centroids are recalculated;
4.
Steps 2 and 3 are repeated until all the cluster assignments are stable (Figure 4B).
The clustering number k determines the visual vocabulary number and the dimension of BOVW features. Considering the noises affecting water column images, an extremely small visual vocabulary number will lead to the confusion of different features, and an extremely large visual vocabulary number will cause a huge waste of calculation resources. Therefore, determining k is important and would affect the final recognition accuracy, which would be discussed in the experimental section.

2.2.3. Image Coding Using BOVW Feature

In the previous section, a k-size visual vocabulary is obtained by clustering all the visual words (SURF eigenvectors). For a single image, the BOVW feature is the k-size vector by counting the occurrence frequencies of all the visual words in the vocabulary. The histograms of visual words are shown in Figure 5.
To make BOVW features suitable for following target recognition and detection, the BOVW feature [x1, x2, … xi, …, xk] needs to be normalized using L2 normalization, as
y i = x i i = 1 k x i 2 ,
where y is the corresponding normalized vector and k is the vocabulary size.
The SURF descriptors extracted from bubble plume targets and background noise image are obviously different in scales and orientations (Figure 3); therefore, BOVW features as bags of these SURF descriptors can effectively represent grayscale, shape, orientation, and texture features of bubble plume targets. In the following sections, BOVW features are used for target recognition and detection.

2.3. Bubble Plume Recognition Using Support Vector Machine

Based on the SURF descriptor differences between bubble plume target and background noise images, the recognition of bubble plume targets can be done by classification of the BOVW features. Taking the target images as positive samples and the background images as negative samples, recognition becomes binary classification. In this section, a SVM classifier is used to binarily classify the BOVW features of bubble plumb target and background noise image. The overall procedure is shown as Figure 6.

2.3.1. Support Vector Machine

As a kernelized method, selection of kernel functions seriously affects the classification accuracy of the SVM classifier. To process the extracted BOVW features (k-dimensional eigenvectors), our work uses the nonlinear SVM classifier using the quadratic polynomial kernel. For any eigenvectors xi and xj, the quadratic polynomial kernel function κ is:
κ x i , x j = ( x i T x j ) 2 .
Based on the SVM dual problem, SVM classification of the BOVW eigenvectors consists of the following steps:
  • Selection of the quadratic polynomial kernel function, where the SVM problem could be converted to the convex quadratic programming problem as follows:
max α i = 1 m α i 1 2 i = 1 m j = 1 m α i α j y i y j ( x i T x j ) 2 subject to  i = 1 m α i y i = 0 , α i 0 , i = 1 , 2 , , m .
2.
Based on the sequential minimal optimization (SMO) algorithm, the optimal solution is
α * = α 1 * , α 2 * , , α N * T
3.
Selection of αj* as one component of α*, satisfying 0 < αj* < C (C is the hyperparameter, called box constraint, to avoid overfitting). Then, we calculate
b * = y j i = 1 m α i y i ( x i T x j ) 2 .
4.
The kernel function is used to replace the inner product, and the quadratic SVM becomes
f x = i = 1 N α i * y i ( x i T x j ) 2 + b * .
The hyperparameter box constraint and feature scale also affect the SVM classification accuracy and need to be updated during SVM training.
After training, the SVM can be used for prediction of input images. The binary loss is a function of the class label and classification score that determines how well a binary learner classifies an image into the class. The prediction score P is calculated using Hinge binary learner loss function which provides a negated average binary loss in a domain as (–∞, ∞).
P = max 0 , 1 y j f j 2 ,
where, yj is a class label for the SVM binary classifier in the set (–1,1), and f(j) is the model score for input image j.

2.3.2. Recognition Procedure of Bubble Plume Targets

To recognize the bubble plume targets in the water column images, we carefully establish the positive and negative sample sets, construct the visual vocabulary based on clustering SURF points, and encode BOVW features of each sample image. Then, the SVM binary classifier is trained and used to recognize the target sample images. The flow diagram of bubble plume target recognition is shown in Figure 7. Only training image samples are used in establishing the visual vocabulary.

2.3.3. Recognition Accuracy Assessment

The prediction results of the target recognition model only contain two results, namely, the bubble plume target images or background noise images; therefore, the confusion matrix can be used to describe model accuracy for binary classification, as shown in Table 1. The accuracy assessments using the confusion matrix are listed in Table 2.

2.4. Target Detection Using BOVW Features

2.4.1. Precise Target Localization

Based on the recognition method, the flow diagram of bubble plume target detection is shown as Figure 8.
A moving search window (Figure 8b) is used to traverse the water column image (Figure 8a), and the score of each window image is predicted by the recognition model based on the BOVW feature and SVM classifier. The size of the moving search window is close to the sample image size, and the moving step was set as 1/4 of the search window size. After traverse, the bubble target is covered by many search windows (Figure 8c), and the outline rectangle of these detection windows can be extracted (as the green boundary in Figure 8d). The initial detection of the bubble plume target is obtained, and more precise detection results (Figure 8g) are processed by precise localization (Figure 8d–f).
The precise localization method is intended to achieve more accurate detection boundary by gradually shortening the boundary range of the detection frame to obtain higher model prediction scores. The steps are as follows:
  • First, gradually shrink the left boundary to the right, calculate the prediction scores of the reduced images, and select the maximum-score position as the final left boundary;
  • Second, gradually shrink the right boundary to the left, calculate the scores of the reduced images, and select the maximum-score position as the final right boundary as the red boundary in Figure 8e;
  • Third, based on the above detection boundary, gradually shrink the top boundary to the bottom, calculate the scores of the reduced images, and select the zero-score position as the final bottom boundary;
  • Finally, gradually shrink the bottom boundary to the top, calculate the scores of the reduced images, and select the zero-score position as the final top boundary as the yellow boundary in Figure 8f.
According to the shape and orientation characteristics of the bubble plume target, the prediction score gradually increases as the initial detected left/right boundary shrinks to the right/left, but gradually decreases when the target begins to get lost in the image (after the maximum score). As to the top and bottom boundaries, the prediction score does not change significantly when the image shrinks in the top-bottom direction. However, when the image shrinks exceeding the bottom/top boundary and does not contain any targets, the score rapidly decreases to the zero value.
For the multiple-target case, independent detection areas can be distinguished and obtained based on the connectivity of detection frames. In each independent detection area, accurate detection of these targets can be achieved using the aforementioned method.

2.4.2. Detection Accuracy

Intersection over Union (IoU) is a common standard for measuring the detection accuracy of image targets. IoU measures the correlation between ground truth and predicted bounding boxes. Higher correlation means higher detection accuracy.
For ground-truth box as A and predicted box as B, the IoU of A and B can be calculated as
I o U = A B A B .
Taking 0.5 as the IoU threshold, the correct detection can be defined as
Correct = true , I o U 0.5 false , I o U < 0.5 .
The correct detection rate is calculated using the correct detected target number Nc divided by the total target number N:
rate c = N c N .

3. Experiments and Results

To verify the validity and performance of the proposed method in this study, the measured multibeam water column data in the Strait of Georgia in 2012 were used for bubble target recognition and detection in the experiment, as shown in Figure 9. The multibeam sonar used was Kongsberg EM 710 with an operating frequency of 73–81 kHz, across-track beam aperture of approximately 130° and fixed beam number of 128 was recorded in the measurement. In the experiment, our method was implemented using MATLAB codes and built-in functions, as explained in Appendix A. The experiment consists of two parts:
  • BOVW feature extraction and training and validation of SVM classifier. During data preparation, the measured multibeam data were used to construct the water column images. The images containing bubble plume target and only background noises were extracted as positive and negative samples and distributed in the training and validation sample sets. Then, the BOVW features were extracted from these images and the SVM classifier was trained using these features;
  • Bubble target detection in water column images. Based on the recognition model using BOVW and SVM, the precise detection method of bubble plume targets was applied to detect all of the bubble plume targets in the water column images of the EM 710 multibeam sonar to prove the validity and generality of our detection method.

3.1. BOVW Feature Extraction and Classification

In this experiment, the recorded multibeam water column data (in *.wcd and *.all file) were decoded based on EM data format. Then, the water column images were constructed using the water column data packets.

3.1.1. Sampling from Multibeam Water Column Images

Positive samples which contain the bubble plume targets, and negative samples which contains only background noises, are manually extracted and labeled to establish the sample set, as shown in Figure 10. To avoid the problem of sample imbalance, the numbers of positive and negative samples are the same as 700. The size of each sample image was unified as 256 × 256 pixels.
The positive samples included partial or whole bubble plume targets in water column images. These positive samples were identified and labeled manually. The negative samples contained various noise background images and some other targets (like fishes) which were not discussed in this study. Meanwhile, because the water column images are fan-shaped, the edge parts of the fan-shaped images were also considered in the positive and negative samples to avoid the impact of edge SURF points on visual vocabulary construction.

3.1.2. Visual Vocabulary Construction and BOVW Feature Encoding

The visual vocabulary is the basis for encoding BOVW features. To construct the visual vocabulary, the total sample set was divided into the training and validation set in a ratio of 7:3. The visual vocabulary was constructed based on the image data of the training set to achieve the BOVW feature encoding of both the training and validation sets.
The SURF point descriptions of all the images in the training set need to be extracted according to the reasonable selection method, and then all SURF points were clustered to construct the visual vocabulary. Considering the sample image size as 256 × 256 px, several different SURF point extraction methods using 8 × 8, 16 × 16, 32 × 32 grids and SURF detector with different vocabulary word number k were selected and compared in the experiment, respectively, and the corresponding results were listed and shown in Table 3.
The training and validation accuracy of each SURF point extraction method using the quadratic SVM classifier were calculated, and the validation accuracies were sorted out as Figure 11.
Table 3 shows that different SURF point extraction methods have significant effects on the accuracy of the recognition model. In the grid extraction methods, with the increase in the grid size, the number of extracted SURF points gradually decreased, but the model accuracies may increase or decrease, indicating the particular importance of that reasonable feature scale. The detector method discovered feature points through the SURF detection algorithm. When the vocabulary word number was set as 300, the training and validation accuracies were both less than 0.5, which showed the difficulty of finding enough feature points in noise-affected sonar images by SURF algorithm.
By comparison, the grid selection method using 32 × 32 px grids was used in the experiments, and the vocabulary word number k was set as 300.

3.1.3. SVM Classifier Training and Validation

To train an SVM classifier, the kernel function needs to be specified first. Different classifiers were used to classify the BOVW features obtained in Section 3.1.2, and the validation accuracy, training time, and prediction speed of each classifier were computed and listed in Table 4. By comparison, the quadratic SVM classifier had the highest validation accuracy, reasonable training time, and prediction speed. Therefore, the quadratic function was chosen as the SVM kernel function.
After determining the kernel function, the hyperparameters, including the box constraint and feature scale, also need to be optimized. The hyperparameter box constraint (also called soft margin) controls the maximum imposed penalty of observed values that violate the margins to prevent model overfitting. If the box constraint is larger, the SVM classifier allocates fewer support vectors, resulting in longer training times.
SVM classifiers are also sensitive to the feature scales. The hyperparameter Feature Scale is used to scale the feature parameters, which means that all the elements of the BOVW features should be divided by the Feature Scale value. During the training process, the hyperparameters of the SVM classifier were optimized by finding the minimum model validation loss when constantly changing the Box Constraint and Feature Scale hyperparameter values, as shown in Figure 12.
After training, the SVM model was used to predict the image data, and the results of some water column images were listed in Table 5. As shown in the table, these images which contained bubble plume targets (Table 5a–d) had more valid SURF points and more abundant visual words. Noise background images (Table 5f–g) usually had fewer valid SURF points and the corresponding visual words were relatively simple.
For images in Table 5c,e, the noise backgrounds were quite similar, so the BOVW features of these two images were also similar to each other. However, obvious differences existed in the frequencies of special visual words (i.e., near index 190) between these two images, which proved the validity of the BOVW features in the recognition of the bubble plume targets. By calculation, the prediction accuracies of the training and validation set were 0.99 and 0.98, respectively.

3.2. Automatic Detection of Bubble Plume Target in Water Column Image

After verifying BOVW features and SVM classifier in recognition of bubble plume targets, we carried out the target detection experiment of water column images based on the proposed procedure (Figure 8).
Fixed and floating anchors are typically used in the detection of the possible targets to find the initial positions, as shown in Figure 13. The usage of fixed anchors (Figure 13a) may result in missing detection of small targets or losing parts of large targets. The usage of floating anchors can effectively avoid these problems, but may result in large initial detection boxes (Figure 13c). Therefore, further precise target localization is needed, as shown in Figure 14.
As shown in Figure 14c,e, the prediction score of the image gradually increased by constantly narrowing the detection box in the left/right directions. When the target boundary was reached, the score reached the maximum value, then started to decrease. When the detection box reached the other boundary of the target, the score would be close to zero because the image contained only background noises. Thus, the position of the maximum prediction score indicated the left and right boundary of the target detection box. Moreover, the maximum-score position in the left/right direction should correspond to the zero-score position in the right/left direction.
After determining the left and right boundaries, the top and bottom boundaries could be further obtained. Due to the orientation of the bubble plume target, the top/bottom boundaries could not be determined by the minimum value of the score curves, as shown in Figure 14b,f. Thus, the position of the top/bottom boundary should be determined by the minimum value of the bottom/top score curves. The final detection box is shown in Figure 14d, with prediction score 3.49 much larger than the initial score 0.88.
More water column images from the experimental data were processed by our detection method, and the detection results were obtained in Figure 15. As shown in Figure 15, the bubble plume targets of different sizes and backscatter strengths had been correctly detected. Taking the IoU value not less than 0.5 as the correct results, we calculated the correct detection rate of bubble plume targets in all the testing water column images as 91.7%. Therefore, the experimental results prove the validity and correctness of our detection method.

4. Discussion

In this section, our recognition and detection method of bubble plume targets in water column images was compared with other methods.

4.1. Feature and Classifier Comparison

To compare with the proposed method, various combinations of feature extraction methods and classifiers were used for the recognition of bubble plume targets in the water column images. These extracted features included gray-level cooccurrence matrix (GLCM), Tamura, local binary pattern (LBP), histogram of oriented gradient (HOG), the combination of these features, and Haar-LBP [33]. The results are shown in Table 6.
The results in Table 6 showed that the GLCM features could achieve the accuracy of 0.824 by combining steps 1 px and 5 px, the Tamura features achieved an accuracy of 0.845 when all six features were used, and the HOG features could reach 0.897 accuracy using 16 × 16 pixel calculation units. However, the recognition accuracies still need to be improved when using these feature descriptions of bubble plume targets.
The combination of LBP features and cubic SVM could reach 0.941 recognition accuracy when selecting 32 × 32 pixel computing units, indicating that LBP features could effectively express the direction characteristics of bubble plume targets. By combining LBP with Haar features, the precision ratio could be 0.994, but the accuracy was slightly improved to 0.958 due to the insufficient recognition accuracy of negative samples (noise background). This result showed that the applicability of LBP directional features on noise background images needed further improvement. The method proposed in our study could obtain a precision ratio of 0.993 and accuracy of 0.986, thereby proving the advantage of BOVW features in recognizing the bubble plume targets.

4.2. Detection Result Comparison

To compare with the previous method [33], we processed the same water column image using histogram similarity detection method and our method to obtain the following results, as shown in Figure 16.
As shown in Figure 16B, the two main problems of the detection method based on histogram similarities were the fixed width of detection boxes and the miss detection of targets in low-echo-intensity areas, resulting in the IoU value less than 0.5. By using a more reasonable detection strategy, our method avoided these two problems and obtained the IoU value as 0.85. Our detection result was very close to the manual ground-truth result. The comparison results further proved the effectiveness of our proposed detection method.

4.3. Advantage Compared with Deep Learning Methods

Deep learning methods are widely applied in the recognition and detection of various image targets, but the high accuracies of deep learning methods need to be guaranteed by a large amount of data samples and corresponding manual labels. For targets in multibeam water column images including bubble plumes, establishing a large number of sample data could be very difficult, which would limit the accuracies of deep learning recognition and detection models.
In our method, based on the characteristics of bubble plume targets, we only need to extract several hundreds of sample data to achieve high-accuracy recognition and detection results. Our method is highly suitable for bubble plume target detection in new water areas without existing sample data. For other types of water column image targets, the proposed method can also be easily retrained according to the special features of these targets.

4.4. Other Important Issues

4.4.1. Using SURF to Detect the Target

To discuss the ability of SURF detection on targets under noise interference in water column images, we directly using SURF method to detect interest points in the image (Figure 17A), and the detection results were obtained as Figure 17B. Moreover, with restrictions on feature direction (near left or right) on scale, the detection results as Figure 17C were obtained. In Figure 17C. the detected SURF points were all around the targets, which proved that conventional SURF with reasonable parameters can extract effectively feature points in water column images.

4.4.2. Ghost Targets and Targets outside the Minimum Slant Range

Besides the real bubble plume targets, ghost targets and targets outside the minimum slant range (MSR) also need to be paid attention, as shown in Figure 18.
The ghost targets are not really existing and are caused by the side lobe, presenting a similar shape to the target. However, due to the orientation of these, ghost targets are quite different; SURFs of ghost targets are different from those of real targets. Therefore, our method using BOVW features achieved good recognition rate on ghost targets.
Our method mainly focuses on the targets inside MSR, because the backscatter samples outside MSR are heavily affected by side lobes. The detection of targets outside MSR is quite a challenge, which would be studied in our future works.

4.4.3. Application on Other Multibeam Water Column Data

The multibeam water column images measured by different sonar in different water conditions could be quite different. The multibeam water column data (Cruise ID: EX1402L3) [47] measured in Gulf of Mexico, 2014, were chosen for this discussion, as shown in Figure 19A. The multibeam sonar was Kongsberg EM 302 with an operating frequency of 26.5–31.7 kHz, across-track beam aperture of 130° and fixed beam number of 288 was used. Due to different frequencies and different water conditions, the extracted features of the water column image (Figure 19B) are quite different from those of the data in our experimental section. The trained model using multibeam data of EM 710 sonar cannot be directly used for data measured using EM 302. To recognize and detect targets in this measurement, the BOVW features and SVM model need to be re-trained. Moreover, how different sonars and water conditions affect the accuracy of our method would be our future research direction.

5. Conclusions

Based on the characteristics of bubble plume targets in multibeam water column images, our proposed method extracts the BOVW features of the targets and uses the SVM classifier for target recognition. On this basis, we further propose a precise target detection method of bubble plume targets in water column images. Through experiments on the measured data, the validity and correctness of each step in the target recognition method were verified, including sample set establishment, parameter selection of BOVW features, and SVM classifier optimization. The validation accuracy of our recognition method was 98.6%. In the target detection experiment, based on the high-accuracy recognition model, the accurate detection rate of bubble target targets in the water column images was 91.7% and most of the detection results were close to the manual ground truths. Compared with various combinations of features and classifiers, the advantages of BOVW and SVM in target recognition was proved. Compared with the previous detection method, the proposed method solved the existing problems and effectively improved the target detection accuracy. The proposed recognition and detection method can also be applied to more types of water column image targets, and also has significance to the exploration and research of underwater resources.

Author Contributions

Conceptualization: J.M. and J.Y. Methodology: J.M. Software: J.Y. Validation: J.M., J.Y. and J.Z. Formal analysis: J.M. and J.Y. Investigation: J.M. and J.Y. Resources: J.Z. Data curation: J.M. Writing—original draft preparation: J.M. Writing—review and editing: J.M., J.Y. and J.Z. Visualization: J.M. and J.Y. Supervision: J.M. and J.Y. Project administration: J.M. Funding acquisition: J.M. and J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

The National Natural Science Foundation of China (grant number 42104036, 41906168), Natural Science Foundation of Anhui Province (grant number 1908085QD161), the University Natural Science Research Key Project of Anhui Province (grant number KJ2019A0024), and Doctoral Research Foundation of Anhui Jianzhu University (grant number 2018QD45) funded this research.

Data Availability Statement

Not applicable.

Acknowledgments

The Guangzhou Marine Geological Survey Bureau provided the experimental data in this study. The authors appreciate their support.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The proposed method was implemented using MATLAB codes and built-in functions. The raw multibeam data were decoded using our written MATLAB code based on EM data format document. Then multibeam water column images were constructed using the decoded data. The samples were extracted and labeled manually, and the target and noise samples were stored in different folders. The BOVW feature extraction were realized using the “bagOfFeatures” function from MATLAB computer vision toolbox. The function parameters need to be carefully selected as discussed in Table 3. The SVM model training were realized using the “templateSVM” function from MATLAB statistics and machine learning toolbox with reasonable parameters. The model prediction scores were obtained using the “predict” function from MATLAB statistics and machine learning toolbox. The detection methods were realized by our MATLAB codes following the steps in Section 2.4.1 and Figure 14. The MATLAB academic license was offered by Anhui University.

References

  1. Colbo, K.; Ross, T.; Brown, C.; Weber, T. A review of oceanographic applications of water column data from multibeam echosounders. Estuar. Coast. Shelf Sci. 2014, 145, 41–56. [Google Scholar] [CrossRef]
  2. Melvin, G.D.; Cochrane, N.A. Multibeam acoustic detection of fish and water column targets at high-flow sites. Estuaries Coasts 2015, 38, 227–240. [Google Scholar] [CrossRef]
  3. Buelens, B.; Pauly, T.; Williams, R.; Sale, A. Kernel methods for the detection and classification of fish schools in single-beam and multibeam acoustic data. ICES J. Mar. Sci. 2009, 66, 1130–1135. [Google Scholar] [CrossRef] [Green Version]
  4. Cox, M.J.; Warren, J.D.; Demer, D.A.; Cutter, G.R.; Brierley, A.S. Three-dimensional observations of swarms of Antarctic krill (Euphausia superba) made using a multi-beam echosounder. Deep Sea Res. Part II Top. Stud. Oceanogr. 2010, 57, 508–518. [Google Scholar] [CrossRef]
  5. Di Maida, G.; Tomasello, A.; Luzzu, F.; Scannavino, A.; Pirrotta, M.; Orestano, C.; Calvo, S. Discriminating between Posidonia oceanica meadows and sand substratum using multibeam sonar. ICES J. Mar. Sci. 2010, 68, 12–19. [Google Scholar] [CrossRef] [Green Version]
  6. De Falco, G.; Tonielli, R.; Di Martino, G.; Innangi, S.; Simeone, S.; Michael Parnum, I. Relationships between multibeam backscatter, sediment grain size and Posidonia oceanica seagrass distribution. Cont. Shelf Res. 2010, 30, 1941–1950. [Google Scholar] [CrossRef]
  7. Moum, J.N.; Farmer, D.M.; Smyth, W.D.; Armi, L.; Vagle, S. Structure and Generation of Turbulence at Interfaces Strained by Internal Solitary Waves Propagating Shoreward over the Continental Shelf. J. Phys. Oceanogr. 2003, 33, 2093–2112. [Google Scholar] [CrossRef]
  8. Leong, D.; Ross, T.; Lavery, A. Anisotropy in high-frequency broadband acoustic backscattering in the presence of turbulent microstructure and zooplankton. J. Acoust. Soc. Am. 2012, 132, 670–679. [Google Scholar] [CrossRef] [Green Version]
  9. Hughes Clarke, J.E.; Lamplugh, M.; Czotter, K. Multibeam water column imaging: Improved wreck least-depth determination. In Proceedings of the Canadian Hydrographic Conference, Halifax, NS, Canada, 6–9 June 2006. [Google Scholar]
  10. Wyllie, K.; Weber, T.; Armstrong, A. Using Multibeam Echosounders for Hydrographic Surveying in the Water Column: Estimating Wreck Least Depths. In Proceedings of the US Hydrographic Conference, National Harbor, MD, USA, 16–19 March 2015. [Google Scholar]
  11. Weber, T.C. Observations of clustering inside oceanic bubble clouds and the effect on short-range acoustic propagation. J. Acoust. Soc. Am. 2008, 124, 2783–2792. [Google Scholar] [CrossRef] [Green Version]
  12. Schneider von Deimling, J.; Brockhoff, J.; Greinert, J. Flare imaging with multibeam systems: Data processing for bubble detection at seeps. Geochem. Geophys. Geosyst. 2007, 8, Q06004. [Google Scholar] [CrossRef] [Green Version]
  13. Schneider von Deimling, J.; Papenberg, C. Detection of gas bubble leakage via correlation of water column multibeam images. Ocean Sci. 2012, 8, 175–181. [Google Scholar] [CrossRef] [Green Version]
  14. Weber, T.C.; Mayer, L.; Jerram, K.; Beaudoin, J.; Rzhanov, Y.; Lovalvo, D. Acoustic estimates of methane gas flux from the seabed in a 6000 km2 region in the Northern Gulf of Mexico. Geochem. Geophys. Geosyst. 2014, 15, 1911–1925. [Google Scholar] [CrossRef] [Green Version]
  15. Dupre, S.; Scalabrin, C.; Grall, C.; Augustin, J.M.; Henry, P.; Şengör, A.; Görür, N.; Çağatay, M.N.; Géli, L. Tectonic and sedimentary controls on widespread gas emissions in the Sea of Marmara: Results from systematic, shipborne multibeam echo sounder water column imaging. J. Geophys. Res. Solid Earth 2015, 120, 2891–2912. [Google Scholar] [CrossRef] [Green Version]
  16. Innangi, S.; Bonanno, A.; Tonielli, R.; Gerlotto, F.; Innangi, M.; Mazzola, S. High resolution 3-D shapes of fish schools: A new method to use the water column backscatter from hydrographic MultiBeam Echo Sounders. Appl. Acoust. 2016, 111, 148–160. [Google Scholar] [CrossRef]
  17. Hughes Clarke, J.E. Applications of multibeam water column imaging for hydrographic survey. Hydrogr. J. 2006, 4, 3–14. [Google Scholar]
  18. Hughes Clarke, J.E.; Brucker, S.; Czotter, K. Improved Definition of Wreck Superstructure using Multibeam Water Column. J. Can. Hydrogr. Assoc. 2006, 5, 1–2. [Google Scholar]
  19. Marques, C.R.; Hughes Clarke, J.E. Automatic mid-water target tracking using multibeam water column. In Proceedings of the CHC 2012, The Arctic, Old Challenges New, Niagara Falls, ON, Canada, 15–17 May 2012. [Google Scholar]
  20. Gardner, J.V.; Malik, M.; Walker, S. Plume 1400 meters high discovered at the seafloor off the northern California margin. Eos Trans. Am. Geophys. Union 2009, 90, 275. [Google Scholar] [CrossRef]
  21. Weber, T.C.; Mayer, L.A.; Beaudoin, J.; Jerram, K.W.; Malik, M.A.; Shedd, B.; Rice, G. Mapping Gas Seeps with the Deepwater Multibeam Echosounder on Okeanos Explorer. Oceanography 2012, 25, 55–56. [Google Scholar] [CrossRef] [Green Version]
  22. Sahling, H.; Römer, M.; Pape, T.; Bergès, B.; dos Santos Fereirra, C.; Boelmann, J.; Geprägs, P.; Tomczyk, M.; Nowald, N.; Dimmler, W. Gas emissions at the continental margin west of Svalbard: Mapping, sampling, and quantification. Biogeosciences 2014, 11, 6029–6046. [Google Scholar] [CrossRef] [Green Version]
  23. Skarke, A.; Ruppel, C.; Kodis, M.; Brothers, D.; Lobecker, E. Widespread methane leakage from the sea floor on the northern US Atlantic margin. Nat. Geosci. 2014, 7, 657–661. [Google Scholar] [CrossRef]
  24. Nakamura, K.; Kawagucci, S.; Kitada, K.; Kumagai, H.; Takai, K.; Okino, K. Water column imaging with multibeam echo-sounding in the mid-Okinawa Trough: Implications for distribution of deep-sea hydrothermal vent sites and the cause of acoustic water column anomaly. Geochem. J. 2015, 49, 579–596. [Google Scholar] [CrossRef] [Green Version]
  25. Philip, B.T.; Denny, A.R.; Solomon, E.A.; Kelley, D.S. Time-series measurements of bubble plume variability and water column methane distribution above Southern Hydrate Ridge, Oregon. Geochem. Geophys. Geosyst. 2016, 17, 1182–1196. [Google Scholar] [CrossRef]
  26. Nikolovska, A.; Sahling, H.; Bohrmann, G. Hydroacoustic methodology for detection, localization, and quantification of gas bubbles rising from the seafloor at gas seeps from the eastern Black Sea. Geochem. Geophys. Geosyst. 2008, 9, Q10010. [Google Scholar] [CrossRef]
  27. Dobeck, G.; Hyland, J.; Smedley, L.D. Automated detection and classification of sea mines in sonar imagery. In Proceedings of the AeroSense ’97, Orlando, FL, USA, 22 July 1997. [Google Scholar]
  28. Tang, X.; Stewart, W.K. Optical and Sonar Image Classification: Wavelet Packet Transform vs Fourier Transform. Comput. Vis. Image Underst. 2000, 79, 25–46. [Google Scholar] [CrossRef] [Green Version]
  29. Reed, S.; Ruiz, I.T.; Capus, C.; Petillot, Y. The fusion of large scale classified side-scan sonar image mosaics. IEEE Trans. Image Process. 2006, 15, 2049–2060. [Google Scholar] [CrossRef]
  30. Rhinelander, J. Feature extraction and target classification of side-scan sonar images. In Proceedings of the 2016 IEEE Symposium Series on Computational Intelligence (SSCI), Athens, Greece, 6–9 December 2016; pp. 1–6. [Google Scholar]
  31. Song, Y.; He, B.; Zhao, Y.; Li, G.; Sha, Q.; Shen, Y.; Yan, T.; Nian, R.; Lendasse, A. Segmentation of Sidescan Sonar Imagery Using Markov Random Fields and Extreme Learning Machine. IEEE J. Ocean. Eng. 2019, 44, 502–513. [Google Scholar] [CrossRef]
  32. Wang, X.; Zhao, J.; Zhu, B.; Jiang, T.; Qin, T. A Side Scan Sonar Image Target Detection Algorithm Based on a Neutrosophic Set and Diffusion Maps. Remote Sens. 2018, 10, 295. [Google Scholar] [CrossRef] [Green Version]
  33. Zhao, J.; Mai, D.; Zhang, H.; Wang, S. Automatic Detection and Segmentation on Gas Plumes from Multibeam Water Column Images. Remote Sens. 2020, 12, 3085. [Google Scholar] [CrossRef]
  34. Williams, D.P.; Dugelay, S. Multi-view SAS image classification using deep learning. In Proceedings of the OCEANS 2016 MTS/IEEE Monterey, Monterey, CA, USA, 19–23 September 2016; pp. 1–9. [Google Scholar]
  35. Zhu, P.; Isaacs, J.; Fu, B.; Ferrari, S. Deep learning feature extraction for target recognition and classification in underwater sonar images. In Proceedings of the 2017 IEEE 56th Annual Conference on Decision and Control (CDC), Melbourne, VIC, Australia, 12–15 December 2017; pp. 2724–2731. [Google Scholar]
  36. Lee, S. Deep learning of submerged body images from 2D sonar sensor based on convolutional neural network. In Proceedings of the 2017 IEEE Underwater Technology (UT), Busan, Korea, 21–24 February 2017; pp. 1–3. [Google Scholar]
  37. Ribeiro, P.O.C.S.; dos Santos, M.M.; Drews, P.L.J.; Botelho, S.S.C. Forward Looking Sonar Scene Matching Using Deep Learning. In Proceedings of the 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA), Cancun, Mexico, 18–21 December 2017; pp. 574–579. [Google Scholar]
  38. Valdenegro-Toro, M. Improving sonar image patch matching via deep learning. In Proceedings of the 2017 European Conference on Mobile Robots (ECMR), Paris, France, 6–8 September 2017; pp. 1–6. [Google Scholar]
  39. Li, C.; Ye, X.; Cao, D.; Hou, J.; Yang, H. Zero shot objects classification method of side scan sonar image based on synthesis of pseudo samples. Appl. Acoust. 2021, 173, 107691. [Google Scholar] [CrossRef]
  40. Liu, D.; Wang, Y.; Ji, Y.; Tsuchiya, H.; Yamashita, A.; Asama, H. CycleGAN-based realistic image dataset generation for forward-looking sonar. Adv. Robot. 2021, 35, 242–254. [Google Scholar] [CrossRef]
  41. Huang, C.; Zhao, J.; Yu, Y.; Zhang, H. Comprehensive Sample Augmentation by Fully Considering SSS Imaging Mechanism and Environment for Shipwreck Detection Under Zero Real Samples. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  42. Specht, M.; Stateczny, A.; Specht, C.; Widźgowski, S.; Lewicka, O.; Wiśniewska, M. Concept of an Innovative Autonomous Unmanned System for Bathymetric Monitoring of Shallow Waterbodies (INNOBAT System). Energies 2021, 14, 5370. [Google Scholar] [CrossRef]
  43. Lewicka, O.; Specht, M.; Stateczny, A.; Specht, C.; Brčić, D.; Jugović, A.; Widźgowski, S.; Wiśniewska, M. Analysis of GNSS, Hydroacoustic and Optoelectronic Data Integration Methods Used in Hydrography. Sensors 2021, 21, 7831. [Google Scholar] [CrossRef] [PubMed]
  44. Kesorn, K.; Poslad, S. An Enhanced Bag-of-Visual Word Vector Space Model to Represent Visual Content in Athletics Images. IEEE Trans. Multimed. 2012, 14, 211–222. [Google Scholar] [CrossRef]
  45. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-Up Robust Features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  46. Zhao, J.; Meng, J.; Zhang, H.; Yan, J. A New Method for Acquisition of High-Resolution Seabed Topography by Matching Seabed Classification Images. Remote Sens. 2017, 9, 1214. [Google Scholar] [CrossRef] [Green Version]
  47. NOAA. NOAA Office of Ocean Exploration and Research: Water Column Sonar Data Collection (EX1402L3, EM302); National Centers for Environmental Information, NOAA: Asheville, NC, USA, 2014. [Google Scholar] [CrossRef]
Figure 1. Bubble plume detection in multibeam water column image. (A) is the principal diagram of target detection in multibeam water column images, where Si is the sampling point at equal time intervals and θ is the beam incident angle; (B) is the measured multibeam water column image containing the bubble plume target.
Figure 1. Bubble plume detection in multibeam water column image. (A) is the principal diagram of target detection in multibeam water column images, where Si is the sampling point at equal time intervals and θ is the beam incident angle; (B) is the measured multibeam water column image containing the bubble plume target.
Remotesensing 14 03296 g001
Figure 2. BOVW feature extraction from water column images based on SURF.
Figure 2. BOVW feature extraction from water column images based on SURF.
Remotesensing 14 03296 g002
Figure 3. SURF differences between bubble plume target and background image in multibeam water column image (Hessian threshold was set as 250). The radiuses of these circles indicate the scales of these detected points, and the lines in these circles mean the main directions of these features.
Figure 3. SURF differences between bubble plume target and background image in multibeam water column image (Hessian threshold was set as 250). The radiuses of these circles indicate the scales of these detected points, and the lines in these circles mean the main directions of these features.
Remotesensing 14 03296 g003
Figure 4. Change of k (as 300) cluster centroids and corresponding assignments after several iterations. (A) is the initial clustering result using k-mean++ algorithm, (B) shows the training processes of k-means clustering, and (C) is the final clustering result.
Figure 4. Change of k (as 300) cluster centroids and corresponding assignments after several iterations. (A) is the initial clustering result using k-mean++ algorithm, (B) shows the training processes of k-means clustering, and (C) is the final clustering result.
Remotesensing 14 03296 g004
Figure 5. BOVW feature encoding of positive ((a): bubble plume target) and negative ((d): background noise) samples. (b,e) show occurrence frequencies of visual words extracted from (a,b), respectively. (c,f) are normalized from (b,e), respectively.
Figure 5. BOVW feature encoding of positive ((a): bubble plume target) and negative ((d): background noise) samples. (b,e) show occurrence frequencies of visual words extracted from (a,b), respectively. (c,f) are normalized from (b,e), respectively.
Remotesensing 14 03296 g005
Figure 6. Procedure of SVM binary classification on BOVW features.
Figure 6. Procedure of SVM binary classification on BOVW features.
Remotesensing 14 03296 g006
Figure 7. Flow diagram of bubble plume target recognition in water column sampling images based on BOVW features. During the training procedure, the parameters in BOVW feature extraction need to be updated by continually changing the interest point extraction method and the number of the visual vocabulary until the recognition accuracy reaches the maximum value.
Figure 7. Flow diagram of bubble plume target recognition in water column sampling images based on BOVW features. During the training procedure, the parameters in BOVW feature extraction need to be updated by continually changing the interest point extraction method and the number of the visual vocabulary until the recognition accuracy reaches the maximum value.
Remotesensing 14 03296 g007
Figure 8. Target detection using SVM classifier and BOVW features. The processing steps are shown in (ag).
Figure 8. Target detection using SVM classifier and BOVW features. The processing steps are shown in (ag).
Remotesensing 14 03296 g008
Figure 9. Experimental regions (Strait of Georgia) of multibeam water column data.
Figure 9. Experimental regions (Strait of Georgia) of multibeam water column data.
Remotesensing 14 03296 g009
Figure 10. Positive and negative sampling from raw water column images.
Figure 10. Positive and negative sampling from raw water column images.
Remotesensing 14 03296 g010
Figure 11. Validation accuracies of recognition model using BOVW feathers with different SURF point selection method.
Figure 11. Validation accuracies of recognition model using BOVW feathers with different SURF point selection method.
Remotesensing 14 03296 g011
Figure 12. Optimization of hype-parameter Feature Scale and Box Constraint by finding the minimum SVM function value.
Figure 12. Optimization of hype-parameter Feature Scale and Box Constraint by finding the minimum SVM function value.
Remotesensing 14 03296 g012
Figure 13. Target detection results using fix anchor boxes (a) and moving anchor boxes (b). (c) is the maximum boundary of (b).
Figure 13. Target detection results using fix anchor boxes (a) and moving anchor boxes (b). (c) is the maximum boundary of (b).
Remotesensing 14 03296 g013
Figure 14. Precise target localization of bubble plume target: (a) prediction score of three detection boxes; (b,c,e,f) prediction score curves in left, right, top, and bottom shrinking direction; (d) final detection box result.
Figure 14. Precise target localization of bubble plume target: (a) prediction score of three detection boxes; (b,c,e,f) prediction score curves in left, right, top, and bottom shrinking direction; (d) final detection box result.
Remotesensing 14 03296 g014
Figure 15. Comparison of our predicted and ground-truth detection results. (ae) are five water column images which contain bubble plume targets. The yellow and green detection boxes show predicted and ground-truth results.
Figure 15. Comparison of our predicted and ground-truth detection results. (ae) are five water column images which contain bubble plume targets. The yellow and green detection boxes show predicted and ground-truth results.
Remotesensing 14 03296 g015
Figure 16. Comparison between our detection method and histogram-similarity detection method. The red, yellow, and green rectangles are the detection boxes of the histogram similarity method, our method, and ground truth, respectively. IoU values were calculated using Equation (13). (A): Raw Water Column Image; (B): Detection Result using Histogram Similarity; (C): Our Detection Result; (D): Manual Ground Truth.
Figure 16. Comparison between our detection method and histogram-similarity detection method. The red, yellow, and green rectangles are the detection boxes of the histogram similarity method, our method, and ground truth, respectively. IoU values were calculated using Equation (13). (A): Raw Water Column Image; (B): Detection Result using Histogram Similarity; (C): Our Detection Result; (D): Manual Ground Truth.
Remotesensing 14 03296 g016
Figure 17. Using SURF directly to detect the target. (A) is the original image, (B) is the SURF detection result of (A), and (C) is the SURF detection result of (A) with restrictions on feature scale and direction.
Figure 17. Using SURF directly to detect the target. (A) is the original image, (B) is the SURF detection result of (A), and (C) is the SURF detection result of (A) with restrictions on feature scale and direction.
Remotesensing 14 03296 g017
Figure 18. Different types of targets in the water column images.
Figure 18. Different types of targets in the water column images.
Remotesensing 14 03296 g018
Figure 19. Multibeam water column data in Gulf of Mexico. (A) is the track lines, and (B) shows a multibeam images of this measurement.
Figure 19. Multibeam water column data in Gulf of Mexico. (A) is the track lines, and (B) shows a multibeam images of this measurement.
Remotesensing 14 03296 g019
Table 1. Confusion matrix of binary classification.
Table 1. Confusion matrix of binary classification.
TruthPredicting Results
TrueFalse
PositiveTrue Positive (TP)False Positive (FP)
NegativeTrue Negative (TN)False Negative (FN)
TP and FP indicate that positive samples are identified as positive and negative samples, respectively, while TN and FN indicate that negative samples are identified as negative and positive samples, respectively.
Table 2. Accuracy assessment of SVM binary classification.
Table 2. Accuracy assessment of SVM binary classification.
Accuracy AssessmentComputational Formula
Accuracy A = T P + T N T P + F P + T N + F N (10)
Precision ratio P = T P T P + F P (11)
Recall ratio R = T P T P + F N (12)
Harmonic mean F 1 = 2 T P 2 T P + F P + F N (13)
Table 3. Parameter selection in constructing visual vocabulary.
Table 3. Parameter selection in constructing visual vocabulary.
SURF Point
Extraction Method
Word Number of Visual
Vocabulary
SURF Point NumberStrongest PointNumberPoint Number for Each
Category
Training
Accuracy
Validation
Accuracy
Grid [8 × 8]2004,014,0803,211,0601,605,5300.620.60
3000.840.83
4000.860.83
5000.710.69
Grid [16 × 16]2001,003,584802,816401,4080.740.73
3000.810.76
4000.740.72
5000.790.77
Grid [32 × 32]100250,880200,704100,3521.000.90
2001.000.93
3001.000.95
4001.000.94
5001.000.94
Detector200419,928331,000165,5000.860.79
3000.470.46
4000.960.89
5000.920.80
The training and validation accuracy were calculated using the same SVM classifier.
Table 4. Comparison of different classifiers using BOVW features.
Table 4. Comparison of different classifiers using BOVW features.
ClassifierValidation
Accuracy (%)
Prediction Speed (Observation/s)Training Time (s)
Medium Tree90.37508.15
Linear Discriminant92.585010.36
Logistic Regression72.155020.63
Linear SVM98.386010.16
Quadratic SVM98.681010.01
Cubic SVM98.480010.56
Medium Gaussian SVM97.9120011.07
Medium KNN95.295012.61
Bagged Trees Ensemble96.960019.65
Table 5. Prediction results of water column images set using BOVW features.
Table 5. Prediction results of water column images set using BOVW features.
Test ImageBOVW FeaturePrediction ScorePrediction ClassManual Label
(a) Remotesensing 14 03296 i001 Remotesensing 14 03296 i002(0, −2.8254)Bubble plumeBubble plume
(b) Remotesensing 14 03296 i003 Remotesensing 14 03296 i004(0, −1.4815)Bubble plumeBubble plume
(c) Remotesensing 14 03296 i005 Remotesensing 14 03296 i006(0, −2.6271)Bubble plumeBubble plume
(d) Remotesensing 14 03296 i007 Remotesensing 14 03296 i008(0, −1.2373)Bubble plumeBubble plume
(e) Remotesensing 14 03296 i009 Remotesensing 14 03296 i010(−1.4502, 0)NoiseNoise
(f) Remotesensing 14 03296 i011 Remotesensing 14 03296 i012(−1.3467, 0)NoiseNoise
(g) Remotesensing 14 03296 i013 Remotesensing 14 03296 i014(−1.2134, −0)NoiseNoise
(h) Remotesensing 14 03296 i015 Remotesensing 14 03296 i016(−0.9665, −0.0335)NoiseNoise
OverallConfusion Matrix 0.95 0.05 0.00 1.00 Validation Accuracy0.98
Table 6. Comparison validation accuracies of various feature extraction and classifier.
Table 6. Comparison validation accuracies of various feature extraction and classifier.
Feature Extraction MethodFeature NumberClassifierAccuracy (%)Precision Ratio (%)Recall Ratio (%)F1 (%)
GLCM (d = 1)12Linear SVM77.5975.3282.0778.55
GLCM (d = 5)12Medium Tree71.0367.8480.0073.42
GLCM (d = 10)12Cosine KNN70.6967.6579.3173.02
GLCM (d = 1&5)24Logistic Regression82.4181.7683.4582.59
Tamura3Medium Tree79.3181.4875.8678.57
Tamura6Fine Tree84.4884.2584.8384.54
LBP (64 × 64)59Medium Gaussian SVM90.0094.6284.8289.45
LBP (64 × 64)10Medium Gaussian SVM79.3178.1581.3879.73
LBP (32 × 32)236Cubic SVM94.1494.4493.7994.12
HOG (32 × 32)36Weighted KNN82.7682.3183.4582.88
HOG (16 × 16)324Quadratic SVM89.6689.6689.6589.65
HOG (8 × 8)1764Linear SVM76.5574.2181.3877.63
GLCM + Tamura30Quadratic SVM91.7290.6093.1091.84
GLCM + LBP83Quadratic SVM91.3890.5492.4191.47
GLCM + HOG60Quadratic SVM90.3487.7493.7990.67
Tamura + LBP65Quadratic SVM93.1092.5293.7993.15
Tamura + HOG42Quadratic SVM90.3490.9189.6690.28
LBP + HOG95Quadratic SVM90.3489.8091.0390.41
GLCM + Tamura + LBP + HOG125Quadratic SVM95.1795.8094.4895.14
Haar-LBP [33]-AdaBoost95.8099.3582.7090.26
BOVW300Quadratic SVM98.6299.3097.9398.61
The parameter d of GLCM feature means the calculation step in pixels. Tamura features contain coarseness, contrast, directionality, line likeness, regularity and roughness. 8 × 8, 16 × 16, 32 × 32 and 64 × 64 in LBP and HOG features means the selected unit sizes in pixel.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Meng, J.; Yan, J.; Zhao, J. Bubble Plume Target Detection Method of Multibeam Water Column Images Based on Bags of Visual Word Features. Remote Sens. 2022, 14, 3296. https://doi.org/10.3390/rs14143296

AMA Style

Meng J, Yan J, Zhao J. Bubble Plume Target Detection Method of Multibeam Water Column Images Based on Bags of Visual Word Features. Remote Sensing. 2022; 14(14):3296. https://doi.org/10.3390/rs14143296

Chicago/Turabian Style

Meng, Junxia, Jun Yan, and Jianhu Zhao. 2022. "Bubble Plume Target Detection Method of Multibeam Water Column Images Based on Bags of Visual Word Features" Remote Sensing 14, no. 14: 3296. https://doi.org/10.3390/rs14143296

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop