Next Article in Journal
A Scheme for the Long-Term Monitoring of Impervious−Relevant Land Disturbances Using High Frequency Landsat Archives and the Google Earth Engine
Previous Article in Journal
p-ICP Coastline Inflection Method for Geolocation Error Estimation in FY-3 MWRI Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Acoustic Target Recognition: A Combination of Multi-Dimensional Fusion Features and Modified Deep Neural Network

1
College of Computer Science and Technology, Harbin Engineering University, Harbin 150001, China
2
College of Computer Science and Technology, Harbin Institute of Technology, Harbin 518000, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(16), 1888; https://doi.org/10.3390/rs11161888
Submission received: 1 July 2019 / Revised: 4 August 2019 / Accepted: 8 August 2019 / Published: 13 August 2019
(This article belongs to the Section Ocean Remote Sensing)

Abstract

:
A method with a combination of multi-dimensional fusion features and a modified deep neural network (MFF-MDNN) is proposed to recognize underwater acoustic targets in this paper. Specifically, due to the complex and changeable underwater environment, it is difficult to describe underwater acoustic signals with a single feature. The Gammatone frequency cepstral coefficient (GFCC) and modified empirical mode decomposition (MEMD) are developed to extract multi-dimensional features in this paper. Moreover, to ensure the same time dimension, a dimension reduction method is proposed to obtain multi-dimensional fusion features in the original underwater acoustic signals. Then, to reduce redundant features and further improve recognition accuracy, the Gaussian mixture model (GMM) is used to modify the structure of a deep neural network (DNN). Finally, the proposed underwater acoustic target recognition method can obtain an accuracy of 94.3% under a maximum of 800 iterations when the dataset has underwater background noise with weak targets. Compared with other methods, the recognition results demonstrate that the proposed method has higher accuracy and strong adaptability.

Graphical Abstract

1. Introduction

With the development of sonar technology, underwater acoustic target recognition has become one of the major functions of sonar systems. It is extensively used for marine biological survey, marine exploration and other scientific activities. However, existing underwater acoustic target recognition still depends on the decisions of well-trained solar persons. It has been difficult to implement continuous monitoring and recognition [1,2]. Therefore, underwater acoustic target recognition with a high recognition accuracy and efficiency attracts extensive attention both in military and civil fields [3,4,5,6,7,8].
Currently, underwater acoustic target recognition contains feature extraction and recognition. Feature extraction is a process which extracts various features from underwater acoustic signals [9,10]. The Mel filter bank has been widely used in feature extraction. It was designed for imitating the band pass filter bank features of the human ear, and it has been the foundation of most speech processing, such as underwater acoustic target recognition, speaker recognition [11,12,13,14,15]. The cepstral or energy features are obtained from the Mel filter bank, which is known as the Mel frequency cepstral coefficient (MFCC). It has been considered the baseline feature for most applications. Lim, T et al. firstly introduced MFCC to recognize underwater acoustic targets. Experimental results demonstrate that the method is very promising for underwater acoustic target recognition [11]. Later Lim and Bae et al. proposed a method which combined the MFCC with neural network. This method also achieved a good result on the experimental dataset [12]. However, the MFCC cannot describe the optimal features for all underwater acoustic targets [16,17]. To solve the disadvantages, the Gammatone frequency cepstral coefficient (GFCC) was introduced into underwater acoustic target recognition [18]. The GFCC based on the Gammatone filter bank can better extract voiceprint features of the underwater acoustic target [18,19]. In addition, feature parameters of the GFCC can improve the recognition rate and robustness under a noisy background [17,18,19,20,21]. In 2014, Zeng and Wang et al. applied GFCC and the Hilbert–Huang transform (HHT) to recognize underwater acoustic targets, which achieved better results than the MFCC. Since then, the GFCC has gradually entered the field of research [22,23,24,25,26,27]. Although the recognition results have been modified, it cannot accurately describe original underwater acoustic signals by only extracting one type of feature. In 2018, Sharma et al. proposed a new method of combining the MFCC and modified empirical mode decomposition (MEMD) to describe mixed multi-dimensional features of original acoustic signals [16]. In feature extraction, MEMD has been mapped to Hilbert space through HHT, the instantaneous energy (IE) and instantaneous frequency (IF) of the intrinsic mode function (IMF) were extracted, and good results have been achieved through the combination with the MFCC [16,21,22,23,24,28,29,30,31]. Inspired by these, a multi-dimensional fusion features method is proposed in this paper. The method combines the advantages of the GFCC with MEMD, and the original underwater acoustic signals can be described from multiple angles.
Recognition is a process of identifying an unknown target signal, which trains the observed target signals according to a certain method. The Gaussian mixture model (GMM) is widely used in the field of recognition and has achieved good results. In 2011, Kotari et al. proposed a method to recognize underwater mines, which is based on the GMM and achieved a good result [32]. On this basis, there are various methods, which have been proposed to improve GMM in underwater acoustic target recognition. Recently, Wang et al. proposed a novel method that systematically combines the deep Boltzmann machine (DBM) with the Dirichlet process based Gaussian mixture model (DP-GMM) to bypass the problem of distribution mismatch. The method can improve the effectiveness and the robustness of the GMM [33,34]. However, compared with the deep learning method, the GMM recognition effect is unsatisfactory in the field of multi-dimensional features [9,10]. GMM is a shallow recognition model, which cannot extract the deep information features of acoustic signals and lacks the ability to obtain the essential features of a large dataset. However, deep neural network (DNN) is a deep recognition model, which can extract abstract and invariant features from a large dataset. Recently, the DNN has gradually attracted increased attention in underwater acoustic target recognition. In 2018, Ibrahim, Ali k et al. recognized the grouper species by their sounds using DNN, and the results are significantly better than the other previously reported methods [35]. Compared with the machine learning model, it can bring great improvement in recognition accuracy. Moreover, the DNN has a strong robustness to recognize underwater acoustic signals under noise conditions. Recognition results have further demonstrated that the DNN has a better learning capability than the GMM [36,37]. However, when the DNN directly recognizes features, it can generate many redundant features which greatly wastes computing resources. In this paper, to overcome these shortcomings, a modified DNN is proposed to accurately recognize underwater acoustic targets. The GMM is used to extract the statistical parameters of the feature matrix, which modify the structure of the DNN.
This paper proposes an MFF-MDNN method which combines multi-dimensional fusion features with a modified DNN to recognize underwater acoustic targets. For solving the problem whereby a single feature cannot better describe underwater acoustic signals, a multi-dimensional fusion features method is given in this paper, which uses the GFCC and MEMD to extract multi-dimensional features. Then, a dimension reduction method is proposed to ensure the same time dimension. It can obtain multi-dimensional fusion features in the original underwater acoustic signals. In addition, the problem of many redundant features being generated in the processing recognition in the DNN was solved by a method utilizing a modified DNN in this paper. The GMM is used to modify the structure of the DNN to improve the accuracy of recognition. The proposed underwater acoustic target recognition method was applied on a dataset, which was divided into six categories, including four types of ships, underwater mammals, and underwater background noise with weak targets. The recognition results show that the method has good effectiveness and adaptability. The proposed method has important theoretical and practical value.
The rest of the paper is organized as follows: Section 2 discusses the multi-dimensional fusion features method. Section 3 describes a modified deep neural network. Section 4 presents the experimental results and analysis. Finally, Section 5 summarizes and concludes this work.

2. Multi-Dimensional Fusion Features Method

2.1. Gammatone Frequency Cepstral Coefficient

The GFCC is one of mainstream feature extraction algorithms in recognition [18,19], Figure 1 shows the specific GFCC process. To obtain the power spectrum, the discrete Fourier transform (DFT) is employed to transform the original signals from the time domain to the frequency domain. The frequency is distorted from the Hertz scale to the Gammatone scale using the Gammatone filter bank. The Gammatone power spectrum is taken as a logarithm. Finally, the GFCC features are extracted by discrete cosine transforms (DCT).
The DFT is obtained by the Fourier transformation as described by the following equation:
x ( k ) = t = 0 M S 1 x ( t ) e j 2 π t k / M S , 0 k M S ,
where x ( t ) is the original underwater acoustic signal, M S is the needed number of the DFT.
The Gammatone filter refers to a causal filter with an infinitely long sequence impulse response. In the filter bank, the time domain impulse response of each Gammatone filter can be considered as the product of the Gammatone function, and the acoustic signals can be calculated as below:
g i ( k ) = k n 1 exp ( 2 π B i k ) cos ( 2 π f i + φ i ) u ( k ) ,   1 i N ,
where n is the order of the filter, B i is the attenuation factor of the filter, f i is the center frequency, φ i is the phase of the filter, u ( k ) is the step function, N is the total number of filters.
When extracting the features from the underwater acoustic signals, the bandwidth of each filter is determined by the critical band of the human ear for simulating human auditory features. The critical frequency band is expressed as
E R B ( f i ) = 24.7 × ( 4.37 f i / 1000 + 1 ) ,
b i = 1.019 E R B ( f i ) ,
where b i is the bandwidth of each sub-band filter in the Gammatone filter bank, which is obtained from the critical band.
The energy spectrum E S ( i ) is shown as
E S ( i ) = ln [ k = 0 Q 1 | x ( k ) | 2 g i ( k ) ] ,
where Q is the number of filters and i = 1 , 2 , , Q , { k b i } i = 0 Q + 1 is the boundary of the filters.
The DCT is applied to calculate the logarithm of the filter bank as described by the following equation:
G F C C ( n ) = k = 0 Q 1 E S ( i ) cos ( ( π n ( i 0.5 ) ) Q ) ,    0 n Q 1 .
To verify the validity of the proposed multi-dimensional fusion features method in this paper, Figure 2 shows the time-domain waveform of the original underwater acoustic signals, which is a section of underwater mammals in the dataset. The feature extraction results of original underwater acoustic signals are given in Figure 3 on the original underwater acoustic signals shown in Figure 2.
As depicted in Figure 3, because the MFCC pays more attention to semantic features, feature extraction results of original underwater acoustic signals are rough in Figure 3a. The GFCC extract features of original underwater acoustic signals in Figure 3b, which can remove considerable redundant noise information in underwater acoustic signals, improve robustness, and retain effective voiceprint feature. Moreover, the extracted features are more precise and accurate.
To further compare the anti-noise performance of GFCC and MFCC in the feature extraction process, Gaussian white noise was added into Figure 2. When the signal-to-noise ratio was 10 dB, the time domain waveform of noise signals is shown in Figure 4. Figure 5 shows the feature extraction results of the MFCC and GFCC algorithm on the original underwater acoustic signals shown in Figure 4.
It can be seen from Figure 5a,b that the GFCC algorithm has strong anti-noise performance and has superiority in the feature extraction of underwater acoustic signals targets. It is beneficial to subsequent recognition research. But the features, distributions, and sizes extracted by the MFCC change significantly and it affects the accuracy of underwater acoustic target recognition.
Meanwhile, to verify the effectiveness and adaptability of the GFCC again, Table 1 shows MFCC-GMM and GFCC-GMM recognition accuracy when dataset did not have underwater background noise with weak targets.
It can be seen from Table 1, the recognition accuracy of the GFCC-GMM is higher than that of MFCC-GMM. The GFCC can better describe underwater acoustic signals. Therefore, The GFCC has better robustness than the MFCC, and extracting voiceprint features has better adaptability to underwater acoustic target recognition.

2.2. Modified Empirical Mode Decomposition

MEMD is an improved algorithm for empirical mode decomposition (EMD). It refers to a signal decomposition method based on the local features of signals. The method exploits the advantages of the wavelet transform. Meanwhile, it solves the problem of selecting a wavelet basis and determining the decomposition scale in the wavelet transform [25]. Therefore, it is more suitable for non-linear and non-stationary signals analysis. MEMD is a white-adapted signal decomposition method, which can be employed for the analysis of underwater acoustic signals. According to empirical mode decomposition, it is assumed that all complex signals consist of simple IMF, and each IMF is independent of each other. The algorithm process of MEMD is shown in Figure 6. It can decompose the different scales or trend components of underwater acoustic signals step by step and produce a series of data sequences with the same feature scales. Compared with the original underwater acoustic signals, the decomposed sequence has stronger regularity.
MEMD calculates the x-coordinates of the interpolation points (IPs) by the following equation:
z ( t ) = d 3 d t 3 x ( t ) , t max = { t : d d t z ( t ) = 0 , d 2 d t 2 z ( t ) < 0 } t min = { t : d d t z ( t ) = 0 , d 2 d t 2 z ( t ) > 0 }
where t max and t min contain the IPs.
To extract the IE and IF, IMFs are usually transformed by the HHT as below:
H ( t ) = 1 π + H ( τ ) t τ d τ ,
where H ( τ ) is the IMF.
The analytic signals can be calculated as follows:
V ( t ) = H ( t ) + j H ( t ) = a ( t ) e j ϑ ( t ) ,
where a ( t ) is the model of H ( t ) , j is the imaginary unit.
a ( t ) is described by the following equation:
a ( t ) = H ( t ) 2 + H ( t ) 2 ,
where ϑ ( t ) is the phase of H ( t ) , it can be calculated as below:
ϑ ( t ) = arctan ( H ( t ) H ( t ) ) .
The IE a 2 ( t ) and IF can be extracted by analyzing the signals. The IF is obtained by the following equation:
f ( t ) = 1 2 π d ϑ ( t ) d t .

2.3. Multi-Dimensional Fusion Features Algorithm

The underwater environment is complex and changeable, so it is difficult to describe underwater acoustic signals with a single feature. GFCC and MFCC are both algorithms for simulating human hearing. The abilities to extract the infrasound or ultrasound features are limited, but the MEMD algorithm can effectively extract these features. Therefore, GFCC and MEMD were developed to extract multi-dimensional features in this paper. On this basis, to ensure the same time dimension, a dimension reduction method is proposed to obtain multi-dimensional fusion features in the original underwater acoustic signals.

2.3.1. Multi-Dimensional Feature Extraction

Figure 7 shows the feature extraction process. The feature extraction algorithm can extract the IF and IE from the IMFs, Meanwhile, it can extract the GFCC features and combine it with IF and IE to improve the capabilities of the algorithm. The process can eventually extract features that contain three dimensions and construct a feature vector.
The global feature vector g can be expressed as
g i = { G i ( t ) , a i 2 ( t ) , f i ( t ) } ,
where t is time series (sampling points), G i ( t ) is the feature value of the GFCC at time i , a i 2 ( t ) is IE at time i , f i ( t ) is IF at time i .
Figure 8 shows the structure of the feature fusion matrix. The feature fusion matrix is sorted in the same time sequence, which holds the same sequence. Furthermore, it can merge multiple features. It is noteworthy that the weight of GFCC is adjusted to the same proportion as that of IE and IF, and the ratio can be considered as a research direction in future studies. Since the features are exceedingly redundant, and the same time series are ensured, it is necessary to perform dimensionality reduction operations, which is similar to random pooling and framing for IE and IF.

2.3.2. Dimension Reduction Method

The HHT is used to extract IE and IF on IMFs in a certain time series (sampling points), and the GFCC divides frames before Gammatone filtering. As a result, the temporal features of IE and IF are different from those of the GFCC. Therefore, it is necessary to reduce the dimension of IE and IF.
Figure 9 shows the dimensionality reduction algorithm structure diagram. It is a one-dimensional feature (IE or IF), and the red box represents the pooled domain. The size and step size of the pooled domain is identical to the frame length when the GFCC frames are divided. The frame length equals the ratio of the one-dimensional feature (IE or IF) to the GFCC feature length.
The specific process of the proposed multi-dimensional fusion features is shown as Algorithm 1.
Algorithm 1 Multi-dimensional Fusion Feature
Input: original underwater acoustic signals s ( t ) ;
Output: fusion feature vector;
Initialization: M S (the number points in the DFT), Q (the number of filters), i = 1 , 2 , , Q , { k b i } i = 0 Q + 1 , N , M ;
Procedure:
1: Let x ( t ) s ( t ) ;
2: Calculate x ( k ) using Equation (1);
3: Calculate T r g i ( k ) using Equation (2);
4: Calculate the energy spectrum E S ( l ) using Equation (3);
5: Calculate the k b i using Equation (6);
6: Calculate the GFCC using Equation (7);
7: Let r 0 ( t ) s ( t ) , h k 0 ( t ) r k 1 ( t ) ;
8: For m ← 1 to M or he number of extrema in r k ( t ) is 2 or less do:
9:  For n ← 1 to N do:
10:    Calculate the x-coordinates of the IPs, i.e., obtain t max and t min using Equation (11);
11:    Calculate the y-coordinates of the IPs, h k n 1 ( t ) ;
12:     y max h k n 1 ( t max ) , y min h k n 1 ( t min ) ;
13:    Create the maxima envelope, e max ( t ) using cubic spline interpolation, with the IPs as { t max , y max } ;
14:    Create the minima envelope, e min ( t ) , using cubic spline interpolation, with the IPs as { t min , y min } ;
15:    Deduce the mean envelope, e ( t ) e max ( t ) + e min ( t ) 2 ;
16:     h k n ( t ) h k n 1 ( t ) e ( t ) ;
17:   h k ( t ) h k N ( t ) , r k ( t ) r k 1 ( t ) h k ( t ) ;
18: Decomposition results: s ( t ) r M ( t ) + k = 1 M h k ( t ) ( h k ( t ) is the kth order of IMFs);
19: Dimension Reduction to IMFs;
20: Calculate the H ( t ) using Equation (8);
21: Calculate the IE and IF using Equation (10) to Equation (12);
22: Construct fusion feature vector using Equation (13).

3. Modified Deep Neural Network

When a DNN is directly used to recognize underwater acoustic targets, due to the direct input of the features of the underwater acoustic signals, there exist many redundant features in the processing of recognition. The GMM can extract the statistical parameters of the underwater acoustic signals, which can reduce the length of the underwater acoustic signals. The GMM can reduce the size of the DNN model, can reduce redundant features, and further improve recognition accuracy. Therefore, The GMM is used to modify the structure of the DNN in this paper. The modified DNN, which is shown in Figure 10, is a fully connected feedforward neural network with multiple hidden layers, and it can better accomplish underwater acoustic target recognition tasks.
The Gaussian probability density function of each component is weighted to yield the GMM probability density function with order, which is expressed as
P ( x | λ ) = i = 1 M w i N i ( x ) ,
where x is the D dimension of the eigenvector, N i ( x ) , i = 1 , 2 , , M is sub-distribution, w i , i = 1 , 2 , , M is mixed weight.
Each sub-distribution N i ( x ) refers to a D dimensional joint Gaussian probability distribution.
N i ( x ) = 1 ( 2 π ) D / 2 | Σ i | 1 / 2 exp { 1 2 ( x μ i ) T Σ i 1 ( x μ i ) } ,
where μ i is mean vector, Σ i is covariance matrix, T is the number of feature vectors, mixed weights w i , i = 1 , 2 , , M satisfies i = 1 M w i = 1 .
The full parameters of the GMM model for each kind of underwater acoustic signal targets are composed by the mean vector of each component, covariance matrix, and the set of mixed weights. The eigenvectors corresponding to the parameters are expressed as follows:
λ = { w i , μ i , Σ i } , i = 1 , 2 , , M .
The feature vector sequence extracted from training data is X = { x t } , t = 1 , 2 , , T and the likelihood probability defined as
P ( X | λ ) = t = 1 T P ( x t | λ ) .
Substituting it into the Gauss density function, as follows:
P ( X | λ ) = t = 1 T lg { i = 1 M w i N i ( x t , μ i , Σ i ) } .
The parameter estimation method adopted in the GMM model is maximum likelihood estimation (MLE). According to the feature vector sequence extracted from training data, the parameters of the model are adjusted continuously until the maximum likelihood probability is P ( X | λ ) . The maximum likelihood probability is λ i , the model parameters are
λ i = arg   max   P ( X | λ ) .
According to Equation (16), the input parameters of the DNN are defined as
r = Σ i w i × μ i μ i , u b m ( d i a g { Σ i } ) 1 / 2 ,
where μ i , u b m is the mean vector of the universal background model (UBM).
Given a dimension input eigenvector, the activation vector of the first hidden layer is expressed as
h ( 1 ) = σ ( W ( 1 ) T x + b ( 1 ) ) ,
where W ( 1 ) T is the first hidden layer weight matrix transpose, dimension is I × N 1 , b ( 1 ) is the size of the offset vector, σ is activation function of the hidden layer.
The activation vectors of the i hidden layer are obtained by the activation vectors h ( i ) of the second hidden layer.
h ( i ) ( x ) = σ ( W ( i ) T h ( i 1 ) ( x ) + b ( i 1 ) ) ,
where N i is the number of neurons in the i hidden layer, W ( i ) T is the transpose of the weight matrix of the i hidden layer, the dimension is N i 1 × N i , and b ( i ) is the offset vector of size.
Since the ReLU function performs well in DNN classification and recognition tasks, the ReLU function serves as hidden layer activation functions. It is specifically defined as:
σ ( a ) = { a i f a > 0 0 i f a 0 ,
The output layer of the DNN uses the SoftMax function to complete the output category. It finally realizes underwater acoustic signals classification and recognition. The SoftMax function is expressed as
s ( z k ) = e z k j = 1 J e z j ,
where z k is the vector of the dimension output layer.
SoftMax regression algorithm serves as the loss function.
J ( w ) = 1 m [ i = 1 m j = 1 d 1 { y ( i ) = d } log e w j T z ( i ) l = 1 d e w l T z ( j ) ] ,
where log e w j T z ( i ) l = 1 d e w l T z ( j ) is the Logarithmic values for softmax function, 1 { } is the characteristic function. y(i) is the predicted label value of the i data in the dataset, d is the label value, when y ( i ) = d is true, return 1, otherwise return 0.

4. Experiment Results and Analysis

This section shows numerical examples to validate the generality and effectiveness of the proposed MFF-MDNN for underwater acoustic target recognition. The dataset was divided into six categories, including four types of ships, underwater mammals, and underwater background noise with weak targets in this paper. The total length of our dataset is almost 20 h. Each acoustic signal was divided into two seconds, and the background noise dataset was used to simulate the general situation of the underwater. The train set was three times as large as the test set in the experiments.
To demonstrate the MEMD suitable for feature extraction in underwater acoustic targets, Figure 11 shows the first 5 IMFs waveforms extracted of the EMD and the MEMD on the original underwater acoustic signals shown in the section of Figure 2. Figure 11a decomposes the IMFs waveform graph of EMD. Figure 11b decomposes the IMFs waveform of the MEMD.
From Figure 11, it can be observed that the underwater environment was extremely complex and that the EMD shows little difference between each order IMF and the original underwater acoustic signals. While MEMD can decompose the original underwater acoustic signals more distinctly, it can improve the effectiveness of feature information extraction in underwater acoustic signals, which is conducive to the subsequent recognition of underwater targets.
Therefore, the proposed multi-dimensional fusion features method has validity to some extent from Figure 3 and Figure 5, Table 1, and Figure 11 in this paper.
Similarly, to verify the superiority of the proposed modified DNN in this paper, Figure 12 shows the recognition accuracy of 30 experiments with a different training set and testing set in the dataset when the maximum iteration was 800, which included the GMM recognition method using the MFCC feature extraction (MFCC-GMM) [11,29], the proposed modified DNN recognition method using MFCC feature extraction (MFCC-MDNN) [11], the proposed modified DNN recognition method using GFCC feature extraction (GFCC-MDNN) [18].
As depicted in Figure 12, the recognition accuracy of the proposed modified DNN is higher than that of the GMM. Although the GMM can recognize underwater acoustic targets, it is a shallow recognition model in which the recognition accuracy is relatively low. Meanwhile, the recognition accuracy of the GFCC-MDNN was higher than that of the MFCC-MDNN, which proves that the GFCC has strong adaptability to modify the DNN in this paper, and it is suitable for underwater acoustic target recognition compared to MFCC.
To further demonstrate the recognition accuracy of the proposed the MFF-MDNN in this paper, Figure 13 shows the recognition accuracy of the proposed modified the DNN recognition method using GFCC feature extraction (GFCC-MDNN) [18], the proposed modified DNN recognition method using MFCC and MEMD feature extraction (MM-MDNN) [16], and the proposed MFF-MDNN.
As seen in Figure 13, the recognition accuracy of the proposed MFF-MDNN was higher than the other method. Therefore, the proposed MFF-MDNN has better recognition accuracy for underwater acoustic target recognition in this paper.
To describe more clearly the reliability and stability of the proposed MFF-MDNN method, Table 2 shows recognition results of 30 experiments with a different training set and testing set in the dataset when the maximum iteration was 800, which includes the MFCC-GMM, GFCC-GMM, MFCC-MDNN, GFCC-MDNN, MM-MDNN, and the proposed MFF-MDNN.
As seen from Table 2, the recognition accuracy of the proposed MFF-MDNN method was higher than that of the other algorithms in the 30 experiments with a different training set and testing set in the dataset when the maximum iteration was 800. In this paper, the proposed multi-dimensional fusion features method can describe underwater acoustic signals from multiple angles. It combines the advantages of the GFCC with MEMD, which is more suitable for underwater acoustic target recognition than the single feature. Moreover, GMM was used to extract the statistical parameters of the feature matrix, which modify the structure of DNN. It can reduce redundant features and further improve recognition accuracy. From the above comparative experiments in Table 1 and Table 2, the following results can be drawn. The proposed MFF-MDNN method can improve accuracy when the dataset has underwater background noise with weak targets. Therefore, the recognition results demonstrate that the proposed MFF-MDNN method has higher accuracy and strong adaptability over other methods.

5. Conclusions

In this paper, a combination of multi-dimensional fusion features and modified DNN method was proposed to recognize underwater acoustic targets. The problem where the single feature could not describe underwater acoustic signals well was solved effectively, using the GFCC and MEMD to extract multi-dimensional features. On this basis, a dimension reduction method was proposed, which fused the multi-dimensional features in the same time dimension. It could obtain multi-dimensional fusion features in the original underwater acoustic signals. In addition, the problem of many redundant features in the processing of recognition was solved by utilizing GMM to modify the structure of DNN. It could improve recognition accuracy. The proposed underwater acoustic target recognition method was employed on a dataset, which was divided into six categories, including four types of ships, underwater mammals, and underwater background noise with weak targets. The average recognition accuracies of the proposed MFF-MDNN, MM-MDNN, GFCC-MDNN, MFCC-MDNN, GFCC-GMM, and MFCC-GMM were 94.3%, 91.1%, 78.2%, 73.1%, 62.2%, and 59.0%, respectively, when the maximum iterations were 800. The recognition results showed that the method had good validity and adaptability. However, MFF-MDNN has a higher time complexity due to the algorithm principle of MEMD and the particularity of underwater acoustic signals. Further studies can be conducted in the future in this direction.

Author Contributions

Conceptualization, X.W., A.L., Y.Z., and F.X.; Funding acquisition, X.W.; Investigation, Y.Z. and F.X.; Methodology, A.L.; Project administration, X.W.; Software, Y.Z. and F.X.; Supervision, X.W. and F.X.; Writing—original draft, A.L. and Y.Z.; Writing—review and editing, X.W., A.L., Y.Z., and F.X.

Funding

This work was supported by the National Natural Science Foundation of China grant number 41876110 and Fundamental Research Funds for the Central Universities grant number 3072019CFT0602.

Acknowledgments

The authors are grateful to the guest editors and anonymous reviewers for their constructive comments based on which the presentation of this paper has been greatly improved.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, H.; Shen, S.; Yao, X.; Sheng, M.; Wang, C. Competitive Deep-Belief Networks for Underwater Acoustic Target Recognition. Sensors 2018, 18, 952. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, X.; Jiao, J.; Yin, J.; Zhao, W.; Han, X.; Sun, B. Underwater sonar image classification using adaptive weights convolutional neural network. Appl. Acoust. 2018, 146, 145–154. [Google Scholar] [CrossRef]
  3. Liang, H.; Liang, Q.L. Target Recognition Based on 3-D Sparse Underwater Sonar Sensor Network. Lect. Notes Electr. Eng. 2019, 463, 2552–2564. [Google Scholar]
  4. Wu, Y.; Li, X.; Wang, Y. Extraction and classification of acoustic scattering from underwater target based on Wigner-Ville distribution. Appl. Acoust. 2018, 138, 52–59. [Google Scholar] [CrossRef]
  5. González-Hernández, F.R.; Sánchez-Fernández, L.P.; Suárez-Guerra, S.; Sánchez-Pérez, L.A. Marine mammal sound classification based on a parallel recognition model and octave analysis. Appl. Acoust. 2017, 119, 17–28. [Google Scholar] [CrossRef]
  6. Can, G.; Akbas, C.E.; Cetin, A.E. Recognition of vessel acoustic signatures using non-linear teager energy based features. In Proceedings of the 2016 International Workshop on Computational Intelligence for Multimedia Understanding (IWCIM), Reggio Calabria, Italy, 27–28 October 2016. [Google Scholar]
  7. Novitzky, M.; Pippin, C.; Collins, T.R.; Balch, T.R.; West, M.E. AUV behavior recognition using behavior histograms, HMMs, and CRFs. Robot 2014, 32, 291–304. [Google Scholar] [CrossRef]
  8. Murino, V. Reconstruction and segmentation of underwater acoustic images combining confidence information in MRF models. Pattern Recognit. 2001, 34, 981–997. [Google Scholar] [CrossRef]
  9. Tegowski, J.; Koza, R.; Pawliczka, I.; Skora, K.; Trzcinska, K.; 71Zdroik, J. Statistical, Spectral and Wavelet Features of the Ambient Noise Detected in the Southern Baltic Sea. In Proceedings of the 23rd International Congress on Sound and Vibration: From Ancient to Modern Acoustics, Int Inst Acoustics & Vibration, Auburn, Al, USA, 10–14 July 2016. [Google Scholar]
  10. Parada, P.P.; Cardenal-López, A. Using Gaussian mixture models to detect and classify dolphin whistles and pulses. J. Acoust. Soc. Am. 2014, 135, 3371–3380. [Google Scholar] [CrossRef] [PubMed]
  11. Lim, T.; Bae, K.; Hwang, C.; Lee, H. Classification of underwater transient signals using MFCC feature vector. In Proceedings of the 9th International Symposium on Signal Processing and Its Applications (ISSPA), Sharjah, United Arab, 12–15 February 2007. [Google Scholar]
  12. Lim, T.; Bae, K.; Hwang, C.; Lee, H. Underwater Transient Signal Classification Using Binary Pattern Image of MFCC and Neural Network. IEICE Trans. Fundam. Electron. Commun. Comput. Sci. 2008, E91A, 772–774. [Google Scholar] [CrossRef]
  13. Jankowski, C., Jr.; Quatieri, T.; Reynolds, D. Measuring fine structure in speech: Application to speaker identification. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, Detroit, MI, USA, 9–12 May 1995; IEEE: Piscataway, NJ, USA; pp. 325–328. [Google Scholar]
  14. Guo, Y.; Gas, B. Underwater transient and non transient signals classification using predictive neural networks. In Proceedings of the 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, St. Louis, MO, USA, 10–15 October 2009; pp. 2283–2288. [Google Scholar]
  15. Wang, W.; Li, S.; Yang, J.; Liu, Z.; Zhou, W. Feature Extraction of Underwater Target in Auditory Sensation Area Based on MFCC. In Proceedings of the 2016 IEEE/OES China Ocean Acoustics Symposium (COA), Harbin, China, 9–11 January 2016. [Google Scholar]
  16. Sharma, R.; Vignolo, L.; Schlotthauer, G.; Colominas, M.; Rufiner, H.L.; Prasanna, S. Empirical mode decomposition for adaptive AM-FM analysis of speech: A review. Speech Commun. 2017, 88, 39–64. [Google Scholar] [CrossRef]
  17. Holambe, R.S.; Deshpande, M.S. Advances in Non-Linear Modeling for Speech Processing; Springer Science and Business Media: Berlin, Germany, 2012. [Google Scholar]
  18. Lian, Z.; Xu, K.; Wan, J.; Li, G. Underwater acoustic target classification based on modified GFCC features. In Proceedings of the IEEE 2nd Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chongqing, China, 25–26 March 2017; pp. 258–262. [Google Scholar]
  19. Mao, Z.; Wang, Z.; Wang, D. Speaker recognition algorithm based on Gammatone filter bank. Comput. Eng. Appl. 2015, 51, 200–203. [Google Scholar]
  20. Grimaldi, M.; Cummins, F. Speaker Identification Using Instantaneous Frequencies. IEEE Trans. Audio Speech Lang. Process. 2008, 16, 1097–1111. [Google Scholar] [CrossRef]
  21. Hasan, T.; Hansen, J.H. Robust speaker recognition in non-stationary room environments based on empirical mode decomposition. In Proceedings of the Interspeech, Florence, Italy, 27–31 August 2011; pp. 2733–2736. [Google Scholar]
  22. Zeng, X.; Wang, S. Underwater sound classification based on Gammatone filter bank and Hilbert-Huang transform. In Proceedings of the 2014 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Guilin, China, 5–8 August 2014; pp. 707–710. [Google Scholar]
  23. Deshpande, M.S.; Holambe, R.S. Speaker Identification Based on Robust AM-FM Features. In Proceedings of the Second International Conference on Emerging Trends in Engineering & Technology, Nagpur, India, 16–18 December 2009; pp. 880–884. [Google Scholar]
  24. Deshpande, M.S.; Holambe, R.S. Am-fm based robust speaker identification in babble noise. Environments 2011, 6, 19. [Google Scholar]
  25. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.-C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  26. E Huang, N.; Wu, M.-L.C.; Long, S.R.; Shen, S.S.; Qu, W.; Gloersen, P.; Fan, K.L.; Wu, M.-L.C. A confidence limit for the empirical mode decomposition and Hilbert spectral analysis. Proc. R. Soc. A Math. Phys. Eng. Sci. 2003, 459, 2317–2345. [Google Scholar] [CrossRef]
  27. Huang, H.; Chen, X.-X. Speech formant frequency estimation based on Hilbert–Huang transform. J. Zhejiang Univ. Eng. Sci. 2006, 40, 1926. [Google Scholar] [CrossRef]
  28. Sharma, R.; Prasanna, S.M. A better decomposition of speech obtained using modified Empirical Mode Decomposition. Digit. Signal Process. 2016, 58, 26–39. [Google Scholar] [CrossRef]
  29. Huang, N.E.; Shen, S.S.P. Hilbert-Huang Transform and Its Applications; World Scientific: Singapore, 2005. [Google Scholar]
  30. Huang, H.; Pan, J. Speech pitch determination based on Hilbert-Huang transform. Signal Process. 2006, 86, 792–803. [Google Scholar] [CrossRef]
  31. Hayakawa, S.; Itakura, F. Text-dependent speaker recognition using the information in the higher frequency band. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, Adelaide, Australia, 19–22 April 1994; IEEE: Piscataway, NJ, USA; pp. 1–137. [Google Scholar]
  32. Kotari, V.; Chang, K.C. Fusion and Gaussian Mixture Based Classifiers for SONAR data. In Signal Processing, Sensor Fusion, and Target Recognition XX; International Society for Optics and Photonics: Orlando, FL, USA, 5 May 2011. [Google Scholar]
  33. Liu, S.; Sim, K.C. On combining DNN and GMM with unsupervised speaker adaptation for robust automatic speech recognition. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 195–199. [Google Scholar]
  34. Wang, Q.; Wang, L.; Zeng, X.; Zhao, L. An Improved Deep Clustering Model for Underwater Acoustical Targets. Neural Process. Lett. 2018, 48, 1633–1644. [Google Scholar] [CrossRef]
  35. Ibrahim, A.K.; Zhuang, H.; Chérubin, L.M.; Schärer-Umpierre, M.T.; Erdol, N. Automatic classification of grouper species by their sounds using deep neural networks. J. Acoust. Soc. Am. 2018, 144, EL196–EL202. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Li, J.; Dai, W.; Metze, F.; Qu, S.; Das, S. A comparison of Deep Learning methods for environmental sound detection. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 126–130. [Google Scholar]
  37. Dumpala, S.H.; Kopparapu, S.K. Improved speaker recognition system for stressed speech using deep neural networks. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1257–1264. [Google Scholar]
Figure 1. The process of the Gammatone frequency cepstral coefficient (GFCC).
Figure 1. The process of the Gammatone frequency cepstral coefficient (GFCC).
Remotesensing 11 01888 g001
Figure 2. Original underwater acoustic signals.
Figure 2. Original underwater acoustic signals.
Remotesensing 11 01888 g002
Figure 3. Features extraction results of the original underwater acoustic signals: (a) Features of the Mel frequency cepstral coefficient (MFCC) in original underwater acoustic signals; (b) Features of the GFCC in original underwater acoustic signals.
Figure 3. Features extraction results of the original underwater acoustic signals: (a) Features of the Mel frequency cepstral coefficient (MFCC) in original underwater acoustic signals; (b) Features of the GFCC in original underwater acoustic signals.
Remotesensing 11 01888 g003
Figure 4. Underwater acoustic signals with a signals-to-noise ratio of 10 dB.
Figure 4. Underwater acoustic signals with a signals-to-noise ratio of 10 dB.
Remotesensing 11 01888 g004
Figure 5. Features extraction results of underwater acoustic signals with a signals-to-noise ratio 10 dB: (a) Features of MFCC in underwater acoustic signals with a signals-to-noise ratio 10 dB; (b) Features of the GFCC in underwater acoustic signals with a signals-to-noise ratio 10 dB.
Figure 5. Features extraction results of underwater acoustic signals with a signals-to-noise ratio 10 dB: (a) Features of MFCC in underwater acoustic signals with a signals-to-noise ratio 10 dB; (b) Features of the GFCC in underwater acoustic signals with a signals-to-noise ratio 10 dB.
Remotesensing 11 01888 g005
Figure 6. The flowchart of modified empirical mode decomposition (MEMD).
Figure 6. The flowchart of modified empirical mode decomposition (MEMD).
Remotesensing 11 01888 g006
Figure 7. The process of feature extraction.
Figure 7. The process of feature extraction.
Remotesensing 11 01888 g007
Figure 8. Fusion feature matrix structure.
Figure 8. Fusion feature matrix structure.
Remotesensing 11 01888 g008
Figure 9. Dimensionality reduction algorithm structure diagram.
Figure 9. Dimensionality reduction algorithm structure diagram.
Remotesensing 11 01888 g009
Figure 10. Schematic diagram of the modified deep neural network (DNN).
Figure 10. Schematic diagram of the modified deep neural network (DNN).
Remotesensing 11 01888 g010
Figure 11. The first 5 intrinsic mode functions (IMFs) waveforms extracted from the empirical mode decomposition (EMD) and MEMD: (a) The first 5 IMFs waveforms extracted from the EMD; (b) The first 5 IMFs waveforms extracted from the MEMD.
Figure 11. The first 5 intrinsic mode functions (IMFs) waveforms extracted from the empirical mode decomposition (EMD) and MEMD: (a) The first 5 IMFs waveforms extracted from the EMD; (b) The first 5 IMFs waveforms extracted from the MEMD.
Remotesensing 11 01888 g011
Figure 12. The recognition accuracy of the Mel frequency cepstral coefficient-Gaussian mixture model (MFCC-GMM), the Mel frequency cepstral coefficient-modified deep neural network (MFCC-MDNN) and the Gammatone frequency cepstral coefficient-modified deep neural network (GFCC-MDNN).
Figure 12. The recognition accuracy of the Mel frequency cepstral coefficient-Gaussian mixture model (MFCC-GMM), the Mel frequency cepstral coefficient-modified deep neural network (MFCC-MDNN) and the Gammatone frequency cepstral coefficient-modified deep neural network (GFCC-MDNN).
Remotesensing 11 01888 g012
Figure 13. The recognition accuracy of the GFCC-MDNN, the MM-MDNN, and the multi-dimensional fusion features and a modified deep neural network MFF-MDNN.
Figure 13. The recognition accuracy of the GFCC-MDNN, the MM-MDNN, and the multi-dimensional fusion features and a modified deep neural network MFF-MDNN.
Remotesensing 11 01888 g013
Table 1. The estimated optimal threshold and η of 20 different underwater sonar images.
Table 1. The estimated optimal threshold and η of 20 different underwater sonar images.
Experiment TimesMFCC-GMMGFCC-GMM
10.850.94
20.880.90
30.880.90
40.860.93
50.870.92
60.890.92
70.890.95
80.880.92
90.870.92
100.900.90
Table 2. The recognition results of 30 experiments with a different training set and testing set in the dataset.
Table 2. The recognition results of 30 experiments with a different training set and testing set in the dataset.
Experiment TimesMFCC-GMMGFCC-GMMMFCC-MDNNGFCC-MDNNMM-MDNNMFF-MDNN
10.590.610.740.780.910.94
20.600.620.720.760.910.95
30.620.650.760.790.910.94
40.640.600.740.790.920.94
50.580.640.730.780.910.95
60.580.620.730.790.910.95
70.600.650.720.790.910.94
80.610.640.710.790.910.94
90.580.610.720.770.920.94
100.580.630.740.780.910.95
110.600.640.740.780.910.93
120.580.650.720.760.910.94
130.580.650.750.780.910.95
140.580.590.740.760.910.94
150.590.600.720.800.920.94
160.580.620.730.790.900.94
170.600.610.740.780.900.92
180.580.620.720.790.900.94
190.570.610.750.780.920.94
200.590.600.740.770.910.94
210.580.630.720.790.910.94
220.580.610.710.760.910.95
230.580.600.730.780.910.94
240.600.620.740.790.920.93
250.610.630.720.790.910.93
260.580.610.740.790.900.94
270.580.600.700.790.910.94
280.600.620.720.780.910.94
290.570.640.740.780.920.94
300.590.640.720.800.910.94

Share and Cite

MDPI and ACS Style

Wang, X.; Liu, A.; Zhang, Y.; Xue, F. Underwater Acoustic Target Recognition: A Combination of Multi-Dimensional Fusion Features and Modified Deep Neural Network. Remote Sens. 2019, 11, 1888. https://doi.org/10.3390/rs11161888

AMA Style

Wang X, Liu A, Zhang Y, Xue F. Underwater Acoustic Target Recognition: A Combination of Multi-Dimensional Fusion Features and Modified Deep Neural Network. Remote Sensing. 2019; 11(16):1888. https://doi.org/10.3390/rs11161888

Chicago/Turabian Style

Wang, Xingmei, Anhua Liu, Yu Zhang, and Fuzhao Xue. 2019. "Underwater Acoustic Target Recognition: A Combination of Multi-Dimensional Fusion Features and Modified Deep Neural Network" Remote Sensing 11, no. 16: 1888. https://doi.org/10.3390/rs11161888

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop