Next Article in Journal
Entropy-Based Video Steganalysis of Motion Vectors
Next Article in Special Issue
Polynomial-Time Algorithm for Learning Optimal BFS-Consistent Dynamic Bayesian Networks
Previous Article in Journal
Entropy Generation Analysis of Laminar Flows of Water-Based Nanofluids in Horizontal Minitubes under Constant Heat Flux Conditions
Previous Article in Special Issue
Q-Neutrosophic Soft Relation and Its Application in Decision Making
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Compression of a Deep Competitive Network Based on Mutual Information for Underwater Acoustic Targets Recognition

School of Marine Science and Technology, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Entropy 2018, 20(4), 243; https://doi.org/10.3390/e20040243
Submission received: 24 January 2018 / Revised: 20 March 2018 / Accepted: 27 March 2018 / Published: 2 April 2018
(This article belongs to the Special Issue Information Theory in Machine Learning and Data Science)

Abstract

:
The accuracy of underwater acoustic targets recognition via limited ship radiated noise can be improved by a deep neural network trained with a large number of unlabeled samples. However, redundant features learned by deep neural network have negative effects on recognition accuracy and efficiency. A compressed deep competitive network is proposed to learn and extract features from ship radiated noise. The core idea of the algorithm includes: (1) Competitive learning: By integrating competitive learning into the restricted Boltzmann machine learning algorithm, the hidden units could share the weights in each predefined group; (2) Network pruning: The pruning based on mutual information is deployed to remove the redundant parameters and further compress the network. Experiments based on real ship radiated noise show that the network can increase recognition accuracy with fewer informative features. The compressed deep competitive network can achieve a classification accuracy of 89.1 % , which is 5.3 % higher than deep competitive network and 13.1 % higher than the state-of-the-art signal processing feature extraction methods.

1. Introduction

Underwater acoustic targets recognition based on ship radiated noise is one of the main functions of passive sonar system. The acquired underwater acoustic signals are usually noisy due to the complexity of sound propagation in shallow sea and the frequent presence of high background noise in the sensor. Nowadays, underwater acoustic targets recognition still depends on the decision of well-trained sonarmen, but it is difficult to implement continuous monitoring and recognition. Hence, an unattended underwater acoustic targets recognition system with high recognition accuracy and efficiency needs to be developed to achieve real-time targets recognition.
In order to build an automatic underwater acoustic targets recognition system, various signal processing strategies were applied to extract features and design classifiers. In addition, feature selection and compression methods were studied to improve the classification accuracy and efficiency. The extracted features mainly include: cepstral features, waveform features, auditory features and wavelet features. Das [1] extracted spectral features and cepstral coefficients. Mel-frequency cepstral coefficients (MFCC) were applied to describe underwater acoustic targets by Zhang [2]. A vessel classifier was developed based on cepstral coefficients and Gaussian mixture models by Santos [3]. Zero-crossing features and peak-to-peak amplitude features were presented to describe rotation of propeller by Meng [4,5]. However, their performances were greatly reduced in noisy shallow sea. Azimi-Sadjadi [6] studied wavelet packets. However, it was difficult to determine the decomposition series of a wavelet for lack of prior knowledge. Siddagangaiah [7] studied the multiscale entropy method to detect and recognize underwater acoustic targets. A common shortcoming of the feature extraction methods discussed above is that they require priori training samples to perform the analysis. In the field of underwater acoustic targets classifier design, multiple support vector machine (SVM) classifiers were integrated to improve classification accuracy and robustness by Yang [8]. A neural classifier based on feed forward neural network was studied by Filho [9]. However, the convergence efficiency remains to be discussed. In the field of feature selection and compression, Yang [8] proposed feature selection algorithms for underwater acoustic target recognition. Wei [10] extracted high-order statistics features and compressed it with principal components analysis. However, the compressibility and efficiency improvement is still worth discussing.
In recent years, deep learning has made great achievements in the field of speech recognition and image recognition [11]. In the field of underwater acoustic targets recognition, Kamal [12] used deep belief network (DBN) and Cao [13] used sparse deep auto encoder. However, the deep neural networks have not improved in accordance with underwater acoustic targets recognition. Yang [14] proved that unsupervised pre-training of DBN with large number of unlabeled ship radiated noise can be viewed as regularization strategy which can help to initialize the networks. However, it did not enhance the clustering characteristics of hidden units. Shen [15] proposed an improved DBN in which the relevance of hidden units can be reduced with the help of grouping strategy and group sparse punitive function. However, it is inefficient to calculate the mutual information between any two features, especially in the case of too many hidden units. In order to compress the deep neural network, Han [16] proposed a compression strategy using pruning, trained quantization and Huffman coding. However, the network pruning according to a threshold value of connection weights failed to preserve the features with more classification information.
A compressed deep competitive network (DCN) based on competitive learning and network pruning is proposed to learn and extract features from ship radiated noise. Deep neural network integrated with competitive learning can make the samples in the same category more aggregated. The pruning based on mutual information is deployed to remove the redundant parameters and further compress the network. For lacking labeled ship radiated noise, unsupervised pre-training with a large amount of unlabeled data is introduced for initializing the networks parameters.
This paper is organized as follows. Relevant issues about the compression of deep competitive network is discussed in Section 2. Section 3 shows the experiment results and relevant discussions, followed by the conclusions in Section 4.

2. Compression of Deep Competitive Network

Framework of the compression of DCN is shown in Figure 1. Each layer of the network is an improved restricted Boltzmann machine (RBM) [17] integrated with competitive learning and network pruning. The training procedure is described as follows. First, RBM is pre-trained with a large amount of unlabeled data in an unsupervised learning way. Second, competitive layer is constructed by adding lateral connection among the grouped hidden units. Gradient algorithm is applied to update parameters of RBM to build the competitive restricted Boltzmann machine (CRBM). Third, mutual information is deployed to compress the CRBM by pruning the redundant hidden units and connection weights. Finally, compressed DCN is constructed by fitting a stack of compressed CRBM and the output of previous layer is input of next layer. The whole model is then discriminatively fine-tuned by a back propagation algorithm with the target of class labels.

2.1. Restricted Boltzmann Machine

Gaussian-Bernoulli RBM (GB-RBM) is used to deal with real-valued ship radiated noise. The visible units v = ( v 1 , v 2 , , v n ) T are connected to hidden units h = ( h 1 , h 2 , , h m ) T by connection weights. The connection weights and biases define a probability distribution over the joint states of the visible and hidden units via the energy function [11,17]:
E ( v , h | θ ) = i = 1 n ( v i a i ) 2 2 j = 1 m b j h j i = 1 n j = 1 m v i W i j h j
θ = ( W i j , a i , b j ) , where W i j represents the connection weight between visible unit i and hidden unit j, a i and b j are biases terms. n and m are the numbers of visible and hidden units, respectively.
The conditional distribution P ( h | v , θ ) is P ( h j = 1 | v , θ ) = σ ( b j + i v i W i j ) , where σ ( x ) = 1 / ( 1 + e x ) . The conditional distribution P ( v | h , θ ) is P ( v i = 1 | h , θ ) = N ( a j + j h j W i j , 1 ) , where N ( μ , V ) obeys the Gaussian distribution with mean μ and variance V. The parameters of RBM are optimized by the gradient:
l n p ( v ) θ = E h | v [ E ( v , h ) θ ] E h , v [ E ( v , h ) θ ]
where E is the expectation.
Unsupervised pre-training with a large amount of unlabeled data is introduced for initializing the RBM [18]. To reconstruct ship-radiated noise accurately, the hidden units of RBM must contain information about aspects of the data that are not relevant to its classification. To remove the redundant features, competitive learning and network pruning based on mutual information is introduced to compress the network.

2.2. Competitive Restricted Boltzmann Machine

In order to increase the activation level of inactivated hidden units and reduce the number of network parameters, competitive learning [19,20] is used to cluster the hidden units, so that the hidden units in the same group share the similar weights. Shen [15] grouped hidden units of RBM based on mutual information, but they can’t find out the corresponding category of each group and it is inefficient to calculate the mutual information. However, a given hidden unit of well-trained RBM has different activation level on different categories. A category to which a unit is responding maximally could be a good first-order representation of what the unit is doing [21]. These hidden units can be viewed as feature detectors of that category.
A statistical method is adopted by calculating the score of a given hidden unit driven by different categories. RBM with m hidden units can be trained by training data with L categories, where h = ( h 1 , h 2 , , h m ) T represent the hidden units and k ( 1 , 2 , , L ) represents the category number. The activation of hidden unit j is h j ( v , θ ) , which is a function of both parameter θ and input sample v . The score of hidden unit j driven by the k t h category is:
s c o r e ( h j k ) = 1 n k p = 1 n k h j ( v k p , θ ) 1 n / k q = 1 n / k h j ( v / k q , θ )
where v k is a sample of the k t h category, n k is the number of samples in the k t h category, v / k is a sample of other categories and n / k is the number of its samples. The hidden units which have the maximum score driven by the k t h category will be included in the k t h group. In general, each hidden unit belongs to only one group and each group contains at least one hidden unit. Each group is assigned to one of the predetermined categories.
Lateral connections are added between hidden units in different groups to build the competitive layer. Lateral inhibition between groups is from small negative weights. The gradient algorithm is derived to optimize the weights of CRBM, the loss function is:
L o s s = 1 2 j = 1 m [ c j h j ( v ) ] 2
The gradient for each grouped weight is calculated and used to update the weights of RBM. Competitive learning is used to identify the shared weights in each group, so that hidden units in the same group will share the similar weights. Weight sharing is determined by integrating competitive learning into the RBM learning algorithm, so that the shared weights can be optimized dynamically according to the loss function during the training procedure.

2.3. Network Pruning Based on Mutual Information

A well-trained RBM still contains lots of hidden units that are not relevant to the classification and this will inevitably have negative effects on the recognition accuracy. According to the grouping strategy above, the number of hidden units in different groups may be unbalanced. In order to further improve the classification accuracy together with reducing the network parameters, network pruning based on mutual information is proposed. Figure 2 is the diagram of network pruning based on mutual information.
The network is compressed by pruning the redundant hidden units, and then the informative hidden units are reserved. The connection weights are pruned together with the removed hidden units. The significance of each hidden unit is evaluated by calculating normalized mutual information (NMI) [22,23] with labels. Let F denote a feature and L represent the labels. N M I ( F , L ) is defined as:
N M I ( F , L ) = H ( F ) H ( F | L ) m a x [ H ( F ) , H ( L ) ]
where H ( · ) is entropy. Entropy of a feature F with a probability mass function p ( f ) is:
H ( F ) = f p ( f ) log 2 p ( f )
The conditional entropy H ( F | L ) is the entropy of a feature F conditional on the knowledge of labels L. If ( F , L ) p ( f , l ) , H ( F | L ) is:
H ( F | L ) = l f p ( l , f ) log 2 p ( f | l )
In order to estimate entropy of feature F, 1R discretization method [24] is used as a preprocessing step. After sorting the continuous values, 1R divides the range of continuous values into a number of disjoint intervals and adjusts the boundaries based on class labels. Each interval should contain a minimum of six instances. For discrete feature variables, both joint and marginal probability tables can be estimated by tallying the samples of categorical variables in the data [25].
It is obvious that N M I ( F , L ) ranges from 0 to 1. N M I = 1 when the two variables are identical and N M I = 0 when the two variables are independent. The pruning is deployed group by group to balance the number of hidden units in different groups. All hidden units with NMI below a threshold are removed from the network. The network is retrained to learn the final connection weights for the remaining hidden units. The pruning based on mutual information compresses the competitive network by reducing the number of hidden units and weight vectors.

2.4. Compressed Deep Competitive Network

Compressed DCN is constructed by fitting a stack of compressed CRBM and the output of previous layer is input of next layer. Each layer of the network is a CRBM integrated with competitive learning and network pruning. By applying greedy layer-wise training, a compressed DCN with many layers is obtained to represent more complex statistical structure in ship radiated noise. The whole model is then discriminatively fine-tuned by a back propagation algorithm with the target of class labels to further improve classification accuracy [26].

3. Experiments and Discussion

3.1. Experimental Datasets

A dataset of ship radiated noise recorded in the South China Sea was used to verify the proposed model. The dataset can be divided into two categories, including surface targets and autonomous underwater vehicles. The data was acquired using an omnidirectional hydrophone placed at the bottom of a shallow water channel of 30 m depth. The targets were approximately 3.5 km away from the hydrophone, and they moved around the same route with different speeds. Signal to noise ratio (SNR) is 2–3dB. Signals were divided into short frames of 186 ms (4096 samples with the sampling rate of 22,050 Hz). Discrete Fourier Transforms (DFT) of these frames was calculated and only absolute values of the DFT were kept. The final dimension of each sample was 2048 for the symmetry of DFT. Thus, there were 21,530 unlabeled samples and 4210 labeled samples. In the labeled samples, 2100 samples were surface targets and 2110 samples were underwater targets. Labeled dataset was split into two subsets: 50 % for training and 50 % for testing. Min-max normalization was applied to fit the input of the network. Figure 3 is the spectrogram (logarithmic amplitude) of the input data. The horizontal axis represents the discrete frequency steps, the vertical axis of the spectrogram typically represents time, and the amount of power is represented as the intensity at each time-frequency point [7].

3.2. Experimental Procedure

Training algorithms for deep learning models are iterative and thus require initial point from which to begin the iterations. Normalized initialization was used in our experiments just as Glorot [27] suggested. The initialization method is to initialize the weights of RBM with n inputs and m outputs by sampling each weight from:
W i j U ( 6 m + n , 6 m + n )
where U represents uniform distribution. We used a momentum of 0.5 and mini-batches of 50 randomly selected samples. Grouping strategy, competitive learning and mutual information network pruning were added to Deep-Learning-Toolbox [28] to perform the experiments. The experimental procedure is illustrated in Figure 4.
  • Pre-training RBM with 21,530 unlabeled samples with learning rate 0.01.
  • Competitive learning with 4210 labeled training data with learning rate 0.001.
  • Network pruning with the threshold of average NMI in each layer.
  • Greedy layer-wise training and supervised fine-tuning with learning rate 0.01. The initial structure of the DCN is 2048-500-500-50-50. After network pruning, the compressed DCN with structure 2048-163-158-34-31 was obtained.
  • The performances of obtained deep features and four widely-used traditional features were compared via t-SNE (t-distributed stochastic neighbor embedding) [29], NMI, classification accuracy and receivers operating characteristic (ROC) curve. In order to make a fair comparison, we used the same classifier for different feature sets. LIB-SVM [30] was used to deal with the small sample size classification problem.
The four traditional feature sets were MFCC features [2], waveform features [4,5], auditory features [31] and wavelet features [6]. MFCC features were extracted by taking the coefficients that make up a Mel-frequency cepstrum. First-order differential Mel-frequency cepstrum coefficients (DMFCC) and second-order differential Mel-frequency cepstrum coefficients (DDMFCC) were calculated. Waveform features were extracted via signal statistical characteristics of zero-crossing wavelength and peek-to-peek amplitude. Auditory features were extracted according to frequency division and masking properties of human auditory system. Wavelet features contained information of entropy of zero-crossing wavelength distribution density of all levels of wavelet signals and low frequency envelope of wavelet decomposition.

3.3. Network Pruning Experiments

To illustrate relationship between the pruning threshold and feature dimension, NMI of features in each group was arranged in descending order. Feature size versus threshold of each layer is shown in Figure 5 . Horizontal line in each figure is average NMI, which can be used to remove the uninformative features effectively. The numbers of hidden units in different groups are more balanced after network pruning with this threshold.
In order to discuss the influence of the pruning threshold on classification accuracy, classification accuracy of SVM under each pruning threshold was calculated, results are shown in Figure 5e. The vertical lines represent the pruning threshold of each layer. Classification accuracy increases with the increase of threshold at the beginning, because irrelevant features are removed. However, the accuracy begins to decrease when the threshold is too high. Average NMI was used as the threshold in the following experiments.

3.4. Clustering Experiments

The goal of clustering experiments was to test whether the competitive learning and network pruning can improve discriminative performance. t-SNE [29] feature visualization method was used to observe the distribution of weight vectors in RBM, CRBM and compressed CRBM. The perplexity of the Gaussian kernel in t-SNE is 30. Hidden units of layer1 were divided into two groups according to categories. The scatter diagram of the grouped weight vectors of RBM viewed by t-SNE is shown in Figure 6a. The similar results of CRBM and compressed CRBM are shown in Figure 6b,c, respectively. Both of them are more distributed than that in Figure 6a. The results indicate that CRBM can learn the differences of categories. The compressed CRBM could represent the ship radiated noise by using fewer features than RBM and CRBM.
The distributions of samples described by DCN features, compressed DCN features and traditional features were observed by t-SNE. In total, 150 samples of each category were selected randomly to draw the scatter diagram. Figure 7 shows the comparison of these features. It is obvious that both DCN features and compressed DCN features produce a better distribution than traditional features. The compressed DCN could get a similar result to DCN with fewer features.

3.5. Features Evaluation

NMI of DCN features, compressed DCN features and traditional features were compared. NMI of each feature and average NMI of each feature set are shown in Figure 8. Features learned by DCN and compressed DCN have higher NMI than traditional features. The average NMI of pruned Layer4 is 0.71, which outperforms other features. There are plenty of DCN features with low NMI, which will have negative effect on the recognition accuracy. Compressed DCN can reduce the feature dimension significantly while reserving the informative features.

3.6. Classification Experiments

Classification performances of deep feed-forward neural network (NN), DBN, DCN and Compressed DCN with the same structure were compared. NN was trained with labeled data. DBN, DCN and compressed DCN were pre-trained in an unsupervised phase, followed by supervised fine-tuning. The accuracy and test time are shown in Table 1. Compared with NN, DBN with unsupervised pre-training can significantly improve the classification accuracy. Compared with DBN, competitive learning mechanism in DCN can help improve the classification accuracy. With the help of network pruning, compressed DCN achieved the highest accuracy, which is significantly higher than other methods. Moreover, the test time of compressed DCN is reduced due to the network pruning.
Classification results of DCN features, compressed DCN features and traditional features were compared. SVM classifiers were used to classify the two targets. Parameters of the classifiers were selected by using 10-fold cross validation, in which SVM was trained and tested 10 times repeatedly, with each of the subsamples(randomly partitioned into 10 equal sized) used exactly once as the validation data. Due to the randomness of the initial point of network parameters illustrated in Equation (8), well-trained networks can have varying classification accuracy in different trials depending on the choice of initialization. In order to verify the robustness of the network, average classification accuracy over 10 random trials was conducted to obtain a more stable and repeatable experimental results.
Assuming that the first class was positive and the second class was negative. ROC curves were constructed by SVM decision function scores obtained on testing data. Figure 9 shows the comparison of ROC curves obtained from DCN features, compressed DCN features and traditional features. The performances of DCN features and compressed DCN features are significantly better than traditional features. Compared with DCN, the compressed DCN has a better performance with fewer features, which indicate that the network pruning can improve the classification accuracy. As shown in Figure 9d, pruned Layer4 can achieve the highest normalized area under ROC curve (AUC). The performance of pruned Layer3 is the second best shown in Figure 9c.
Classification results of SVM are shown in Table 2. For DCN and compressed DCN, the classification accuracy gradually improved as the number of layers increased. The classification accuracy obtained from pruned Layer4 is 89.1 % , which is the highest accuracy in our experiments. The features of pruned Layer3 give an accuracy of 83.0 % , which perform the second best. Due to the network pruning, the test time can be effectively reduced with the decrease of the feature dimension and this provides conditions for real-time targets classification by sonar devices. Compared with results in Table 1, SVM classifier can achieve equivalent or even higher accuracy on the dataset.
Feature selection algorithm based on NMI was used for each feature set. Features in training dataset were sorted in descending order, and then they were selected incrementally one by one to make up the input feature subsets for SVM classifier, the corresponding features of testing dataset were selected for testing. Results are shown in Figure 10. By applying the feature selection algorithm, the highest accuracy obtained from DCN features is 84.6 % shown in Figure 10d and the highest accuracy obtained by traditional features is 78.8 % shown in Figure 10e. The pruned layer4 could achieve the highest accuracy of 89.1 % which is 5.3 % higher than layer4 shown in Figure 10d and 13.1 % higher than MFCC shown in Figure 10e. Accuracy obtained from DCN features begins to drop significantly when the feature dimension is too large. However, the accuracy obtained from compressed DCN features increases almost monotonously with the growth of feature dimension, which indicate that the compressed network has fewer redundant features. Compared with DCN, the compressed DCN can reduce feature dimension, while improving the classification accuracy.

4. Conclusions

A compressed deep competitive network is presented by integrating competitive learning into the restricted Boltzmann machine learning algorithm and pruning the network based on mutual information. Conclusions are summarized as follows: By applying our algorithm to the underwater acoustic targets recognition, compared with traditional features, the deep competitive network features are more relevant with labels. The deep competitive network can be greatly compressed by pruning redundant hidden units based on mutual information. Support vector machine trained on the features learned by the proposed model can achieve higher classification accuracy with fewer features. The high accuracy and efficiency facilitate the application of the algorithm to real time underwater acoustic targets recognition.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (Grant No. 51179157) and grant from Science and Technology on Underwater Acoustic Antagonizing Laboratory.

Author Contributions

Sheng Shen and Honghui Yang contributed to the idea of the incentive mechanism and responsible for the theoretical analysis. Meiping Sheng contributed to the structure and content. All of the authors were involved in writing the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Das, A.; Kumar, A.; Bahl, R. Marine vessel classification based on passive sonar data: The cepstrum-based approach. IET Radar Sonar Navig. 2013, 7, 87–93. [Google Scholar] [CrossRef]
  2. Zhang, L.; Wu, D.; Han, X.; Zhu, Z. Feature Extraction of Underwater Target Signal Using Mel Frequency Cepstrum Coefficients Based on Acoustic Vector Sensor. J. Sens. 2016, 2016, 1–11. [Google Scholar] [CrossRef]
  3. Santos-Domínguez, D.; Torres-Guijarro, S.; Cardenal-López, A.; Pena-Gimenez, A. ShipsEar: An underwater vessel noise database. Appl. Acoust. 2016, 113, 64–69. [Google Scholar] [CrossRef]
  4. Meng, Q.; Yang, S.; Piao, S. The classification of underwater acoustic target signals based on wave structure and support vector machine. J. Acoust. Soc. Am. 2014, 136, 2265. [Google Scholar] [CrossRef]
  5. Meng, Q.; Yang, S. A wave structure based method for recognition of marine acoustic target signals. J. Acoust. Soc. Am. 2015, 137, 2242. [Google Scholar] [CrossRef]
  6. Azimi-Sadjadi, M.R.; Yao, D.; Huang, Q.; Dobeck, G.J. Underwater target classification using wavelet packets and neural networks. IEEE Trans. Neural Netw. 2000, 11, 784. [Google Scholar] [CrossRef] [PubMed]
  7. Siddagangaiah, S.; Li, Y.; Guo, X.; Chen, X.; Zhang, Q.; Yang, K.; Yang, Y. A Complexity-Based Approach for the Detection of Weak Signals in Ocean Ambient Noise. Entropy 2016, 18, 101. [Google Scholar] [CrossRef]
  8. Yang, H.; Gan, A.; Chen, H.; Pan, Y. Underwater acoustic target recognition using SVM ensemble via weighted sample and feature selection. In Proceedings of the 2016 13th International Bhurban Conference on International Bhurban Conference on Applied Sciences and Technology, Islamabad, Pakistan, 12–16 January 2016; pp. 522–527. [Google Scholar]
  9. Filho, W.S.; Seixas, J.M.D.; Moura, N.N.D. Preprocessing passive sonar signals for neural classification. IET Radar Sonar Navig. 2011, 5, 605–612. [Google Scholar] [CrossRef]
  10. Wei, X. On feature extraction of ship radiated noise using 11/2 d spectrum and principal components analysis. In Proceedings of the IEEE International Conference on Signal Processing, Communications and Computing, Hong Kong, China, 5–8 August 2016; pp. 1–4. [Google Scholar]
  11. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  12. Kamal, S.; Mohammed, S.K.; Pillai, P.R.S.; Supriya, M.H. Deep learning architectures for underwater target recognition. In Proceedings of the Ocean Electronics, Kochi, India, 23–25 October 2014; pp. 48–54. [Google Scholar]
  13. Cao, X.; Zhang, X.; Yu, Y.; Niu, L. Deep learning-based recognition of underwater target. In Proceedings of the IEEE International Conference on Digital Signal Processing, Beijing, China, 16–18 October 2017; pp. 89–93. [Google Scholar]
  14. Yang, H.; Shen, S.; Yao, X.; Han, Z. Underwater Acoustic Target Feature Learning and Recognition using Hybrid Regularization Deep Belief Network. Xibei Gongye Daxue Xuebao/J. Northwest. Polytech. Univ. 2017, 35, 220–225. [Google Scholar]
  15. Shen, S.; Yang, H.; Han, Z.; Shi, J.; Xiong, J.; Zhang, X. Learning robust features from underwater ship-radiated noise with mutual information group sparse DBN. In Proceedings of the INTER-NOISE and NOISE-CON Congress and Conference Proceedings, Hamburg, Germany, 21–24 August 2016; pp. 4855–5847. [Google Scholar]
  16. Han, S.; Mao, H.; Dally, W.J. Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding. Fiber 2015, 56, 3–7. [Google Scholar]
  17. Hinton, G.E. A Practical Guide to Training Restricted Boltzmann Machines. Momentum 2012, 9, 599–619. [Google Scholar]
  18. Bengio, Y. Deep Learning of Representations for Unsupervised and Transfer Learning. Workshop Unsuperv. Transf. Learn. 2012, 7, 1–20. [Google Scholar]
  19. Xie, X.; Hahnloser, R.H.; Seung, H.S. Selectively grouping neurons in recurrent networks of lateral inhibition. Neural Comput. 2002, 14, 2627–2646. [Google Scholar] [CrossRef] [PubMed]
  20. Rumelhart, D.E.; Zipser, D. Feature Discovery by Competitive Learning; Ablex Publishing Corp.: New York, NY, USA, 1988; pp. 151–193. [Google Scholar]
  21. Erhan, D.; Bengio, Y.; Courville, A.; Vincent, P. Visualizing Higher-Layer Features of a Deep Network; University of Montreal: Montreal, QC, Canada, 2009. [Google Scholar]
  22. Yang, H.; Shen, S. The Feature Selection of Pattern Recognition; Publishing House of Electronics Industry: Beijing, China, 2016. [Google Scholar]
  23. Cover, T.M.; Thomas, J.A. Elements of Information Theory (Wiley Series in Telecommunications and Signal Processing); Wiley-Interscience: Cambridge, UK, 2017. [Google Scholar]
  24. Liu, H.; Hussain, F.; Tan, C.L.; Dash, M. Discretization: An Enabling Technique. Data Min. Knowl. Discov. 2002, 6, 393–423. [Google Scholar] [CrossRef]
  25. Peng, H.; Long, F.; Ding, C. Feature Selection Based on Mutual Information: Criteria of Max-Dependency, Max-Relevance, and Min-Redundancy. IEEE Comput. Soc. 2005, 27, 1226–1238. [Google Scholar]
  26. Raina, R.; Battle, A.; Lee, H. Self-taught Learning Transfer Learning from Unlabeled Data. In Proceedings of the 24th International Conference on Machine Learning, Corvalis, OR, USA, 20–24 June 2007; pp. 759–766. [Google Scholar]
  27. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. J. Mach. Learn. Res. 2010, 9, 249–256. [Google Scholar]
  28. Palm, R.B. Prediction as a Candidate for Learning Deep Hierarchical Models of Data; Technical University of Denmark: Kongens Lyngby, Denmark, 2012. [Google Scholar]
  29. Hinton, G.E. Visualizing High-Dimensional Data Using t-SNE. Vigiliae Christ. 2008, 9, 2579–2605. [Google Scholar]
  30. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27:1–27:27. [Google Scholar] [CrossRef]
  31. Yang, L.X.; Chen, K.A.; Zhang, B.R.; Liang, Y. Underwater acoustic target classification and auditory feature identification based on dissimilarity evaluation. Acta Phys. Sin. 2014, 63, 134304. [Google Scholar]
Figure 1. Compressed deep competitive network. The left part of the diagram shows the stacked compressed CRBM. White circles indicate the pruned hidden units. Then, the entire network is fine-tuned with class labels shown in the top of the diagram. The right part of the diagram represents the algorithm flow of compressed CRBM, where the competitive learning and the mutual information based network pruning are two core ideas of the model.
Figure 1. Compressed deep competitive network. The left part of the diagram shows the stacked compressed CRBM. White circles indicate the pruned hidden units. Then, the entire network is fine-tuned with class labels shown in the top of the diagram. The right part of the diagram represents the algorithm flow of compressed CRBM, where the competitive learning and the mutual information based network pruning are two core ideas of the model.
Entropy 20 00243 g001
Figure 2. Network pruning based on mutual information. Left part of the diagram represents estimating normalized mutual information (NMI) from learned features. Right part of the diagram is 1R discretization preprocessing step.
Figure 2. Network pruning based on mutual information. Left part of the diagram represents estimating normalized mutual information (NMI) from learned features. Right part of the diagram is 1R discretization preprocessing step.
Entropy 20 00243 g002
Figure 3. Spectrogram of ship radiated noise.
Figure 3. Spectrogram of ship radiated noise.
Entropy 20 00243 g003
Figure 4. Experiment Procedure.
Figure 4. Experiment Procedure.
Entropy 20 00243 g004
Figure 5. Feature size versus threshold value in each layer, the horizontal lines are the thresholds selected in each layer. (a) Layer1; (b) Layer2; (c) Layer3; (d) Layer4; (e) Accuracy versus the threshold.
Figure 5. Feature size versus threshold value in each layer, the horizontal lines are the thresholds selected in each layer. (a) Layer1; (b) Layer2; (c) Layer3; (d) Layer4; (e) Accuracy versus the threshold.
Entropy 20 00243 g005
Figure 6. Distribution of weights vectors viewed by t-SNE. (a) restricted Boltzmann machine (RBM); (b) competitive restricted Boltzmann machine (CRBM); (c) The proposed compressed CRBM.
Figure 6. Distribution of weights vectors viewed by t-SNE. (a) restricted Boltzmann machine (RBM); (b) competitive restricted Boltzmann machine (CRBM); (c) The proposed compressed CRBM.
Entropy 20 00243 g006
Figure 7. Training samples viewed by t-SNE. (a) Layer1 in deep competitive network (DCN); (b) Layer2 in DCN; (c) Layer3 in DCN; (d) Layer4 in DCN; (e) pruned Layer1 in the proposed model; (f) pruned Layer2 in the proposed model; (g) pruned Layer3 in the proposed model; (h) pruned Layer4 in the proposed model; (i) Mel-frequency cepstral coefficients (MFCC), differential MFCC (DMFCC) and second-order differential MFCC (DDMFCC); (j) waveform features; (k) auditory features; (l) wavelet features.
Figure 7. Training samples viewed by t-SNE. (a) Layer1 in deep competitive network (DCN); (b) Layer2 in DCN; (c) Layer3 in DCN; (d) Layer4 in DCN; (e) pruned Layer1 in the proposed model; (f) pruned Layer2 in the proposed model; (g) pruned Layer3 in the proposed model; (h) pruned Layer4 in the proposed model; (i) Mel-frequency cepstral coefficients (MFCC), differential MFCC (DMFCC) and second-order differential MFCC (DDMFCC); (j) waveform features; (k) auditory features; (l) wavelet features.
Entropy 20 00243 g007
Figure 8. Comparison of each feature via normalized mutual information.
Figure 8. Comparison of each feature via normalized mutual information.
Entropy 20 00243 g008
Figure 9. Receivers operating characteristic (ROC) curves of each feature set. (a) Layer1 in the proposed model; (b) Layer2 in the proposed model; (c) Layer3 in the proposed model; (d) Layer4 in the proposed model; (e) MFCC, DMFCC, DDMFCC; (f) Auditory, Waveform, Wavelet.
Figure 9. Receivers operating characteristic (ROC) curves of each feature set. (a) Layer1 in the proposed model; (b) Layer2 in the proposed model; (c) Layer3 in the proposed model; (d) Layer4 in the proposed model; (e) MFCC, DMFCC, DDMFCC; (f) Auditory, Waveform, Wavelet.
Entropy 20 00243 g009
Figure 10. Feature selection results. (a) Layer1 in the proposed model; (b) Layer2 in the proposed model; (c) Layer3 in the proposed model; (d) Layer4 in the proposed model; (e) MFCC, DMFCC, DDMFCC, (f) Auditory, Waveform, Wavelet.
Figure 10. Feature selection results. (a) Layer1 in the proposed model; (b) Layer2 in the proposed model; (c) Layer3 in the proposed model; (d) Layer4 in the proposed model; (e) MFCC, DMFCC, DDMFCC, (f) Auditory, Waveform, Wavelet.
Entropy 20 00243 g010
Table 1. Classification results of the Neural Network.
Table 1. Classification results of the Neural Network.
MethodsAccuracy/%Test Time/s
NN75.50.1657
DBN81.50.1624
DCN82.20.1632
Compressed DCN88.60.0974
Table 2. Classification results of support vector machine (SVM).
Table 2. Classification results of support vector machine (SVM).
MethodsFeaturesDimensionAccuracy/%Variance/ × 10 3 Test Time/s
TraditionalMFCC [2]1278.74.90.0139
DMFCC [2]1271.35.30.0135
DDMFCC [2]1270.85.60.0135
Waveform [4,5]872.49.10.0552
Auditory [31]2478.47.20.1071
Wavelet [6]1473.87.40.0614
DCNLayer150071.24.43.8241
Layer250074.64.53.8250
Layer35079.43.90.1667
Layer45082.43.70.1672
Compressed DCNPruned Layer116375.44.60.8067
Pruned Layer215877.84.40.7213
Pruned Layer33483.03.80.1498
Pruned Layer43189.13.40.1423

Share and Cite

MDPI and ACS Style

Shen, S.; Yang, H.; Sheng, M. Compression of a Deep Competitive Network Based on Mutual Information for Underwater Acoustic Targets Recognition. Entropy 2018, 20, 243. https://doi.org/10.3390/e20040243

AMA Style

Shen S, Yang H, Sheng M. Compression of a Deep Competitive Network Based on Mutual Information for Underwater Acoustic Targets Recognition. Entropy. 2018; 20(4):243. https://doi.org/10.3390/e20040243

Chicago/Turabian Style

Shen, Sheng, Honghui Yang, and Meiping Sheng. 2018. "Compression of a Deep Competitive Network Based on Mutual Information for Underwater Acoustic Targets Recognition" Entropy 20, no. 4: 243. https://doi.org/10.3390/e20040243

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop