Next Article in Journal
Tidal Exclusion Barriers Fragment an Invertebrate Community into Taxonomically and Functionally Distinct Estuarine and Wetland Assemblages
Previous Article in Journal
Underwater Target Bearing Estimation Performance of Bottom-Mounted Extended Coprime Sparse Array
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel ViT Model with Wavelet Convolution and SLAttention Modules for Underwater Acoustic Target Recognition

Ocean College, Jiangsu University of Science and Technology, Zhenjiang 212003, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2025, 13(4), 634; https://doi.org/10.3390/jmse13040634
Submission received: 23 February 2025 / Revised: 10 March 2025 / Accepted: 21 March 2025 / Published: 22 March 2025

Abstract

:
Underwater acoustic target recognition (UATR) technology plays a significant role in marine exploration, resource development, and national defense security. To address the limitations of existing methods in computational efficiency and recognition performance, this paper proposes an improved WS-ViT model based on Vision Transformers (ViTs). By introducing the Wavelet Transform Convolution (WTConv) module and the Simplified Linear Attention (SLAttention) module, WS-ViT can effectively extract spatiotemporal complex features, enhance classification accuracy, and significantly reduce computational costs. The model is validated using the ShipsEar dataset, and the results demonstrate that WS-ViT significantly outperforms ResNet18, VGG16, and the classical ViT model in classification accuracy, with improvements of 7.3%, 4.9%, and 2.1%, respectively. Additionally, its training efficiency is improved by 28.4% compared to ViT. This study demonstrates that WS-ViT not only enhances UATR performance but also maintains computational efficiency, providing an innovative solution for efficient and accurate underwater acoustic signal processing.

1. Introduction

Underwater acoustic target recognition (UATR) technology plays a vital role in fields such as marine exploration, resource development, and national defense security. With the increasing scale of marine activities, accurate identification and classification of underwater targets have become crucial for ensuring the safety and efficient utilization of underwater environments. Exploring advanced UATR methods contributes to the development of related fields, including seabed mapping, marine biodiversity monitoring, and ship target recognition [1]. Moreover, this technology is widely applied in scientific research areas such as marine behavioral analysis and underwater archeological investigations. As a result, UATR technology has become indispensable in modern underwater engineering and national defense, driving advancements in the safety and efficiency of marine activities. In the early stages, the identification of underwater acoustic targets relied solely on human effort, specifically using the human ear. To classify underwater acoustic signals more efficiently, researchers began leveraging the time–frequency characteristics of these signals, such as waveform structure [2], Mel-Frequency Cepstral Coefficients (MFCC) [3], Short-Time Fourier Transform (STFT) [4], and Wavelet Transform (WT) [5]. Each method offers distinct advantages depending on the signal characteristics: Fourier transform proves particularly effective for analyzing stable acoustic signals, such as the continuous hum of an engine, while Wavelet Transform demonstrates superior performance in processing dynamic and non-stationary signals, particularly in the analysis of propeller noise. In recent years, with the advancement of computer technology, methods based on deep learning (DL) for UATR have rapidly developed [6,7] and have been widely applied in fields such as natural language processing, speech recognition, and medical diagnosis [8,9]. For instance, Song et al. [7] utilized a CNN model to process one-dimensional time-series signals for underwater target recognition, while Xue et al. designed and employed a ResNet model with an attention mechanism to improve recognition accuracy [10]. Additionally, Kamal et al. [11] extracted features using a combination of CNN and Bi-LSTM models.
Since its introduction, the Transformer architecture [12] has garnered widespread attention. Feng et al. [13] were the first to apply a Transformer-based model for underwater acoustic target recognition, achieving promising results. Li et al. [14] proposed a spectrogram Transformer model (STM) for underwater acoustic target recognition, which effectively overcomes the limitations of traditional CNNs in capturing rich spectral features. Yao et al. [15] proposed Mobile_ViT, a hybrid model that fuses local features from MobileNet with global context from a Transformer for underwater acoustic target recognition, achieving high accuracy with reduced complexity. Tang et al. [16] proposed an underwater acoustic target recognition method based on a pre-trained Transformer, which effectively overcomes the limitations of traditional CNNs in capturing global features. Dong et al. [17] proposed a Transformer network based on a cross-attention mechanism (CAF-ViT) that achieves efficient underwater acoustic target recognition by fusing time-frequency features such as LOFAR, Mel spectrograms, and wavelet packets. Despite its powerful performance, the high computational cost remains a significant challenge. To address this, Finder et al. proposed a method that leverages Wavelet Transform (WT) to expand the network’s receptive field without excessive parameterization [18]. On the other hand, Guo et al. introduced a Simplified Linear Attention (SLA) module aimed at enhancing the efficiency of Transformer models while maintaining or even improving their performance [19]. Inspired by these advancements, we propose a novel approach applied to the Vision Transformer (ViT) model. Our goal is to develop a model that is both efficient and powerful, balancing computational efficiency and performance. The main contributions of this paper include the following:
(1) We propose a novel WS-ViT model by integrating Wavelet Transform Convolution (WTConv) and Simplified Linear Attention (SLAttention) into the Vision Transformer (ViT) architecture. The WTConv module effectively complements the signal features obtained from the Short-Time Fourier Transform by employing multi-scale decomposition to further extract local time-frequency details, while the SLAttention module reduces computational complexity.
(2) Extensive experiments on the ShipsEar dataset demonstrate the superiority of WS-ViT. It outperforms ResNet18, VGG16, and the classical ViT in classification accuracy, showcasing higher computational efficiency and strong generalization capabilities.

2. Theory and Models

2.1. Attention in Transformer Models

Image classification, as a core task in computer vision, has evolved significantly over the years. Traditional Convolutional Neural Networks (CNNs) have achieved remarkable success in image classification tasks, with the introduction of classic architectures such as AlexNet and VGG laying the foundation for deep learning in the field of image processing. However, as network depth and computational power increased, the limitations of CNNs became apparent, particularly in handling long-range dependencies and capturing global contextual information. Against this backdrop, Transformer-based vision models have gradually emerged as a promising alternative.
The core component of the Transformer model is the self-attention mechanism, which determines the representation (embedding) of each position in the input sequence by calculating the similarity between all positions. The self-attention mechanism computes attention weights for each position relative to others using three sets of vectors: Query (Q), Key (K), and Value (V). For each position in the input, we first map it to three distinct vectors through linear transformations: the Query vector, the Key vector, and the Value vector. The specific calculation formulas are as follows:
Q i = x i W q   K i = x i W k   V i = x i W v
where x i represents the vector representation of the i-th token in the input sequence, where i ranges from 1 to N, and N denotes the length of the input sequence. W q , W k , W v are weight matrices learned during training. Next, the similarity between the Query vector Q i and the Key vectors K j of all other positions is calculated, typically using the dot product to measure their similarity. The formula is as follows:
s c o r e ( Q i , K j ) = Q i K j T
where j ranges from 1 to N. To prevent excessively large values, the dot product is typically scaled, resulting in the Scaled Dot-Product Attention. The formula is as follows:
s c o r e ( Q i , K j ) = Q i K j T d k
where d k represents the dimensionality of the Key vectors, and the scaling factor d k helps prevent the dot product from becoming too large, which could lead to gradient vanishing or exploding problems. Then, the similarities of all Keys are normalized using the Softmax function, resulting in the attention weight for each Key being as follows:
W e i g h t s ( Q i , K j ) = S o f t m a x Q i K j T d k
The Softmax function is commonly used in multi-class classification problems to convert a vector of raw scores (logits) into a probability distribution. Specifically, for a given vector a = [ a 1 , a 2 , , a n ] , the Softmax function is defined as:
S o f t m a x ( a i ) = e a i j = 1 n e a j
where a i represents the i-th element of the input vector. Finally, the attention weights are used to perform a weighted sum of the Value vectors, resulting in the output being:
O u t p u t i = j = 1 N Weights ( Q i , K j ) V j
The values of output reflect the importance of features at different positions in the input sequence.
To enable the model to learn diverse representations from multiple subspaces, the Transformer introduces the Multi-Head Attention mechanism. The core idea of Multi-Head Attention is to project the Query, Key, and Value vectors into multiple distinct subspaces, compute attention separately in each subspace, concatenate the results, and then apply a linear transformation to produce the final output.

2.2. Vision Transformer

The Vision Transformer (ViT), proposed by Google Research in 2020 [20], is a Transformer-based model for image classification. The innovation of ViT lies in its departure from traditional Convolutional Neural Networks (CNNs). Instead, it treats an image as a sequence of patches and processes them as sequential data within the Transformer architecture. The core idea of the model is that spatial information in images can be captured through the self-attention mechanism, enabling efficient image classification. The structure of the model is illustrated in Figure 1.
The first step in ViT is to divide the input image into a set of fixed-size, non-overlapping patches. Suppose the input image has dimensions H × W × C, where H is the height, W is the width, and C is the number of channels. The image is divided into patches of size P × P, with each patch containing P × P × C pixels. Each patch is flattened and mapped to a vector through a linear transformation (typically a fully connected layer), where the dimensionality of the vector usually matches the input dimension of the Transformer model. In this way, the input image is transformed into a sequence of embeddings, with each patch corresponding to a token in the Transformer. Typically, the input image size for the ViT model is 224 × 224 × 3, and each patch size is 16 × 16. This results in 14 × 14 = 196 patches, with each patch having dimensions 16 × 16 × 3. Each patch is flattened into a vector, and after linear embedding, a sequence of size 196 × D is obtained, where D is the dimensionality of the embedding space. Since the Transformer architecture inherently lacks positional information, ViT introduces positional encoding to incorporate the spatial relationships between patches. Positional encoding allows the model to recognize the relative or absolute positions of patches within the image. ViT adopts a method similar to the original Transformer, using sine and cosine functions to generate positional encodings. These encodings are added element-wise to the patch embeddings, providing positional information for each patch. This enables ViT to capture spatial structures in the image through the self-attention mechanism, even without convolutional operations. The Transformer encoder is the core component of ViT. It processes the sequence of patch embeddings through multiple self-attention layers and feed-forward neural networks. For classification, a special classification token ([CLS] token) is typically introduced, serving as the global representation of the image. After processing by the Transformer encoder, the model extracts global features from the final representation of the [CLS] token. These global features are then passed to a classification head (usually a fully connected layer) for the final category prediction.

2.3. WS-ViT Model

The model proposed in this paper improves on the traditional ViT by integrating the WTConv and SLAttention modules into the ViT architecture for ship radiation noise recognition. This enhancement allows the model to better extract effective features when processing input data with high spatiotemporal complexity, thereby improving the accuracy of image classification. The main structure of the model is shown in Figure 2. First, the audio signal is converted into a time–frequency spectrogram using the Short-Time Fourier Transform (STFT), which is then fed into the network. The basic calculation formula is as follows:
X ( m , n ) = k = 0 L 1 x ( k ) g ( k m ) e j 2 π n k L
where x(k) represents the signal sequence, and g(km) is the window function. The ranges of m and n are determined by the signal length L and the window length N. Specifically, 0 ≤ m ≤ L − N and 0 ≤ n ≤ N − 1.
Before the image enters the Patch Embedding stage, it is processed through a Wavelet Transform Convolution (WTConv) layer to obtain enhanced features. The principle involves convolving the input image with decomposition filters to generate multi-scale feature components Y L L , Y L H , Y H L and Y H H . These components represent the low-frequency parts in both horizontal and vertical directions, the low-frequency part in the horizontal direction and the high-frequency part in the vertical direction, the high-frequency part in the horizontal direction and the low-frequency part in the vertical direction, and the high-frequency parts in both horizontal and vertical directions, respectively. After wavelet decomposition, each feature component is further convolved to produce enhanced features Y , which can be expressed as:
Y = W Y
where W is convolution kernel.
After decomposition and convolution, the inverse Wavelet Transform (IWT) is used to reconstruct the multi-scale features back to the original feature dimensions. The reconstruction can be expressed as:
X = Y G
where G is the wavelet reconstruction filter and X is the reconstructed signal. Finally, residual connections are used to retain the original input information on top of the wavelet convolution operation. The formula for the residual connection is:
Z = X + X
where X is original input image.
In the Transformer encoder part, the traditional attention mechanism (Attention) can be expressed as:
Q = X W Q ,   K = X W K ,   V = X W V , O i = j = 1 N Sim ( Q i , K j ) j Sim ( Q i , K j ) V j
where W Q , W K , W V C × C are the weight matrices, Sim ( , ) is the similarity function and N is the length of the input sequence. For traditional attention, the similarity function can be expressed as:
Sim softmax ( Q i , K j ) = exp Q i K j T C
where C represents the dimensionality of the Key vectors. The Sim softmax ( , ) represents using the softmax function to characterize the similarity between the Query and the Key.
The attention mechanism based on softmax leads to high computational complexity. In the SLA module, the similarity function modified with the ReLU function, along with the use of Depthwise Separable Convolution (DWC) technology, can improve computational efficiency. Its formula is expressed as:
Sim SLA ( Q i , K j ) = ReLU ( Q i ) ReLU ( K j ) T , O ˜ i = j = 1 N Sim SLA ( Q i , K j ) j Sim SLA ( Q i , K j ) V j , O SLA = O ˜ + DWC ( V )

3. Results and Discussion

3.1. Dataset and Preprocessing

The ship noise dataset used in this paper is the ShipsEar dataset [21]. This dataset was recorded between 2012 and 2013 from various regions along the Spanish coast. The original classification of the dataset divides 90 different ship noises and background noises into five categories, A, B, C, D, and E, with detailed classifications, shown in Table 1. The raw audio lengths vary, so normalization was applied to the dataset. The audio was sampled at a rate of 48,000 Hz, and silent segments were removed. The processed audio signals were then framed and labeled with a length of 1 s, resulting in labeled samples from categories A to E, with 1888, 1574, 4286, 2465, and 1147 samples per category, totaling 11,360 samples. Subsequently, for each category, 1000 samples were randomly selected and split into training and validation sets, and Test Set I, with a 6:2:2 ratio. An additional 1000 samples were randomly selected from the remaining data as Test Set II.

3.2. Training Parameters

Combined with the model method proposed in this paper, the training parameters used in the experiments are shown in Table 2.

3.3. Evaluation Metrics

Accuracy is one of the most commonly used metrics in classification models. It represents the ratio of correctly predicted samples to the total number of samples. The calculation formula is:
Accuracy = T P + T N T P + T N + F P + F N
where True Positive (TP) represents the number of samples that are actually positive and correctly predicted as positive by the model. True Negative (TN) denotes the number of samples that are actually negative and correctly predicted as negative by the model. False Positive (FP) refers to the number of samples that are actually negative but incorrectly predicted as positive by the model. False Negative (FN) indicates the number of samples that are actually positive, but incorrectly predicted as negative by the model. Accuracy provides an intuitive measure of the overall prediction correctness of the model and is widely applicable to various classification problems. It is particularly effective in scenarios where the class distribution is relatively balanced, as it can better reflect the model’s performance.
The confusion matrix is a visualization tool widely used in supervised learning for evaluating the performance of classification models. It is constructed based on the actual labels and predicted labels, effectively summarizing the model’s classification results and presenting them in a visual format.

3.4. Analysis of Experimental Results

By using Test Set I, the confusion matrix of the experimental results in this paper is shown in Figure 3. Subfigures (a), (b), (c), (d), and (e) in Figure 3 represent the recognition performance of categories A, B, C, D, and E, respectively.
From the confusion matrix, it can be observed that the WS-ViT model performs excellently in terms of recognition accuracy for each category. Specifically, the recognition accuracies for categories A, D, and E all reached 99%, while the recognition accuracies for categories B and C were 98%, respectively. The overall average recognition accuracy was 98.6%. The reason for these results might be that the ship noise features of categories B and C are relatively similar to those of the other categories, which made it somewhat difficult for the model to distinguish between them. Despite the slightly lower recognition accuracy for categories B and C compared to the other categories, the accuracy remains high, indicating that the model demonstrates good robustness when handling these categories.
To further validate the superiority of the WS-ViT model, this paper compares it with the ResNet18, VGG16, and classic ViT models. The comparative experiments for all models were conducted on Test Set II, which consists of 1000 randomly selected samples covering five categories. The comparison results are shown in Table 3. The WS-ViT model achieved a classification accuracy of 99.5%. on Test Set II, significantly outperforming the other models. Specifically, the WS-ViT model improved upon the ResNet18 model by 7.3%, the VGG16 model by 4.9%, the ResNet152model by2.7% and the classic ViT model by 2.1%. This result indicates that after incorporating the WTConv and SLAttention modules, the WS-ViT model not only maintains the superior performance of the ViT model in classification tasks, but also further enhances classification accuracy.
To evaluate the computational efficiency of the WS-ViT model, this paper compares the training time of the WS-ViT model with the classic ViT model. The classic ViT model has a training time of 194 s per epoch, while the WS-ViT model has a training time of only 139 s per epoch, reducing the training time by 28.4%. The experiments further compared the floating-point operations (FLOPs) of the two models: the classic ViT model requires 11.29 G FLOPs, while the WS-ViT model reduces this to 9.03 G FLOPs, representing a 20% decrease. The experimental results are shown in Table 4. This optimization is due to the introduction of the WTConv and SLAttention modules in the WS-ViT model, which effectively reduce computational redundancy and decrease the reliance on global attention computations, thus significantly improving computational efficiency. Specifically, the WTConv module performs multi-scale feature extraction on the input image through a Wavelet Transform Convolution layer, and the SLAttention module further reduces computation by simplifying the linear attention mechanism, thereby enhancing the model’s training efficiency. As a result, the WS-ViT model not only maintains high classification accuracy but also significantly improves training efficiency, providing strong support for large-scale data processing in practical applications.
To further validate the generalization ability of the WS-ViT model, experiments were conducted using Test Set II. The confusion matrix for Test Set II is shown in Figure 4. From the confusion matrix, it can be seen that the WS-ViT model performs excellently on Test Set II as well. Specifically, the recognition accuracies for categories A, C, and E all reached 100%, while the recognition accuracies for categories B and D were 98.93% and 98.51%, respectively. This result indicates that the WS-ViT model not only performs outstandingly on the training and validation sets but also demonstrates strong generalization ability on unseen test data.
Based on the above experimental results, the WS-ViT model performs excellently in the underwater acoustic target recognition task. It not only significantly outperforms other comparative models in classification accuracy, but also achieves a notable improvement in training efficiency. Specifically, the WS-ViT model effectively extracts the spatiotemporal complex features of underwater acoustic signals by introducing the WTConv and SLAttention modules, thereby enhancing the model’s classification performance. At the same time, the WTConv module performs multi-scale feature extraction on the input image using Wavelet Transform Convolution layers, reducing the model’s reliance on global attention computations, which in turn lowers computational complexity. The SLAttention module simplifies the linear attention mechanism, further reducing computation burden and improving the model’s training efficiency. Furthermore, the performance of the WS-ViT model on Test Set II further validates its strong generalization ability. The model still maintains a high classification accuracy on unseen test data, demonstrating its broad potential for practical applications. In particular, in areas such as marine exploration, resource development, and national defense security, the WS-ViT model provides an innovative solution for efficient and precise underwater acoustic signal processing.

4. Conclusions

This paper proposes an improved WS-ViT model to meet the efficiency and accuracy requirements for underwater acoustic target recognition (UATR). By integrating the WTConv and SLAttention modules into the ViT architecture, the WS-ViT model achieves optimization in both computational efficiency and classification performance. Experimental results show that the WS-ViT model outperforms ResNet18, VGG16, and the classic ViT model in terms of classification accuracy, with improvements of 7.3%, 4.9%, and 2.1%, respectively, while its training efficiency has improved by 28.4% compared to the classic ViT model. The research in this paper provides an efficient and powerful solution for underwater acoustic target recognition, with broad application potential and practical value. However, despite the promising performance of the proposed method on the existing dataset, several limitations remain. First, the dataset used in this study (e.g., ShipsEar) is limited in terms of diversity and scale, potentially failing to fully represent the complex acoustic environments encountered in real-world scenarios. Second, the absence of recently published public datasets has hindered further validation and comprehensive performance evaluations of underwater acoustic target recognition algorithms. Future work should prioritize the development of more diverse and representative datasets to enable more accurate and reliable assessment of algorithm performance in practical applications. Additionally, we will enhance inference speed to meet real-time requirements through lightweight network architecture design, model compression via knowledge distillation, integrated with multi-threaded asynchronous processing and hardware acceleration.

Author Contributions

Conceptualization, H.G., B.W., T.F. and B.L.; funding acquisition, B.W.; experiments, H.G., B.W., T.F. and B.L.; method, H.G., B.W., T.F. and B.L.; data analysis, H.G., B.W., T.F. and B.L.; writing, H.G., B.W., T.F. and B.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Postgraduate Research & Practice Innovation Program of Jiangsu Province, grant number KYCX24_4181.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yuan, F.; Ke, X.; Cheng, E. Joint Representation and Recognition for Ship-Radiated Noise Based on Multimodal Deep Learning. J. Mar. Sci. Eng. 2019, 7, 380. [Google Scholar] [CrossRef]
  2. Meng, Q.; Yang, S. A Wave Structure Based Method for Recognition of Marine Acoustic Target Signals. J. Acoust. Soc. Amer. 2015, 137, 2242. [Google Scholar] [CrossRef]
  3. Zhang, L.; Wu, D.; Han, X.; Zhu, Z. Feature Extraction of Underwater Target Signal Using Mel Frequency Cepstrum Coefficients Based on Acoustic Vector Sensor. J. Sensors 2016, 2016, 7864213. [Google Scholar] [CrossRef]
  4. Zhang, Q.; Da, L.; Zhang, Y.; Hu, Y. Integrated Neural Networks Based on Feature Fusion for Underwater Target Recognition. Appl. Acoust. 2021, 182, 108261. [Google Scholar]
  5. Li, H.; Cheng, Y.; Dai, W.; Li, Z. A Method Based on Wavelet Packets-Fractal and SVM for Underwater Acoustic Signals Recognition. In Proceedings of the 2014 12th International Conference on Signal Processing (ICSP), Hangzhou, China, 19–23 October 2014. [Google Scholar]
  6. Hu, G.; Wang, K.; Peng, Y.; Qiu, M.; Shi, J.; Liu, L. Deep Learning Methods for Underwater Target Feature Extraction and Recognition. Comput. Intell. Neurosci. 2018, 2018, 1214301. [Google Scholar] [CrossRef] [PubMed]
  7. Xiaoping, S.; Jinsheng, C.; Yuan, G. A New Deep Learning Method for Underwater Target Recognition Based on One-Dimensional Time-Domain Signals. In Proceedings of the 2021 OES China Ocean Acoustics (COA), Harbin, China, 14–17 July 2021; pp. 1048–1051. [Google Scholar]
  8. Albawi, S.; Mohammed, T.A.; Al-Zawi, S. Understanding of a Convolutional Neural Network. In Proceedings of the 2017 International Conference on Engineering and Technology (ICET), Antalya, Turkey, 21–23 August 2017. [Google Scholar]
  9. Hershey, S.; Chaudhuri, S.; Ellis, D.P.W.; Gemmeke, J.F.; Jansen, A.; Moore, R.C.; Plakal, M.; Platt, D.; Saurous, R.A.; Seybold, B.; et al. CNN Architectures for Large-Scale Audio Classification. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017. [Google Scholar]
  10. Xue, L.; Zeng, X.; Jin, A. A Novel Deep-Learning Method with Channel Attention Mechanism for Underwater Target Recognition. Sensors 2022, 22, 5492. [Google Scholar] [CrossRef] [PubMed]
  11. Kamal, S.; Satheesh Chandran, C.; Supriya, M.H. Passive Sonar Automated Target Classifier for Shallow Waters Using End-to-End Learnable Deep Convolutional LSTMs. Eng. Sci. Technol. Int. J. 2021, 24, 860–871. [Google Scholar]
  12. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  13. Feng, S.; Zhu, X. A Transformer-Based Deep Learning Network for Underwater Acoustic Target Recognition. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1505805. [Google Scholar] [CrossRef]
  14. Li, P.; Wu, J.; Wang, Y.; Lan, Q.; Xiao, W. STM: Spectrogram Transformer Model for Underwater Acoustic Target Recognition. J. Mar. Sci. Eng. 2022, 10, 1428. [Google Scholar] [CrossRef]
  15. Yao, H.; Gao, T.; Wang, Y.; Wang, H.; Chen, X. Mobile_ViT: Underwater Acoustic Target Recognition Method Based on Local–Global Feature Fusion. J. Mar. Sci. Eng. 2024, 12, 589. [Google Scholar] [CrossRef]
  16. Tang, J.; Ma, E.; Qu, Y.; Gao, W.; Zhang, Y.; Gan, L. UAPT: An Underwater Acoustic Target Recognition Method Based on Pre-Trained Transformer. Multimed. Syst. 2025, 31, 50. [Google Scholar] [CrossRef]
  17. Dong, W.; Fu, J.; Zou, N.; Zhao, C.; Miao, Y.; Shen, Z. CAF-ViT: A Cross-Attention Based Transformer Network for Underwater Acoustic Target Recognition. Ocean Eng. 2025, 318, 120049. [Google Scholar] [CrossRef]
  18. Finder, S.E.; Amoyal, R.; Treister, E.; Freifeld, O. Wavelet Convolutions for Large Receptive Fields. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; Springer Nature: Cham, Switzerland, 2024; pp. 363–380. [Google Scholar]
  19. Guo, J.; Chen, X.; Tang, Y.; Wang, Y. SLAB: Efficient Transformers with Simplified Linear Attention and Progressive Re-parameterized Batch Normalization. arXiv 2024, arXiv:2405.11582. [Google Scholar]
  20. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  21. Santos-Domínguez, D.; Torres-Guijarro, S.; Cardenal-López, A.; Pena-Gimenez, A. ShipsEar: An Underwater Vessel Noise Database. Appl. Acoust. 2016, 113, 64–69. [Google Scholar] [CrossRef]
Figure 1. The Main Structure of the ViT Model.
Figure 1. The Main Structure of the ViT Model.
Jmse 13 00634 g001
Figure 2. Main Architecture of the WS-ViT Model.
Figure 2. Main Architecture of the WS-ViT Model.
Jmse 13 00634 g002
Figure 3. Confusion matrix for Recognition Results of Each Category.
Figure 3. Confusion matrix for Recognition Results of Each Category.
Jmse 13 00634 g003
Figure 4. Classification Results of Test Set II Under the Proposed Method.
Figure 4. Classification Results of Test Set II Under the Proposed Method.
Jmse 13 00634 g004
Table 1. Detailed Classification of the ShipsEar Dataset.
Table 1. Detailed Classification of the ShipsEar Dataset.
ClassTypes of ShipsSamples
Afishing boats, trawlers, mussel boats, tugboats and dredgers1888
Bmotorboats, pilot boats and sail boats1574
Cpassenger ferries4268
Docean liners and ro-ro vessels2465
Ebackground noise1147
Table 2. Training parameters.
Table 2. Training parameters.
Training ParametersValues/Setting
batch size8
training epochs200
initial learning rate0.001
optimization algorithmSDG
learning ratecosine annealing
loss functioncross-entropy
Table 3. Classification results of different networks.
Table 3. Classification results of different networks.
ModelAccuracy
ResNet1892.2%
VGG1694.6%
ResNet15296.8%
ViT97.4%
WS-ViT99.5%
Table 4. Comparison of training time and FLOPs.
Table 4. Comparison of training time and FLOPs.
ModelTraining Time (Seconds per Epoch)FLOPs (G)
ViT19411.29
WS-ViT1399.03
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, H.; Wang, B.; Fang, T.; Liu, B. A Novel ViT Model with Wavelet Convolution and SLAttention Modules for Underwater Acoustic Target Recognition. J. Mar. Sci. Eng. 2025, 13, 634. https://doi.org/10.3390/jmse13040634

AMA Style

Guo H, Wang B, Fang T, Liu B. A Novel ViT Model with Wavelet Convolution and SLAttention Modules for Underwater Acoustic Target Recognition. Journal of Marine Science and Engineering. 2025; 13(4):634. https://doi.org/10.3390/jmse13040634

Chicago/Turabian Style

Guo, Haoran, Biao Wang, Tao Fang, and Biao Liu. 2025. "A Novel ViT Model with Wavelet Convolution and SLAttention Modules for Underwater Acoustic Target Recognition" Journal of Marine Science and Engineering 13, no. 4: 634. https://doi.org/10.3390/jmse13040634

APA Style

Guo, H., Wang, B., Fang, T., & Liu, B. (2025). A Novel ViT Model with Wavelet Convolution and SLAttention Modules for Underwater Acoustic Target Recognition. Journal of Marine Science and Engineering, 13(4), 634. https://doi.org/10.3390/jmse13040634

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop