Next Article in Journal
Trajectory Tracking Control of an Autonomous Underwater Vehicle Under Disturbance and Model Uncertainty
Previous Article in Journal
Cyclostratigraphic Analysis and Depositional Environment Evolution of the Third Member of Eocene Shahejie Formation in the Laizhou Bay Sag, Southern Bohai Bay
Previous Article in Special Issue
A Channel-Aware AUV-Aided Data Collection Scheme Based on Deep Reinforcement Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning to Equalize for Single-Carrier Underwater Acoustic Communications

1
School of Low-Altitude Equipment and Intelligent Control, Guangzhou Maritime University, Guangzhou 510725, China
2
The Key Laboratory of Marine Environmental Survey Technology and Application, Ministry of Natural Resources, Guangzhou 510300, China
3
School of Electronic and Information Engineering, South China University of Technology, Guangzhou 510641, China
4
School of Artificial Intelligence, Guangzhou Maritime University, Guangzhou 510725, China
*
Authors to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2025, 13(11), 2209; https://doi.org/10.3390/jmse13112209
Submission received: 22 October 2025 / Revised: 10 November 2025 / Accepted: 17 November 2025 / Published: 20 November 2025

Abstract

Learning-based equalizers for multicarrier communication systems have been widely studied over underwater acoustic (UWA) channels. In this article, a learning-based equalizer is utilized for single-carrier (SC) underwater acoustic communications. A comprehensive comparison is made between existing deep learning (DL)-based approaches and a classical equalizer designed with adaptive filtering principles. It motivates the design of equalization for SC communications over underwater acoustic channels. To overcome distortion over the UWA channel, we propose a sliding deep learning-based equalizer that uses a sliding nonlinear network for equalization rather than a single-layer linear method. Moreover, to accelerate convergence during training, we proposed a preprocessing-based training phase. To mitigate the impact of time-varying channels, we additionally propose a meta-learning-enhanced adaptive filter algorithm for online adaptive equalization, named Meta-DNN. Based on the proposed DL equalizer, we leverage the pilot and data relationship to perform online transfer to achieve better BER performance. Moreover, to make this work more convincing, we test bit-error-rate (BER) performance across reproducible, realistic multi-scenario channels.

1. Introduction

Underwater communications are well needed for underwater information transmission, such as subsea facilities, test areas for autonomous vessels, underwater Internet of Things, and marine observatories [1,2]. In recent decades, numerous approaches using various media have been introduced to facilitate information interaction [3]. So far, light, magnidle, radio, and acoustic all aim to achieve high-rate, stable, and long-distance underwater communications [4,5]. Due to the nature of water, only the underwater acoustic (UWA) method can effectively achieve medium- and long-distance communication. Unfortunately, UWA communications face many challenges caused by physical factors, such as Doppler shifts, multi-path effects, and propagation loss [6].
In the literature, many methods have been proposed to reduce the distortion caused by UWA channels. Single-carrier (SC) and multi-carrier (MC) communication are two classical communication approaches [7]. In particular, orthogonal frequency division multiplexing (OFDM) is a typical multi-carrier modulation technique that is already widely applied to UWA communications [8]. OFDM can easily achieve frequency-domain equalization to combat multi-path interference and inter-symbol interference (ISI). However, UWA communications is a wideband system, and the small ratio of the carrier frequency to the signal bandwidth makes UWA OFDM more sensitive to Doppler shift. Meanwhile, the high peak-to-average ratio (PAPR) also poses problems for OFDM in long-distance transmission [9]. From this perspective, SC communications receive more attention because of their higher tolerance to Doppler shift than MC modulation over UWA channels and their lower PAPR. Generally, the equalization of SC modulation can be achieved in the time domain and/or frequency domain [10]. The recovery procedure can be performed in the time domain, especially under non-stationary conditions caused by platform and boundary movement. Meanwhile, the recovery procedure can also be performed in the frequency domain, and the complexity of frequency-domain equalization is relatively low. It is a pity that the classical adaptation-based equalizer usually uses a simple structure for reducing ISI, such as least mean square (LMS) [11] or least symbol error rate [12]. In addition, communication over the time-varying channel is a quite a prominent topic. In [13], a multi-scale time-varying multi-path amplitude model is proposed by using singular spectrum analysis, and a time-varying impulse response simulation framework is developed. In [14], an adaptive channel prediction scheme that extrapolates the channel knowledge estimated from a block of training symbols is proposed, and the predicted channel is used to decode consecutive data blocks. In [15], the received data are divided into subblocks, and the channel estimation of each subblock is regarded as a task. Meanwhile, a factor graph is proposed for multi-task channel estimation of subblocks to overcome the UWA time-varying channel.
Recently, deep learning has emerged as a novel approach for equalization, especially in challenging communication environments, e.g., molecular and UWA communications, whose communication channels are difficult to model accurately [16]. Many DL-based methods have been applied to UWA communications [7,17]. In [18], a hybrid architecture combining a convolutional neural network with a multi-layer perceptron was adopted, and a skip connection mechanism was introduced, leading to an effective reduction in the bit error rate (BER). To address data mismatch issues, the meta-learning algorithm was introduced in [19], which effectively improved the generalization of the DL-based receiver. In [20], an OFDM integrated receiver based on multi-task learning was proposed, improving the generalization of DL receivers through shared subtask parameters and a designed multi-task regularization loss function. In [21], inter-carrier interference was effectively suppressed through the use of autoencoder-based feature extraction and a sliding convolutional kernel structure, thereby reducing computational complexity while maintaining performance. In [22], downlink UWA communication using a deep neural network (DNN) utilizing a one-dimensional convolution neural network (CNN) is proposed, and the DNN-based DL NOMA UWA receiver outperformed the classical successive interference cancellation receiver. In [23], a DL-based receiver for single-carrier communication is proposed to mitigate time-varying UWA channels. In [24], a receiver system that explores the machine learning technique of a deep belief network is designed to combat distortion caused by the Doppler effect and multi-path propagation. There is little research on deep learning-based time-domain equalization methods for UWA single-carrier communication. The changing conditions of the UWA channel require retraining data-driven models from scratch for each new dataset, which poses a key limitation. Based on an online training mechanism, meta-learning, also known as learning to learn [25], is increasingly applied to UWA domains such as UWA OFDM, including channel estimation and detection, and it has become an effective solution [19,26]. In addition, some underwater communications systems are furnished for in-situ applications that operate in relatively stable locations. The parameters of in-situ equipment communication scenarios, including depth, distance, temperature, salinity, and even working hours, are either fixed or change periodically. Meanwhile, many direct and stochastic channel-replay tools have been found effective for validating the performance of UWA communications. Therefore, how to achieve better communication through known channels also becomes an interesting subject.
Motivated by the above reasons, we aim to apply deep learning methods to improve BER performance in underwater SC communications. The main contribution can be concluded as follows:
  • To overcome the distortion of the UWA channel, we proposed a sliding deep learning-based framework for SC communications that uses a multi-layer neural network with ReLU nonlinearity to eliminate channel distortion rather than a single-layer linear method. The sliding framework accounts for the time-varying characteristics of the UWA channel, and multi-layer, non-linear neural networks can improve equalization performance. Additionally, to accelerate training convergence, we used a pre-processing-based training phase.
  • Considering the impact of time-varying channels, we leverage the pilot and data symbol relationship to perform online transfer learning. A meta-learning-based SC equalizer is proposed in the time domain. In this process, the pilot is treated as labeled data, and the data symbol is regarded as the unlabeled data. A few-step learning procedure is performed, and the DL network is updated using the pilot. Then, the data symbol is equalized by the updated network. To make this article more convincing, the real-world UWA channels collected in different experiments are used to corroborate the effectiveness of the proposed algorithm.
The rest of this paper is divided into the following sections: Section 2 presents the structure of the single-carrier communication system. In Section 3, the learning-to-equalize approach is explained. Then, we present experimental settings and results in Section 4 and provide the conclusion in Section 5.
Notation: Column vectors and matrices are in the form of lowercase and capital bold letters, respectively. R indicates the set of real numbers. | · | represents the cardinality of a set. · stands for the Frobenius norm. R [ · ] means the real part operator. ∇ denotes the gradient operator.

2. System Model

The structure of the single-carrier communication system with N symbols is shown in Figure 1. At the transmitter, the transmitted bitstream, binary bit sequence { b 1 , , b i , , b M K ,   b i 0 , 1 } , is split into K groups with i = 1 , , M K , and each group is mapped to one of the symbols of the 2 M -QAM alphabet A . The K generated symbols x 1 , x 2 , , x K are framed with pilots, then fed to a pulse-shaping filter to give the time-domain signal. Then, after the addition of pilots, the signal is transmitted over a time-varying channel.
At the receiver, the received signal can be expressed as
y = H x + w ,
where y is a vector which contains [ y 1 , y t , , y K ] T , x is a vevtor which represents [ x 1 , x 1 , , x K ] T , and w denotes the additive white Guassian noise. H is a channel response matrix, which can be expressed as
H = h [ 1 , 1 ] h [ 1 , L 1 ] h [ 3 , 1 ] h [ 3 , L 1 ] h [ t , 1 ] h [ t , L 1 ] .
For UWA channels, we have two specific conditions. One is the time-varying (TV) channel, which is a tough case. TV channels will change within a data block. Each row in the channel matrix H has a different value, meaning the channel h [ t ] = [ h [ t , 1 ] , , h [ t , l ] ] is different at each time step. That is, each row in the matrix is distinct yet shares related properties. Another is the quasi-static (QS) channel, which is a mild status. QS channels are relatively fixed within a frame, i.e., time-invariant within a data block. In this condition, we assume that the CIR h [ t ] of each time slot t is the same. That is, each row in the matrix is the same.

3. Learning to Equalize

For UWA channels, channel parameters vary with the communication situation and are difficult to estimate due to their time-varying characteristics. The relationship between the transmitted symbols and the observed symbols can be described by
P c h { y 1 , , y k , , y K | x 1 , , x k , , x K ; H } ,
where x k means the transmitted symbol and y k represents the received symbol.
For convenience, we provide a summary of learning to equilization. The previous symbols can be used to equalize the received signal. For example, the symbol x k can be estimated by using the sequence { y k , y k 1 , , y 1 } , despite the ISI caused by channel multi-path. The view of learning can be divided into intra-frame and inter-frame learning methods. A comparison of the intra-frame learning and the inter-frame learning is shown in Figure 2.

3.1. Intra-Frame Learning-Based Adaptive Equalizer

Intra-frame learning methods use only the pilots contained in a communication round to train the filter weight w. Only the online training process is available throughout the entire learning stage. The most typical algorithm is the classical adaptive filter method. The classical adaptive equalization methods are in-frame learning approaches. The specific process of the intra-frame learning is as follows: assume X = { x 1 , , x k , , x K } is the data sequence sent from transmitter, with K modulated symbols, and Y = { y 1 , , y k , , y K } is the corresponding observed data at the reiceiver. A pair of training data and label can then be noted as { Y k m , k , X k } with Y k m , k = { y k m , , y k 1 , y k } , X k = { x k } , where k means the k-th symbol of the transmitted sequence and m represents the number of observation symbols for the k-th symbol, also known as the filter length. The sliding architecture is shown in Figure 3.
In this class of algorithms, only a simple network is considered. The simple framework can be easily trained on sequences and has a strong mathematical basis. The drawback is that the performance is unsatisfactory for tough channel conditions. The most famous adaptive equalization method is the LMS algorithm, which has a nonlinear feedback architecture. The LMS algorithm can be divided into two steps. The first step is the forward part, which equalizes the received signal and estimates the error. The second step is the adaptation process, which updates the current equalizer value based on the error. The forward architecture is a winner filter. The linear scheme makes it easy to analyze using mathematical methods. The winner filter can be represented in discrete form, and the input signal can be formulated as a discrete-time sequence. The output is guaranteed as an equalized symbol. According to optimization theory, the LMS algorithm updates the filter coefficients at each time step using the steepest descent method.

3.2. Inter-Frame Learning-Based Deep Learning Equalizer

Inter-frame learning methods contain both online and offline training phases that utilize this information. The inter-frame learning viewpoint is consistent with deep learning methods. Based on the LMS methods, we introduce a fully connected (FC) network-based framework, called FC-DNN, to equalize the SC signal in a sliding manner. We denote the FC-DNN equalizer as f ( θ ) , where θ R n are the parameters of the neural network. Define y as the network input and x as the recovered symbols, respectively, where x l { 1 , 1 } . The input vector has dimensions 2 M × 1 , which concatenates the real and imaginary parts of the input signal y , where M is the filter length. The input is processed through a sequence of deterministic, layered transformations. The NN procedure is a cascade of linear transformations followed by element-wise non-linearities, progressively mapping the input to the desired output space using the network’s learned parameters.
Therefore, the DNN equalizer f ( θ ) can be expressed as
x ^ = f ( θ ) = f t a n h ( V ) ( f R e l u ( v ) ( f R e l u ( 1 ) ( y ) ) ) ,
where V is the number of layers, and the v-th layer contains a total of N v neurons, each connected to all neurons in the v-th layer through the connection weight matrix. The parameters of the network are optimized during the offline learning phase.
The input of the first layer is the samples of y [ k ] , which are selectively chosen from the observed signals through pre-processing. This dataset is then used to train the FC-DNN equalizer, which classifies y [ k ] as s 1 or s 2 . The output of the hidden layers are activated by f R e l u ( z v ) = max ( 0 , z v ) , which is a non-linear function, to provide a normalized output and keep the output within [ 0 , + ] , and z v represents the output of the v-th hidden layer. The width of the l-th layer is denoted d v . Note that w l is the parameter of the l-th layer. Therefore, we can use θ = { ( w 1 , b 1 ) , , ( w v , b v ) , , ( w V , b V ) } R to express the set of whole network hyper-parameters, where w v means the v-th layer weigth and b v represents the v-th layer bias with v { 1 , , V } . Due to the ReLU activation function, each neuron’s output can be represented by two limit states, denoted as f l ( z i ) { 0 , 1 } . The output layer consists of a single neuron that outputs the estimates of the binary bits to be detected.
The receiver can be trained by optimizing the loss function
L ( x k , x ^ k ; θ ) = 1 N k = 1 K | x k x ^ k | 2 ,
where x ^ k denotes the output of the basic DNN scheme.
By employing Formula (5), the FC-DNN equalizer f ( θ ) arrives at an optimum set of connection weights θ o p t which minimizes the optimal average cost function. Then, the well-trained DNN f ( θ o p t ) can be employed for signal equalization using Formula (4).

3.3. Comparison of the Complexity of Intra-Frame Learning and Inter-Frame Learning Equalizer

The complexity comparisons between intra-frame and inter-frame learning are shown in Table 1, including addition, multiplication, and memory. In this article, the LMS and NLMS were selected as representative algorithms for intra-frame learning, while the DNN was chosen as a representative for inter-frame learning. There is no doubt that the complexity and storage parameters of inter-frame learning methods are higher than those of classical adaptive equalizers. Concretely, the LMS and NLMS algorithms, due to their simple structures, exhibit computational and memory requirements that scale linearly with the filter length N, making them highly efficient. Among them, NLMS introduces a normalization step, trading approximately twice the computational cost of LMS for more stable convergence performance. In contrast, the complexity of DNNs depends entirely on their network scale, with both computational load and memory usage growing rapidly with the number and width of network layers, often reaching quadratic or even higher orders. Thus, despite its powerful capabilities, the DNN requires significantly more computational resources than the former two algorithms.

4. The Proposed Meta-Learning-Based Inter-Frame Learning Strategy

In a time-varying condition, each communication round at a different slot will be distorted by various channel impulse responses. Therefore, the well-trained neural network (NN) should be fine-tuned to adapt to the channel for precise communication. In a classical adaptive filter algorithm, the filter parameters are updated based on a pilot or training sequence. From the same perspective, we aim to use the pilot to optimize the DL equalizer’s parameters for improved performance. To further improve the DL equalizer f ( θ ) ’s BER performance, we proposed a meta-learning-based training strategy for the SC system in the time domain using inter-frame learning, dubbed Meta-DNN. A flowchart of the meta-learning-based training strategy is shown in Figure 4. In real-world communication environments, the trained parameters need to be fine-tuned to adapt to the new environment. Hence, the meta-learning scheme is employed for inter-frame learning. The communication frame contains a pilot segment and a data segment. We divide them into two categories: the support set and the query set. The support set contains a pilot symbol of multiple communication rounds, and the query set includes a data symbol generated from multiple communication rounds. The support set for meta-learning can be denoted as D t P i l o t , where D t P i l o t = { ( ( y k m P , , y k 1 P , y k P ) , x k P ) } is the received pilot signal pairs of the t-th time slot. Moreover, the query set is noted as D t D a t a , where D t D a t a = { ( ( y k m Q , , y k 1 Q , y k Q ) , x k Q ) } are the received data signal pairs of the t-th time slot. The meta-learning-based inter-frame learning strategy is described in Algorithm 1.
Algorithm 1 The meta-learning-based inter-frame learning strategy.
Require: Support set generated by pilot symbol D t P i l o t and the query set generated by data symbol D t D a t a , the update steps T, step size α and β .
  1:
Initialize: θ
  2:
Online training stage, both support set and query set are employed for updating
  3:
for  t = 1 to T do
  4:
   Using D t P i l o t update the parameters according to
  5:
    ϕ t ( θ ) = θ t α θ L ( θ t , D t P i l o t ) ;
  6:
  Then, using D t D a t a update the parameter according to
  7:
    θ t + 1 = θ t β θ L ( ϕ t ( θ t ) , D t D a t a ) ;
  8:
   obtain the θ t + 1
  9:
end for
10:
Offline adaption stage, only support set is used to fine-tune the parameters according to Formula (7).
Ensure:  θ opt
The neural network was first updated using the support data D t P i l o t , which were generated from different time slot t communications. Given the limited pilot length, the sliding step size is set to 1 to obtain more fine-tuning data for the query data. For the proposed algorithm, given the model parameters of the DL equalizer, we assume the DL equalizer can update its own parameters by few-step learning procedures according to the gradient descent based on D t P i l o t , namely
ϕ t ( θ ) = θ t α θ L ( θ t , D t P i l o t ) ,
where α is the adaptation learning rate.
Then, the loss L ( ϕ t , D t P i l o t ) is evaluated according to Formula (5) and θ is updated by using the query data D t D a t a ulteriorly according to
θ t + 1 = θ t β θ L ( ϕ t ( θ t ) , D t D a t a ) .
where β is the update learning rate. It is worth mentioning that during online training, both pilot symbol D t P i l o t and data symbol D t D a t a are used to train the DL network. On the contrary, during the offline testing phase, only pilot data is used to update the network to adapt to data under different conditions.

5. Simulation and Discussion

In our experiments, PyTorch 1.2.0 is chosen as the development framework. The SC system with 1024 bits is considered. 4-QAM is used as the modulation scheme. SC blocks contain the pilots and transmitted symbols. We adopt the mean squared error (MSE) loss function and the adaptive moment estimation (Adam) optimizer with a 0.001 learning rate. A four-layer FC-DNN structure with one input layer, two hidden layers, and one output layer is used, where the number of neurons in each layer is 160, 80, 40, and 1, respectively. In our experiments, each frame contains M = 200 pilots and 1000 QPSK symbols. The local adaptation rate α is 0.001, and the update rate β is 0.0001. The step size of the LMS and NLMS is 2 × 10 5 . The Norway–Oslofjord (NOF) and Norway–Continental Shelf (NCS) channels are used in this section [27] as the sea-trail-measured channels. The center frequency is 14 kHz. The maximum Doppler shifts of the NOF and NCS channels are 3.9 Hz and 15.7 Hz, respectively. The delay coverage of the NOF and NCS channels is 128 ms and 32 ms, respectively. Meanwhile, the channel response can be replayed as the QS channel and the TV channel. Hence, four conditions are considered in the simulations:, namely QS-NOF, QS-NCS, TV-NOF, and TV-NCS. In each channel type, 80 % is used for training and 20 % for testing. The training epochs are set to 100 or 500, and the batch size is 10 frames per epoch, each containing at least 2000 pilot symbols and 20,000 data symbols. Meanwhile, the channel response will change from frame to frame.

5.1. Pre-Equalization Methods

The comparison algorithms are the normalized LMS (NLMS) algorithm and the matching pursuit (MP) algorithm. Meanwhile, we introduce a pre-equalization method to eliminate the partial channel influence. Therefore, considering the pure DL receiver requires a longer training period, we employ a pre-equalization approach to perform pre-processing and obtain a better initial value. Pre-equalization can provide a better initial value and accelerate convergence during training. We combine our proposed FC-DNN equalizer and the NLMS time-domain equalizer to form a hybrid equalization procedure, which is a novel and effective approach for communication systems, as it can benefit from two equalizers at different stages of the equalization process.

5.2. BER Performance

Figure 5 systematically compares the BER performance of seven equalization algorithms, including NLMS, MP, MP with NLMS, FC-DNN, Meta-DNN, MP with FC-DNN, and MP with Meta-DNN, across typical quasi-static and time-varying scenarios. The results demonstrate that under quasi-static channel conditions, all algorithms exhibit a steady decline in BER as SNR increases. Among them, DNN-based algorithms, particularly Meta-DNN, show significant advantages in the low-SNR region, reducing BER more than traditional methods do. In time-varying channels, the performance of adaptive algorithms generally degrades. Notably, the MP with the Meta-DNN combination maintains optimal performance, highlighting its strong robustness against channel time variations. Traditional methods, such as NLMS, perform worst in TV-NCS channels, with BER failing to drop below 10 2 . MP with Meta-DNN consistently achieves the best performance across all channel conditions, reaching a BER of 10 3 at 20 dB SNR over TV-NCS channels. Standalone Meta-DNN slightly underperforms MP with Meta-DNN but still significantly outperforms FC-DNN. MP with FC-DNN approaches Meta-DNN performance in static channels but shows a noticeable gap in time-varying scenarios. The traditional MP with the NLMS scheme only approaches DNN-based performance in static channels at a high SNR. NLMS and MP consistently perform worst, with BER remaining around 10 1 over the time-varying NCS channel. The result confirms the superior adaptability of Meta-DNN-based methods, particularly in dynamic channel environments, while traditional approaches struggle to maintain reliable performance.

5.3. Training Epoch Performance

Figure 6 presents the learning-based receiver without pre-processing, also named the pure learning-based receiver, and the learning-based receiver with pre-processing. From Figure 6a, we know that the TV channel needs more training epochs than the QS channel. We also compare the training speed across two channel types. The learning process under the mild channel, NOF, exhibits a faster learning speed compared to that under the tough channel, NCS. From Figure 6b, the pre-equalization-based approach has fewer training epochs. It is because equalization is applied before the DL equalizer is used, which improves the initialization of the training data. Therefore, it speeds up learning convergence. However, there is a drawback: when the signal is distorted by a severe channel, the classical pre-equalization method cannot restore the signal to a relatively normal initial state, even under adverse conditions. It results in a higher MSE for the pre-equalization-based training process compared to the pure learning approach. The comparison of the MSE convergence characteristics of the proposed method under four typical channel environments reveals the key impact of channel characteristics on model training. Experimental results show that MSE decreases monotonically across all channel conditions. However, there are significant differences: in the quasi-static NOF channel (QS-NOF), MSE converges fastest and performs best, eventually stabilizing at the order of 10 2 , while in the time-varying channel (TV-NOF/TV-NCS), MSE is always high, especially in short-term training, with significant fluctuations. Specifically, the training process can be divided into two stages: in the early stage, 0–100 epochs, the gradient descent efficiency of the QS-NOF channel is five times that of the TV-NCS channel; in the middle stage, 100–500 epochs, QS-NOF reaches stability in 200 rounds, while TV-NOF needs 400 rounds to approach convergence, and TV-NCS has never fully converged. Further analysis reveals that the impact of time variability on MSE is 1.6 times that of NCS characteristics, directly resulting in a final MSE for TV-NCS that is higher than that for QS-NOF. Based on these findings, it is recommended to adopt a channel-aware training strategy in engineering practice: for an ideal environment, it can be used as a benchmark for algorithm performance verification; for time-varying channels, it is recommended to use a dynamic learning rate and extend the training cycle to more than 600 rounds; for NCS channels, it is necessary to add a delay spread compensation module to the network and improve the nonlinear activation function. These optimization measures will effectively enhance the model’s adaptability and convergence performance across different channel conditions.

5.4. Pilot Number Effect

Figure 7 shows the performance impact of different training sequence lengths, also known as pilot numbers, on the BER as the number of training rounds changes under the condition of time-invariant channels. Pilots number P = 200 , P = 400 , and P = 1000 are all considered. The BER decreases significantly with increasing the number of training rounds, and all curves show a monotonically decreasing trend, consistent with the expectation that the DNN model is gradually optimized during the training process. The longer the training sequence P, the faster the BER decreases. When P = 1000, the BER curve decreases the steepest, and the BER is close to 10 3 at about 150 rounds of training, finally stabilizing at the lowest level. The convergence speed of P = 400 and P = 200 slows down successively. In particular, P = 200 requires more than 300 rounds to achieve the performance of P = 1000 at 150 rounds. The challenge of time-invariant Rayleigh channels and multi-path fading requires the DNN to have strong nonlinear fitting capabilities, and increasing P directly improves the model’s representation ability, thereby approaching the theoretical optimal solution faster. However, as P increases, so does the redundancy overhead and network complexity. P = 400 can be used as a balanced choice, with a final BER close to P = 1000, a moderate number of training rounds, and higher spectrum efficiency. P = 200 is suitable only for scenarios with strict resource constraints and where significant performance loss can be tolerated. The more pilots, the lower the BER we can achieve. However, the disadvantage is that too many pilots will slow down the communication rate. Therefore, we have a balance between the BER and the communication rate.

5.5. Filter Length Effect

Figure 8 clearly shows the effect of filter length, including M = 30, M = 50, and M = 100, on the MSE loss during neural network training. As demonstrated, a longer filter length is associated with a lower MSE. The MSE loss across all configurations decreases significantly with increasing training epochs before eventually stabilizing, consistent with typical training convergence. Faster convergence is achieved with larger M values. The curve for M = 100 drops most steeply and stabilizes at around 20 epochs, whereas the convergence for M = 30 is the slowest, requiring more epochs to reach a comparable loss level. At convergence, the lowest BER, approximately 10 1 , is attained by M = 100, indicating that the convergence of the second-stage equalizer can be accelerated by increasing the length of the first-stage filter. Although the final loss values for M = 30 and M = 50 are similar, M = 50 holds a clear advantage in the early training phase. A moderate increase in filter length can optimize the initial optimization.

5.6. Constellation Performance

Figure 9 presents the output constellations of a hybrid equalizer over the NOF channel under 25 dB and 30 dB. In all cases, the first equalizer employs an MP channel estimator and an MMSE equalizer. The second equalizer, however, varies across the subfigures: Figure 9a,d use the NLMS algorithm; Figure 9b,e use a standard neural network (NN); and Figure 9c,f uses a meta-learning-based neural network. Our proposed NN-based algorithm as a secondary equalizer obtains a separated constellation in Figure 9b,e compared with the traditional hybrid algorithm combined with the MP channel estimation and MMSE equalization. In particular, the NNs combined with meta-learning can get the most separated constellation in Figure 9c,f than the NNs without transfer learning. By comparison, the equalization algorithm using meta-learning exhibits better convergence in constellation diagrams. Moreover, we can also know that the constellation of NNs based on the receiver has better limitations within the range [ 1 , 1 ] . It is because we use a nonlinear tanh function as the activation function of the output layer.

6. Conclusions

In this paper, we concluded the investigation into learning-based equalization for single-carrier UWA communications. Meanwhile, a sliding deep neural network equalizer was proposed to overcome severe channel distortion, accompanied by a pre-processing strategy designed to accelerate training. To address time-varying channels, an adaptive online equalization scheme, Meta-DNN, was introduced and enhanced by meta-learning. In this scheme, pilot–data relationships were utilized to enable efficient transfer and improvement of BER performance. The superior and robust performance of the proposed approaches is confirmed through extensive tests conducted under reproducible, multi-scenario channels. In the future, we will improve the generalization of the proposed equalization method and leverage more advanced networks to enhance BER performance further. In addition, we will focus on potential deployment strategies, such as model quantization and pruning, to reduce computational and memory requirements for real-time inference.

Author Contributions

Conceptualization, H.Z., K.Y. and D.X.; methodology, H.Z., K.Y. and Y.W.; software, H.Z., Q.W. and K.Y.; validation, H.Z., Y.W. and K.Y.; formal analysis, H.Z., Y.W. and K.Y.; investigation, H.Z., Y.W. and K.Y.; resources, H.Z., Y.W. and K.Y.; data curation, H.Z., Y.W. and K.Y.; writing—original draft preparation, H.Z., D.X. and Q.W.; writing—review and editing, H.Z., K.Y. and D.X.; visualization, H.Z.; supervision, D.X. and Y.C.; project administration, H.Z. and D.X.; funding acquisition, H.Z. and D.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China under Grant 62401167, in part by the Youth S&T Talent Support Programme of Guangdong Provincial Association for Science and Technology (GDSTA) under Grant SKXRC2025418, in part by the Featured Innovation Projects for Guangdong Provincial Higher Education Institutions under Grant 2025KTSCX109 and in part by the Key Laboratory of Marine Environmental Survey Technology and Application, Ministry of Natural Resources, P. R. China, under Grant MESTA-2023-B001.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

The authors would like to thank the anonymous reviewers for their careful assessment of our work.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhao, H.; Ji, F.; Wang, Y.; Yao, K.; Chen, F. Space–Air–Ground–Sea Integrated Network with Federated Learning. Remote Sens. 2024, 16, 1640. [Google Scholar] [CrossRef]
  2. Ullah, I.; Ali, F.; Sharafian, A.; Ali, A.; Naeem, H.M.Y.; Bai, X. Optimizing underwater connectivity through multi-attribute decision-making for underwater IoT deployments using remote sensing technologies. Front. Mar. Sci. 2024, 11, 1468481. [Google Scholar] [CrossRef]
  3. Jahanbakht, M.; Xiang, W.; Hanzo, L.; Rahimi Azghadi, M. Internet of Underwater Things and Big Marine Data Analytics—A Comprehensive Survey. IEEE Commun. Surv. Tutor. 2021, 23, 904–956. [Google Scholar] [CrossRef]
  4. Theocharidis, T.; Kavallieratou, E. Underwater communication technologies: A review. Telecommun. Syst. 2025, 88, 54. [Google Scholar] [CrossRef]
  5. Li, Z.; Chitre, M.; Stojanovic, M. Underwater acoustic communications. Nat. Rev. Electr. Eng. 2025, 2, 83–95. [Google Scholar] [CrossRef]
  6. Zhao, H.; Yang, C.; Xu, Y.; Ji, F.; Wen, M.; Chen, Y. Model-Driven Based Deep Unfolding Equalizer for Underwater Acoustic OFDM Communications. IEEE Trans. Veh. Technol. 2023, 72, 6056–6067. [Google Scholar] [CrossRef]
  7. Zhao, H.; Wen, M.; Ji, F.; Liang, Y.; Yu, H.; Yang, C. Deep learning aided underwater acoustic OFDM receivers: Model-driven or data-driven? Digit. Commun. Netw. 2025, 11, 866–877. [Google Scholar] [CrossRef]
  8. Li, B.; Zhou, S.; Stojanovic, M.; Freitag, L.; Willett, P. Multicarrier Communication over Underwater Acoustic Channels with Nonuniform Doppler Shifts. IEEE J. Ocean. Eng. 2008, 33, 198–209. [Google Scholar] [CrossRef]
  9. Wu, J.; Qiao, G.; Qi, X. The research on improved companding transformation for reducing PAPR in underwater acoustic OFDM communication system. Discret. Dyn. Nat. Soc. 2016, 2016, 3167483. [Google Scholar] [CrossRef]
  10. Liang, Y.; Yu, H.; Xu, L.; Zhao, H.; Ji, F.; Yan, S. Joint Bayesian Channel Estimation and Data Detection for Underwater Acoustic Communications. IEEE Trans. Commun. 2024, 72, 5868–5883. [Google Scholar] [CrossRef]
  11. Tao, J.; Wu, Y.; Han, X.; Pelekanakis, K. Sparse Direct Adaptive Equalization for Single-Carrier MIMO Underwater Acoustic Communications. IEEE J. Ocean. Eng. 2020, 45, 1622–1631. [Google Scholar] [CrossRef]
  12. Chen, F.; Lin, S.; Zheng, B.; Li, Q.; Wen, M.; Liu, Y.; Ji, F. Minimum Symbol-Error Rate Based Adaptive Decision Feedback Equalizer in Underwater Acoustic Channels. IEEE Access 2017, 5, 25147–25157. [Google Scholar] [CrossRef]
  13. Yan, H.; Liu, S.; Pan, C.; Kuang, B.; Wang, S.; Qiao, G. Simulation of Non-Stationary Mobile Underwater Acoustic Communication Channels Based on a Multi-Scale Time-Varying Multipath Model. J. Mar. Sci. Eng. 2025, 13, 1765. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Venkatesan, R.; Dobre, O.A.; Li, C. Efficient Estimation and Prediction for Sparse Time-Varying Underwater Acoustic Channels. IEEE J. Ocean. Eng. 2020, 45, 1112–1125. [Google Scholar] [CrossRef]
  15. Liang, Y.; Yu, H.; Ji, F.; Chen, F. Multitask Sparse Bayesian Channel Estimation for Turbo Equalization in Underwater Acoustic Communications. IEEE J. Ocean. Eng. 2023, 48, 946–962. [Google Scholar] [CrossRef]
  16. Huang, L.; Wang, Y.; Zhang, Q.; Han, J.; Tan, W.; Tian, Z. Machine Learning for Underwater Acoustic Communications. IEEE Wirel. Commun. 2022, 29, 102–108, Digit. Commun. Netw. 2025, 11, 866–877. [Google Scholar] [CrossRef]
  17. Khan, S.; Ullah, I.; Ali, F.; Shafiq, M.; Ghadi, Y.Y.; Kim, T. Deep learning-based marine big data fusion for ocean environment monitoring: Towards shape optimization and salient objects detection. Front. Mar. Sci. 2023, 9, 1094915. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Chang, J.; Liu, Y.; Xing, L.; Shen, X. Deep learning and expert knowledge based underwater acoustic OFDM receiver. Phys. Commun. 2023, 58, 102041. [Google Scholar] [CrossRef]
  19. Zhang, Y.; Wang, H.; Li, C.; Chen, D.; Meriaudeau, F. Meta-learning-aided orthogonal frequency division multiplexing for underwater acoustic communications. J. Acoust. Soc. Am. 2021, 149, 4596–4606. [Google Scholar] [CrossRef] [PubMed]
  20. Zhao, H.; Ji, F.; Wen, M.; Yu, H.; Guan, Q. Multi-task learning based underwater acoustic OFDM communications. In Proceedings of the IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Xi’an, China, 17–20 August 2021; pp. 1–5. [Google Scholar]
  21. Liu, J.; Ji, F.; Zhao, H.; Wen, M. CNN-based underwater acoustic OFDM communications over doubly-selective channels. In Proceedings of the IEEE 94th Vehicular Technology Conference (VTC2021-Fall), Virtual, 27 September–28 October 2021; pp. 1–6. [Google Scholar]
  22. Zuberi, H.H.; Liu, S.; Bilal, M.; Alharbi, A.; Jaffar, A.; Mohsan, S.A.H.; Miyajan, A.; Khan, M.A. Deep-neural-network-based receiver design for downlink non-orthogonal multiple-access underwater acoustic communication. J. Mar. Sci. Eng. 2023, 11, 2184. [Google Scholar] [CrossRef]
  23. Zhang, Y.; Li, J.; Zakharov, Y.V.; Li, J.; Li, Y.; Lin, C.; Li, X. Deep Learning Based Single Carrier Communications over Time-Varying Underwater Acoustic Channel. IEEE Access 2019, 7, 38420–38430. [Google Scholar] [CrossRef]
  24. Lee-Leon, A.; Yuen, C.; Herremans, D. Underwater Acoustic Communication Receiver Using Deep Belief Network. IEEE Trans. Commun. 2021, 69, 3698–3708. [Google Scholar] [CrossRef]
  25. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the International Conference on Machine Learning (PMLR), Sydney, Australia, 6–11 August 2017; Volume 70, pp. 1126–1135. [Google Scholar]
  26. Zhao, H.; Ji, F.; Li, Q.; Guan, Q.; Wang, S.; Wen, M. Federated Meta-Learning Enhanced Acoustic Radio Cooperative Framework for Ocean of Things. IEEE J. Sel. Top. Signal Process. 2022, 16, 474–486. [Google Scholar] [CrossRef]
  27. van Walree, P.A.; Socheleau, F.-X.; Otnes, R.; Jenserud, T. The Watermark Benchmark for Underwater Acoustic Modulation Schemes. IEEE J. Ocean. Eng. 2017, 42, 1007–1018. [Google Scholar] [CrossRef]
Figure 1. The single-carrier communication system structure.
Figure 1. The single-carrier communication system structure.
Jmse 13 02209 g001
Figure 2. A comparison of the inter-frame learning and intra-frame learning.
Figure 2. A comparison of the inter-frame learning and intra-frame learning.
Jmse 13 02209 g002
Figure 3. The sliding signal processing structure.
Figure 3. The sliding signal processing structure.
Jmse 13 02209 g003
Figure 4. The meta-learning based inter-frame learning strategy.
Figure 4. The meta-learning based inter-frame learning strategy.
Jmse 13 02209 g004
Figure 5. The BER performance comparision of seven equalization schemes over typical channel conditions: (a) quasi-static NOF channels; (b) time-varying NOF channels; (c) quasi-static NCS channels; (d) time-varying NCS channels.
Figure 5. The BER performance comparision of seven equalization schemes over typical channel conditions: (a) quasi-static NOF channels; (b) time-varying NOF channels; (c) quasi-static NCS channels; (d) time-varying NCS channels.
Jmse 13 02209 g005
Figure 6. The training epoch comparision of the pure learning based receiver and pre-equalization-based DL receiver. (a) Only deep learning-based methods; (b) pre-equalization-based DL methods.
Figure 6. The training epoch comparision of the pure learning based receiver and pre-equalization-based DL receiver. (a) Only deep learning-based methods; (b) pre-equalization-based DL methods.
Jmse 13 02209 g006
Figure 7. The performance impact of different training sequence lengths P.
Figure 7. The performance impact of different training sequence lengths P.
Jmse 13 02209 g007
Figure 8. The effect of different filter length M.
Figure 8. The effect of different filter length M.
Jmse 13 02209 g008
Figure 9. Hybrid equalizer’s output constellation under 25 dB and 30 dB: (a) the MP channel estimation-based equalizer under 25 dB. (b) The MP channel estimation and NN under 25 dB. (c) The MP channel estimation and meta-learning NN-based equalizer under 25 dB. (d) The MP channel estimation-based equalizer under 30 dB. (e) The MP channel estimation and NN-based equalizer under 30 dB. (f) The MP channel estimation and meta-learning NN-based equalizer under under 30 dB.
Figure 9. Hybrid equalizer’s output constellation under 25 dB and 30 dB: (a) the MP channel estimation-based equalizer under 25 dB. (b) The MP channel estimation and NN under 25 dB. (c) The MP channel estimation and meta-learning NN-based equalizer under 25 dB. (d) The MP channel estimation-based equalizer under 30 dB. (e) The MP channel estimation and NN-based equalizer under 30 dB. (f) The MP channel estimation and meta-learning NN-based equalizer under under 30 dB.
Jmse 13 02209 g009
Table 1. Comparison of the complexity of the LMS, NLMS, and DNN equalizers.
Table 1. Comparison of the complexity of the LMS, NLMS, and DNN equalizers.
AlgorithmAdditionMultiplicationMemory
LMSNN + 1N
NLMS2N2N + 1N
DNN v = 2 V N v × N v 1 v = 2 V N v × N v 1 v = 1 V N v
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, H.; Yao, K.; Xiang, D.; Wang, Q.; Chen, Y.; Wang, Y. Learning to Equalize for Single-Carrier Underwater Acoustic Communications. J. Mar. Sci. Eng. 2025, 13, 2209. https://doi.org/10.3390/jmse13112209

AMA Style

Zhao H, Yao K, Xiang D, Wang Q, Chen Y, Wang Y. Learning to Equalize for Single-Carrier Underwater Acoustic Communications. Journal of Marine Science and Engineering. 2025; 13(11):2209. https://doi.org/10.3390/jmse13112209

Chicago/Turabian Style

Zhao, Hao, Kexing Yao, Dan Xiang, Qisen Wang, Yankun Chen, and Yan Wang. 2025. "Learning to Equalize for Single-Carrier Underwater Acoustic Communications" Journal of Marine Science and Engineering 13, no. 11: 2209. https://doi.org/10.3390/jmse13112209

APA Style

Zhao, H., Yao, K., Xiang, D., Wang, Q., Chen, Y., & Wang, Y. (2025). Learning to Equalize for Single-Carrier Underwater Acoustic Communications. Journal of Marine Science and Engineering, 13(11), 2209. https://doi.org/10.3390/jmse13112209

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop