Next Article in Journal
Collapse Behavior of Onshore and Spar-Floating Wind Turbine Towers During Blade Pitch Malfunction
Next Article in Special Issue
In Situ Measurement of Oceanic 3D-Volume Two-Component Turbulence Based on Holographic Astigmatic Particle Tracking Velocimetry
Previous Article in Journal
An Incorporating Pore Water Pressure Constitutive Model for Overconsolidated Clay and Calibration of Transient FE Parameters
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction Model for Maritime 5G Signal Strength Based on ConvLSTM-PSO-XGBoost Algorithm

1
Department of Information Engineering, Zhejiang Ocean University, Zhoushan 316022, China
2
Key Laboratory of Marine Environmental Survey Technology and Application, Ministry of Natural Resources, Guangzhou 510300, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2026, 14(4), 377; https://doi.org/10.3390/jmse14040377
Submission received: 13 January 2026 / Revised: 8 February 2026 / Accepted: 13 February 2026 / Published: 16 February 2026
(This article belongs to the Special Issue Advances in Ocean Observing Technology and System)

Abstract

The accurate prediction of signal strength plays an important role in estimating radio signal quality, thus forming the essential foundation for the planning, optimization, and reliable operation of modern wireless network systems. This paper proposes a new hybrid model for predicting maritime 5G signal strength, combing Convolutional Long Short-Term Memory (ConvLSTM) with Particle Swarm Optimization-extreme Gradient Boosting (PSO-XGBoost). The model was developed and validated using a dataset comprising 22 columns, 2994 rows, and 21 features, collected via a research vessel in Zhoushan Port, China. Four evaluation metrics, Root Mean Square Error (RMSE), Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), and the coefficient of determination (R2) were employed to assess model performance and interpretability. Comparative experiments against various popular models demonstrated the hybrid model’s superior performance in predicting maritime 5G signals. Its accuracy surpassed both standalone ConvLSTM and XGBoost models, while achieving lower MAE and RMSE values compared to various popular models. This study provides a method for predicting coverage conditions based on navigation and environmental data, without relying on radio key performance indicators. Furthermore, it supplies high-quality signal data to advance the modeling of marine communication channels.

1. Introduction

1.1. Background

Over 70% of the Earth’s surface is covered by oceans and seas [1]. With the continuous increase in maritime activities, the demand for reliable maritime communication technologies is also rising. Effective maritime communications are critical for numerous applications [2], including environmental monitoring, offshore oil extraction, wind power generation, maritime surveillance, fisheries, and tourism. Simultaneously, maritime wireless channel modeling forms the core foundation for designing and optimizing marine communication systems. By accurately describing the propagation characteristics of electromagnetic waves in marine environments, it enhances communication reliability, efficiency, and coverage.
Maritime wireless channel modeling presents a challenging research topic, aiming to characterize the channel properties between vessels [3], satellites [4], drones [5], unmanned surface vehicles [6], and other maritime infrastructure. To achieve efficient wireless communication systems, establishing accurate maritime wireless communication channel models is crucial [7]. Currently, maritime wireless channel modeling relies on collected data; yet, acquiring reliable maritime communication data for channel modeling remains challenging. Additionally, maritime 5G signals face complex maritime environments and vessel motion/interference, rendering many of the collected data unusable [8]. For instance, air temperature and humidity affect sea surface reflectivity [9,10] and spatial refractive index [11], thereby affecting signal strength. Concurrently, sea waves influence the dynamic multipath propagation of signals while inducing Doppler shifts and alterations in antenna directivity, thereby causing fluctuations in maritime wireless communications [12]. This complex and dynamic environment poses significant challenges for modeling maritime wireless channels. Therefore, this paper addresses the current difficulty in acquiring data for maritime modeling by providing usable data for maritime wireless channel modeling through predicting 5G signals at sea.

1.2. Related Works

Driven by rapid advancements in computer science and big data, artificial intelligence (AI) and neural networks have evolved significantly. AI technologies offer novel approaches for modeling marine wireless channels. AI applications span machine learning [13], computer vision [14] and natural language processing [15]. These advancements have enabled applications such as autonomous driving [16], image recognition [17] and facial recognition [18]. Some scholars have employed machine learning to predict maritime 5G signal strength in complex marine environments. However, model usability hinges on prediction accuracy. Thus, developing a predictive model capable of accurately forecasting maritime 5G signal strength is essential.
As artificial intelligence systems have gained widespread adoption, common methods for predicting maritime 5G signal strength previously included machine learning-based and deep learning-based spatial feature models. Ahmed [19] elaborated on the application prospects of AI in Industry Internet 4.0 and pointed out that the application prospects of AI at sea are broad. Sánchez et al. [20] investigated Wi-Fi coverage modeling in Checa, Ecuador, using deterministic and empirical propagation models with Raspberry Pi and GPS measurements, demonstrating that a logarithmic regression-adjusted model achieved significantly improved accuracy (path loss exponent of 8.96) for predicting signal strength in community Wi-Fi network planning. Palacios Játiva et al. [21] analyzed the performance of four machine learning algorithms (Coalition Game Theory, Naive Bayesian Classifier, Support Vector Machine, and Decision Trees) for spectrum detection and decision in Cognitive Mobile Radio Networks using Network Simulator 3, demonstrating that SVM outperformed other algorithms across all metrics including probability of detection, probability of false alarm, classification quality, and simulation time. Machine learning-based models can be implemented using Python’s Sklearn package [22] or Tensorflow [23] and PyTorch [24]. Currently, approaches to predicting maritime 5G signal strength can be divided into three categories: traditional machine learning techniques, deep learning machine techniques, and hybrid model techniques.

1.2.1. Traditional Machine Learning Research

Through traditional machine learning techniques serving as a critical means to extract patterns from data for prediction and decision making through algorithms and statistical methods, we have witnessed the sequential emergence of various algorithms throughout their extensive development history. Each algorithm, driven by its inherent characteristics and evolving demands of the era, has collectively advanced the field. In the early stages, linear regression emerged, focusing on modeling linear relationships by constructing linear equations between input and output variables to achieve data fitting and prediction. As data complexity increased, decision tree algorithms were introduced. Decision trees employ a tree-like structure to decompose the decision-making process into a series of feature-based judgment nodes. Starting from the root node, the process branches downward according to different feature conditions, ultimately arriving at leaf nodes to yield decision outcomes. This intuitive tree-based representation renders the decision process transparent, facilitating comprehension and interpretability. Subsequently, to address more complex classification tasks, particularly in high-dimensional data spaces, the Support Vector Regression (SVR) algorithm was developed. SVR is a supervised learning model used for regression problems. It attempts to find an optimal hyperplane such that the distances (errors) from all training data points to this plane are as small as possible, while controlling the complexity of the model to avoid overfitting. Subsequently, XGBoost was developed primarily to address the suboptimal performance of models such as SVR, linear regression, and decision trees in handling certain complex data structures [25]. In signal prediction, these algorithms can effectively extract features and make predictions. For instance, C.Peréz-Wences et al. [26] employed SVR algorithm for power prediction, while Chaudhary, P et al. [27] proposed an enhanced linear regression framework for the common Bernoulli–Gaussian noise problem in wireless communication. Through the gradient descent optimization algorithm, more accurate channel estimation was achieved in non-Gaussian noise environments, providing an efficient solution for 5G/6G communication systems. G. Jangir et al. [28] employed the regression method of the machine learning model decision tree for prediction. However, when dealing with large datasets, traditional machine learning algorithms may suffer from insufficient computational efficiency, limiting real-time prediction capabilities [29]. Furthermore, traditional methods heavily rely on features; if features are not effectively selected and processed, model performance often suffers significantly.

1.2.2. Research on Deep Learning Technologies

Deep learning technology, a branch of machine learning, fundamentally operates through multi-layer neural networks to perform feature extraction and pattern recognition [30]. Historically, deep learning has evolved from early single-layer neural networks to encompass diverse models such as convolutional neural networks (CNNs), recurrent neural networks (RNNs), and generative adversarial networks (GANs). CNNs excel at processing image data, automatically extracting local features, while RNNs demonstrate significant advantages in handling time-series data, capturing dynamic characteristics of data over time.
Deep learning demonstrates formidable modeling capabilities in signal strength prediction, automatically learning complex data features and proving particularly well suited for processing image and time-series data. For instance, researchers can extract valuable information from a signal’s spatial characteristics using convolutional neural networks, while recurrent neural networks can model temporal variations in signals to achieve more precise predictions. CNNs are employed for prediction, leveraging their powerful feature extraction capabilities to achieve significant improvements in accuracy. By constructing multi-layer CNNs, signal strength across different sea areas can be forecasted, with the model maintaining high predictive precision even in complex environments. Additionally, RNN models excel at processing time-series signals, effectively capturing dynamic characteristics of signal variations—making them particularly well suited for scenarios with frequent changes in marine environments. For instance, Oroza et al. [31]. employed four machine learning models to predict Received Signal Strength Indicator (RSSI). Peng et al. [32]. utilized a deep learning model to predict Received Signal Strength Probability (RSRP), evaluating results using MAE.
Deep learning technology is revolutionizing protocol design in the fields of maritime Internet of Things (IoT) communication and machine learning, with the DL-HEED protocol serving as a prime example. The core innovation of the DL-HEED protocol lies in its utilization of GNNs to capture complex relationships among nodes, integrating multi-dimensional features such as residual energy, node degree, spatial location, and signal strength to enable context-aware dynamic decision making.
In practical maritime communication applications, deep learning-driven protocols also demonstrate immense potential. For instance, Juwaied et al. [33] proposed DL-HEED, a novel deep learning-based clustering protocol that integrates Graph Neural Networks (GNNs) into the HEED algorithm for energy-efficient cluster head selection in heterogeneous wireless sensor networks, achieving up to 60% improvement in network lifetime and energy efficiency compared to traditional HEED through intelligent, context-aware decision making using node features including residual energy, degree, spatial position, and signal strength. Rigas et al. [34] proposed an end-to-end deep learning framework for fault detection in marine machinery, leveraging Graph Attention Networks (GATs) to extract spatio-temporal information from multivariate time-series sensor data, providing explainable predictions through a feature-wise scoring mechanism and addressing maritime-specific challenges including scarce labeled data and proprietary datasets through a scalable edge-to-cloud architecture using MQTT protocol for data collection and a custom evaluation metric balancing prediction accuracy and fault identification timeliness. In marine IoT semantic communication, Han et al. [35] proposed a Turbo-based deep semantic auto encoder that combines Transformers with Turbo coding to achieve efficient text transmission under low signal-to-noise ratio conditions. Hussain et al. [36] proposed the Cooperative-Relay Neighboring-Based Energy Efficient Routing (CR-NBEER) protocol for Marine Underwater Sensor Networks, implementing advanced relay optimization and cooperative routing schemes that achieved superior performance compared to existing protocols, including Co-UWSN, CEER, and NBEER, with simulation results demonstrating reduced end-to-end delay (19.3 ms average), lower energy consumption (7.644 j average), a higher packet delivery ratio (94.831% average), and reduced path loss (79.9 dB average). These studies demonstrate that deep learning technologies effectively address challenges such as energy constraints, high dynamic variability, and stringent reliability requirements in maritime/IoT communication environments through intelligent routing selection, data aggregation, and anomaly detection mechanisms. These advancements provide technical support for the construction of efficient, reliable, and intelligent marine information systems.
Although deep learning technologies demonstrate numerous advantages in marine 5G signal strength prediction—such as high accuracy and adaptive capabilities—they still face considerable challenges. Compared to traditional machine learning methods, deep learning models typically require substantial data and powerful computational resources, with particularly significant computational resource consumption during the training phase [37]. Furthermore, deep learning models exhibit relatively low interpretability, which may pose difficulties in practical applications, especially when explanations and analyses of prediction results are required.

1.2.3. Research on Hybrid Model Techniques

With the exponential growth of training datasets, individual deep learning models struggle to handle such complex and massive data. Concurrently, hybrid models adapt better to increasingly complex data by integrating multiple deep learning models. Hybrid models offer the following advantages: First, by combining the strengths of multiple individual models, hybrid models achieve breakthrough improvements in predictive performance [38]. Second, hybrid models enhance adaptability to complex data and distribution shifts through diverse architectural designs and decision strategies. Third, while improving performance, hybrid models substantially reduce computational costs [39]. Fourth, ensemble models demonstrate exceptional task adaptability and flexibility. Therefore, many studies began to use hybrid models to predict signal strength. Abimannan et al. [40] found that hybrid learning methods have become powerful approaches to overcome the limitations of single learning models in terms of generalization, robustness, and performance. Meti [41] used an ensemble model to predict throughput in the deployment of a Wireless Local Area Network (WLAN).
Despite the increasingly widespread application of fusion models in maritime communication, research on the prediction of maritime 5G signal strength remains relatively scarce. Accurate prediction of wireless signal strength is of paramount importance for maritime wireless channel modeling, 5G network planning, mobility management, and resource scheduling. Meanwhile, most previous studies have relied on statistical methods or traditional machine learning models, failing to fully account for the complex and dynamic environmental factors in real-world scenarios, which results in limitations in prediction accuracy and applicability.
As a pioneering effort to address these challenges, this paper makes the following contributions: Firstly, the proposed ConvLSTM- PSO-XGBoost fusion model stands out conceptually for its systematic three-stage decomposition design. Initially, ConvLSTM is employed to extract spatiotemporal correlation features from raw signal sequences. This choice outperforms pure time-series models (e.g., TCN) and pure attention-based models (e.g., Transformer), offering lower computational overhead and better data efficiency in short-term high-frequency prediction tasks compared to Transformer. Secondly, a PSO-optimized XGBoost module is introduced for residual correction. The global search capability of PSO effectively avoids local optima traps in the high-dimensional hyperparameter space of XGBoost, significantly enhancing model generalization performance compared to default parameters or grid search strategies. Moreover, XGBoost’s regularization mechanism and tree ensemble structure demonstrate unique noise resistance advantages in handling tabular environmental features and fitting residuals unexplained by ConvLSTM. Finally, an entropy-weighted fusion is adopted to dynamically adjust the output weights of the two sub-models. This design surpasses traditional fixed-weight fusion and computationally expensive attention-based fusion mechanisms by quantifying the uncertainty of each model’s predictions in real time to allocate confidence levels, thereby improving prediction accuracy and robustness.
Overall, ConvLSTM is responsible for spatiotemporal pattern modeling, PSO-XGBoost handles nonlinear residual refinement, and entropy-based fusion manages adaptive weight scheduling. Together, they form a comprehensive solution that achieves improved accuracy, enhanced robustness, and preserved interpretability. Compared to alternative architectures such as Transformer and TCN, this solution demonstrates clearer engineering value and applicability boundaries in real-world deployment scenarios characterized by limited data and complex environments.

2. Maritime 5G Signal Strength Prediction Model

2.1. Particle Swarm Optimization Algorithm

PSO algorithm is a swarm-based stochastic optimization technique [42]. It simulates the foraging behavior of flocks, where particles move through the solution space to find optimal solutions [43]. Several critical parameters require configuration and selection in the PSO algorithm, such as the inertia weight and acceleration coefficient. These parameters determine how particle velocity and position are updated, significantly impacting algorithm performance and convergence. The particle velocity update formula and position update formula are as follows:
V i d k + 1 = ω V i d k + C 1 r 1 ( P i d X i d k ) + C 2 r 2 ( P g d X i d k )
X i d k + 1 = X i d k + V i d k + 1
where V i d k + 1 denotes the particle’s velocity after the kth iteration, X i d k + 1 primarily represents the particle’s position after the kth iteration, k denotes the iteration count, ω represents the inertia weight, C1 and C2 represent the learning factors, r1 and r2 denote acceleration factors, and Pid and Pgd, respectively, denote the local optimum of the particle swarm. Figure 1 elaborates on the workflow of PSO in detail.

2.2. XGBoost Model

XGBoost is an optimized implementation of Gradient Boosting Trees (GBDT), enhancing performance through the regularization of the objective function and parallel computation. By employing ridge regression, the model minimizes loss and variance while preserving objective function information.
The XGBoost model predicts final results through an additive expression composed of K (number of trees) base models, as follows:
y ^ i = k = 1 K f k ( x i )
where xi represents the ith sample, ŷi denotes the predicted output for the ith sample, k indicates the number of base models, and fk represents the kth base model.
The associated loss function takes the following form:
L = i = 1 n l ( y i , y ^ i )
where yi is the true value corresponding to the ith sample, ŷi denotes the predicted output for the ith sample, and L is the loss function.
The objective function is defined as follows:
O b j ( k ) = i = 1 n l [ y i , y ^ i ( k 1 ) ] + f t ( x i ) + k Ω ( f k )
Ω ( f ) = γ T + 1 2 λ | | w | | 2
where ŷi(k−1) is the sum of the outputs from the previous (k − 1) trees; fk(xi) is the output of the kth tree; and l is the differentiable convex loss function measuring the discrepancy between the predicted value ŷi and the true value yi. Ω(f) is the penalty term for model complexity; γ is the regularization parameter for the number of leaves; λ is the regularization parameter for leaf weights; w is the value of the leaf node; and T is the number of leaf nodes.

2.3. PSO-XGBoost Model

PSO-XGBoost is a hybrid model combining the PSO with XGBoost. It enhances model prediction performance by optimizing XGBoost hyperparameters via PSO. Its core idea is to leverage PSO’s global search capability to identify optimal XGBoost hyperparameter combinations, addressing the inefficiency and local optima trapping issues of traditional grid search. Figure 2 elaborates on the workflow of the PSO-XGBoost model.

2.4. Convolutional Long Short-Term Memory Model

ConvLSTM was proposed by Xingjian Shi et al. [44] to address spatio-temporal sequence prediction problems. Its core innovation replaces the fully connected operations in traditional LSTMs with convolutional operations, enabling the model to simultaneously capture temporal dependencies and spatial features. This makes it suitable for spatio-temporal structured data such as radar echo map prediction and video frame generation.
ConvLSTM inherits the gating mechanism of LSTM, but all inputs, hidden states, and cell states are 3D tensors (channel × height × width), enabling local feature interactions through convolutional operations.
The input gate it controls the proportion of current input data Xt stored in the memory cell Ci, as detailed in the following formula:
i t = σ ( W x i X t + W h i H t 1 + W c i ° C t 1 + b i )
where * denotes convolution; ∘ denotes Hadamard product; σ() denotes the sigmoid activation function; Wxi represents the weight for the current input Xt; Whi denotes the weight for the previous hidden state Ht−1; Wci denotes the weight for the previous memory cell Ct−1; and bi represents the threshold for the input gate.
The forget gate ft controls the proportion of the previous memory cell state Ct−1 preserved into the current memory cell Ct, with the specific formula as follows:
f t = σ ( W x f X t + W h f H t 1 + W c f ° C t 1 + b f )
where Wxf is the weight of the current input Xt; Whf is the weight of the previous hidden state Ht−1; Wcf is the weight of the previous cell state Ct−1; and bf is the respective thresholds for the forget gates.
The output gate ot controls the proportion of the current memory cell state Ct stored in the ConvLSTM network’s output value Ht, as expressed by the following formula:
o t = σ ( W x o X t + W h o H t 1 + W c o ° C t + b o )
where Wxo is the weight for the current input Xt; Who is the weight for the previous hidden state Ht−1; Wco is the weight for the previous memory cell Ct−1; and bo represents the threshold for the forget gate.
The memory cell Ct is obtained through a nonlinear transformation involving the output of the forget gate, the state of the memory cell at the previous time step, the input gate, the current input, and the hidden state at the previous time step:
C ˜ t = tanh ( W x c X t + W h e H t 1 + b c )
C t = f t ° C t 1 + i t ° C ˜ t
where Wxc represents the weight of the current input Xt; Whe represents the weight of the previous hidden state Ht−1; and bc denotes the threshold of the memory cell.
The hidden state Ht is obtained by applying the Hadamard product to the output of the output gate and the output of the Tanh activation function applied to the current memory cell state Ct.
H t = o t ° tanh ( C t )
Figure 3 illustrates the fundamental operational principle of the ConvLSTM architecture.

2.5. ConvLSTM-PSO-XGBoost Model

The prediction sets generated by ConvLSTM and PSO-XGBoost are fused using an entropy-weighted fusion method to combine the predictions from both models into new prediction data. The formulas for calculating the prediction error and weights in the fusion process are as follows.
ω 1 = 1 e r r 1 1 e r r 1 + 1 e r r 2 ,   ω 2 = 1 ω 1
y e n s e m b l e = w 1 × y 1 + w 2 × y 2
where err1 = |ytrue − y1| + 10−8 represents the prediction error of Model 1 and err2 = |ytrue − y2| + 10−8 is the prediction error of Model 2. w1 and w2 represent the weights of the two models. y1 and y2 represent the predictions of the two models. yensemble is the final prediction value after fusion.
Figure 4 illustrates the comprehensive architecture of the ConvLSTM-PSO-XGBoost model in detail.

3. Model Training

3.1. Experimental Data Collection

The experimental data in this paper are all collected from a scientific research vessel operating along the coast of Zhoushan City, Zhejiang Province, in the East China Sea. The testing was conducted from 07:30 to 14:48 on 29 April 2025. Data were collected over a total sampling duration of approximately seven hours at fixed 10-s intervals, using a fixed frequency of 700 MHz (Band 28). The vessel’s voyage commenced within the dashed-line area depicted in Figure 5. The vessel completed a total voyage of 63.14 nautical miles, equivalent to 116.93 km. The experimental equipment is a maritime 5G terminal developed by Super Radio. This terminal is specifically engineered for maritime environments, featuring high precision and low power consumption. The 5G terminal automatically switches frequency bands based on current network conditions. It transmits real-time data from various sensors, including a 9-axis motion sensor (comprising a 3-axis gyroscope, 3-axis accelerometer, and 3-axis compass), and is equipped with a Global Positioning System (GPS).

3.2. Feature Variable Selection

To accurately predict the target variable RSRP, it is essential to both understand the vessel motion attitude (roll, pitch, yaw, surge, sway, heave, etc.) and surrounding environment information (temperature, field strength, distance, etc.). Fifteen relevant influencing factors are listed below, followed by a brief description.
x1: surge, measured by an accelerometer to reflect the linear acceleration of the vessel in the X-axis direction, indicating the longitudinal motion state (forward/backward acceleration) of the vessel. The unit is usually g (9.8 m/s2).
x2: sway, measured by an accelerometer to reflect the linear acceleration of the vessel in the Y-axis direction, reflecting the lateral motion state of the vessel (acceleration of left and right swaying), and is used to monitor the force on the hull and the stability of navigation.
x3: heave, measured by an accelerometer to reflect the linear acceleration of the vessel in the Z-axis direction, reflecting the vertical motion state of the vessel (up and down jolting acceleration). It is commonly used for wave height estimation and navigation comfort assessment.
x4: roll, the angular velocity of the ship’s rotation around the X-axis, reflecting the rate of change of the longitudinal tilt of the hull. The unit is typically degrees per second (°/s) or radians per second (rad/s).
x5: pitch, the angular velocity of the ship’s rotation around the Y-axis, reflecting the speed of the ship’s heading change. The unit is usually degrees per second (°/s) or radians per second (rad/s).
x6: yaw, the angular velocity of the ship’s rotation around the Z-axis, reflecting the speed of the ship’s lateral tilt change. The unit is usually degrees per second (°/s) or radians per second (rad/s).
x7: roll angle, the angle of rotation around the X-axis, reflecting the longitudinal tilt of the hull. The unit is typically degrees (°).
x8: pitch angle, the angle of rotation around the Y-axis, reflecting the degree of deviation from the vessel’s course.
x9: heave angle, the angle of rotation around the Z-axis, reflecting the degree of lateral tilt of the hull.
x10: longitudinal magnetic (LM) field, measured by a magnetometer to determine the intensity of the Earth’s magnetic field in the X-axis direction of the vessel’s coordinate system. It is commonly used for course calculation and attitude correction, with units typically being Tesla (T) or Gauss (Gs).
x11: athwartship magnetic (AM) field, the intensity of the Earth’s magnetic field in the Y-axis direction of the vessel’s coordinate system.
x12: vertical magnetic (VM) field, the intensity of the Earth’s magnetic field in the Z-axis direction of the vessel’s coordinate system. It can be used to detect electromagnetic interference in the environment and assist in the calibration of the heading sensor.
x13: surrounding temperature (ST), the measurement of the air temperature around a vessel, typically expressed in °C or °F.
x14: ship speed (SS), the speed at which a vessel moves through water, typically measured in knots.
To extract the most critical input variables for predicting 5G signal strength, a feature selection strategy combining Pearson correlation coefficient analysis with the PSO algorithm was employed. The Pearson correlation coefficient [45], also known as the Pearson product–moment correlation coefficient, measures the linear correlation between two variables X and Y, with values ranging from −1 to 1. A value of 1 indicates perfect positive correlation, −1 indicates perfect negative correlation, and 0 indicates no linear relationship between the variables.
For two variables X = {x1, x2, …, xn} and Y = {y1, y2, …, yn}, the Pearson correlation coefficient r is calculated as follows:
r = n x i y i x i y i n x i 2 ( x i ) 2 n y i 2 ( y i ) 2
where r represents the correlation coefficient, xi represents the feature value of the ith sample in the input features, and yi represents the feature value of the ith sample in the output features.
As shown in Figure 6 (Pearson correlation heatmap), the LM field exhibits the highest correlation coefficient. Based on feature scoring, surge and heave were excluded in subsequent tests, retaining the remaining data.

3.3. Data Processing

Data preprocessing is preparatory work conducted before testing, including verifying dataset integrity, removing outliers, standardizing data, and partitioning the dataset. Standardization ensures each feature’s physical quantity is on the same order of magnitude, thereby simplifying subsequent experiments. The preprocessed dataset contains no missing values or outliers.
The data processing employed normalization, a technique prevalent in deep learning, which mitigates the likelihood of gradient explosion, accelerates convergence, stabilizes the training process, and ultimately enhances model performance. The objective of normalization is to transform the data into a distribution with a mean of zero and a standard deviation of one, thereby aligning it with the standard normal distribution. Consequently, prior to model training, all raw data—excluding binary variables and dummy variables—were normalized using the following formula:
z = x μ σ
where z is the standardized value, x is the raw feature value, μ is the feature mean and σ is the feature standard deviation.
The comparison between the standardized processed data and the raw data is shown in Table A1 in Appendix A.

3.4. Hyperparameter Tuning

In the model training process, 20% of the data are designated as the test set, which is exclusively utilized for data prediction to preclude any interference from feature selection and model evaluation on the predictive outcomes. The remaining 80% of the data are allocated as the training set. A temporal partitioning approach is employed for dataset construction, with a window length of five time steps, thereby eliminating temporal overlap and leakage between the test set and the training set.
The optimization of hyperparameters proceeded in two phases. The initial phase focused on configuring the ConvLSTM architecture and its associated training parameters. Subsequently, the XGBoost model’s hyperparameters were automatically optimized via the PSO algorithm. The ConvLSTM model employed a three-layer architecture with progressively decreasing filter sizes (32, 16, and 8) to capture spatiotemporal features at multiple scales. Each ConvLSTM2D layer utilized a 3 × 3 kernel with appropriate padding to preserve spatial dimensions. The input tensor shape was configured as (5, 1, 15, 1) to accommodate the temporal sequence length and spatial dimensions. A fully connected layer with 64 ReLU-activated neurons followed by a single output neuron was employed for the final prediction. Training was conducted using the Mean Squared Error (MSE) loss function with a batch size of 32 and a maximum of 150 epochs. To prevent overfitting, we implemented early stopping with a patience of 10 epochs, monitoring the validation loss. The training would terminate if the validation loss failed to improve for 10 consecutive epochs.
For the XGBoost model, we implemented the PSO algorithm to automatically search for optimal hyperparameters. The PSO algorithm was configured with a swarm size of 20 particles and a maximum of 30 iterations. The algorithm parameters were set as follows: inertia weight, individual learning factor, and global learning factor. The minimum fitness threshold was set to 0.001.
Figure 7 displays the training curves for ConvLSTM. ConvLSTM employs one-dimensional feature vectors (15 features) and transforms them into a pseudo-spatial representation by reshaping the original three-dimensional tensor of dimensions (number of samples, 5, 15) into a five-dimensional tensor of dimensions (number of samples, 5, 1, 15, 1). The figure shows two curves: training loss represents iterations during training, while verification loss indicates iterations during prediction. It is evident that convergence is achieved around iteration 10, with MSE loss stabilizing in subsequent iterations.
By employing the PSO algorithm to optimize XGBoost hyperparameters, the optimal parameter ranges for XGBoost were ultimately determined, as shown in Table 1.
Through iterative refinement, the optimal parameters for PSO-XGBoost and ConvLSTM are presented in Table 2. Other models employ network search methodologies to identify the parameter configurations that yield optimal performance in predictive tasks. For precise definitions of each hyperparameter, refer to the documentation of corresponding functions in scikit-learn.

3.5. Evaluation Metrics

This paper mainly adopts R2, MAE, RMSE, and MAPE as the evaluation metrics of the prediction results. The calculation formulas are as follows:
R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2
M A E = 1 n i = 1 n y i y ^ i
R M S E = 1 n i = 1 n ( y i y ^ i ) 2
M A P E = 1 n i = 1 n y i y ^ i y i × 100 %
where yi is actual value, ŷi is predicted value, y ¯ is average of actual values, and n is the number of test set samples.

3.6. ConvLSTM-PSO-XGBoost Model Fusion Approach

ConvLSTM effectively extracts spatio-temporal correlation features from data through convolutional operations and gating mechanisms. Meanwhile, PSO-XGBoost enhances the efficiency of traditional XGBoost in handling high-dimensional features and modeling nonlinear relationships by optimizing hyperparameters via the PSO algorithm. Due to the significant heterogeneity in model architecture and characteristics between the two, the fusion model demonstrates multiple outstanding advantages when dealing with complex data scenarios. Specifically, it exhibits complementarity. ConvLSTM can effectively capture long-term temporal dependencies within the data by virtue of its unique structure, delving into the underlying patterns of data changes over time. Meanwhile, PSO-XGBoost excels at handling complex interactions among high-dimensional features, accurately grasping the information embedded in different feature combinations. These two models complement each other, jointly providing a more comprehensive and in-depth perspective for the analysis of complex data. On the other hand, the fusion model demonstrates remarkable robustness. Single models are often prone to interference when confronted with noisy data, leading to deviations in analysis results. In contrast, the fusion model, by integrating the strengths of different models, can effectively reduce its sensitivity to noisy data. Even in the presence of a certain degree of noise interference in the data, it can still maintain relatively stable and reliable performance. Furthermore, the fusion model possesses strong generalization ability. Through the adoption of an ensemble strategy, it organically combines the advantages of multiple models, avoiding the overfitting issues that may occur in single models and reducing the risk of overfitting. As a result, it can achieve favorable application results across a wider range of data distributions and scenarios.
The fusion achieves information complementarity by concatenating or transforming intermediate features from ConvLSTM and PSO-XGBoost. Its core lies in constructing a unified feature space. An effective model fusion method directly impacts the fusion results of ConvLSTM and PSO-XGBoost. By testing different fusion approaches and comparing the MAE and RMSE of the final fused models, the optimal fusion method is selected. The optimal fusion method for ConvLSTM-PSO-XGBoost is entropy-weighted fusion. Comparison diagrams for each fusion method are shown in Figure 8.

4. Experiments and Analysis

4.1. Experimental Overview

The experiments were conducted on a PC running the Windows operating system and equipped with an AMD Ryzen 7 5800H processor (Lenovo, Beijing, China). Programming was performed using Python 3.13. To evaluate the performance of ConvLSTM-PSO-XGBoost, we compared it with seven other models.
A comparative analysis of 20 experimental trials demonstrates that the integrated model yields optimal predictive results.
When compared to linear regression (LR), RMSE decreased by 0.3268 (58.42%); MAE decreased by 0.3076 (66.93%).
When compared to Lasso Regression (Lasso), RMSE decreased by 0.3274 (58.46%); MAE decreased by 0.3080 (66.96%).
When compared to ridge regression (Ridge), RMSE decreased by 0.3281 (58.52%); MAE decreased by 0.3086 (67.00%).
When compared to ElasticNet Regression (ElasticNet), RMSE decreased by 0.3274 (58.46%); MAE decreased by 0.3080 (66.96%).
When compared to Decision Tree Regression (DTR), RMSE decreased by 0.1109 (32.29%); MAE decreased by 0.0707 (31.75%).
When compared to KNeighbors Regressor (KNR), RMSE decreased by 0.2892 (55.42%); MAE decreased by 0.2566 (62.80%).
When compared to Support Vector Regression (SVR), RMSE decreased by 0.2071 (47.10%); MAE decreased by 0.1980 (56.57%).
The deviation between the predictions of the proposed model and the actual values compared against these seven models is illustrated in Table 3.
The prediction curve of the proposed model is shown in Figure 9. The comparison between the ideal line, confidence interval, and actual prediction points in the figure demonstrates outstanding performance in regions with strong signal strength (−40 to −60 dBm).
Figure 10 presents the model evaluation metrics, highlighting that the ConvLSTM-PSO-XGBoost model achieved an R2 value of 0.944, an RMSE of 0.233, an MAE of 0.152, and an MAPE of 2.745. These results demonstrate that this model outperforms the other nine models evaluated.

4.2. Analytical Study

To comprehensively validate the effectiveness of each component within the proposed ConvLSTM-PSO-XGBoost integrated model, extensive ablation experiments were conducted. All experiments were performed under identical training conditions to ensure a fair comparison. The necessity of the hybrid architecture was evaluated by comparing the performance of individual components with that of the fully integrated system. This study assessed the respective contributions of LSTM, ConvLSTM, PSO-optimized XGBoost, and their integrated framework.
Experimental process can be described as follows:
(1) XGBoost (PSO) only: evaluate the performance of XGBoost alone when optimized by PSO (without ConvLSTM).
(2) ConvLSTM only: evaluate the performance of ConvLSTM alone.
(3) XGBoost (random parameters): evaluate XGBoost with manually set (non-optimized) hyperparameters.
(4) LSTM only: evaluate LSTM (without convolutional layers) for spatiotemporal tasks.
(5) Complete system (ConvLSTM + PSO-XGBoost Fusion): evaluate the full hybrid model.
Table 4 presents a comparative analysis of the performance across various model architectures. The ConvLSTM-PSO-XGBoost ensemble achieved the optimal performance, with a Mean Absolute Error (MAE) of 0.1446. The use of XGBoost with PSO alone yielded an MAE of 0.2609, while the standalone ConvLSTM model attained an MAE of 0.2585. Among all variants, the LSTM model exhibited the lowest performance, with an MAE of 0.3633.
Figure 11 presents the prediction outcomes of various models from the ablation study. The ablation experiments revealed several important findings. Both models are necessary: neither XGBoost nor ConvLSTM alone can achieve the performance of the ensemble, confirming the complementarity of tree-based models and sequence-based models. Additionally, PSO optimization is effective: the performance improvement from random to PSO-optimized parameters demonstrate the effectiveness of particle swarm optimization in hyperparameter tuning. Besides, hybrid model outperforms single models: the fusion of ConvLSTM and XGBoost produced the best performance, validating the hypothesis that combining different modeling paradigms can capture more comprehensive patterns.

4.3. Feature Importance Analysis

This study employs the feature importance analysis method based on XGBoost to evaluate the contribution of each feature to RSRP prediction using the gain metric. This approach calculates the reduction value of the loss function brought by each feature at every split in the decision trees, with the computational formula as follows:
Importance ( f ) = t T G a i n t N f
where Gaint represents the objective function improvement achieved by splitting on feature f in the i tree, while Nf denotes the total number of times this feature is utilized for splitting.
Table 5 displays the importance scores of the features. According to the table, the top three contributing features are LM (9.2%), ST (9.0%), and Dis (8.7%).
Figure 12 presents the results of the feature importance scoring analysis.

4.4. Performance Analysis Under Different Conditions

To evaluate the model’s generalization capability and understand its performance boundaries, we conducted a comprehensive analysis of prediction accuracy under varying operational conditions.
Table 6 presents comparative experiments that were conducted at varying speeds while maintaining a fixed position relative to the shoreline. Specifically, at distances of 20, 40, and 60 km offshore, vessel operations were carried out at speeds of 10, 18, and 25 km/h, respectively. As illustrated by the data, the model’s prediction accuracy exhibits an overall declining trend as the distance increases from 20 to 60. The R2 value decreases from 0.9494 to 0.8135, while error metrics such as RMSE, MAE, and MAPE all show an upward trajectory. Under a constant distance, an increase in speed (from 10 to 25) also contributes to a reduction in R2 and an amplification of errors. For instance, at a distance of 40, R2 declines from 0.9238 to 0.8525, and MAPE rises from 3.09% to 5.83%. Comprehensive analysis indicates that the model’s prediction accuracy is negatively correlated with both navigation distance and speed, with optimal performance observed under conditions of short distance and low speed. The experimental results are presented as follows.

4.5. Model Scoring

To compare the utility of the nine evaluation metrics, a model scoring process was designed. By categorizing metrics into positive and negative indicators, the scoring formula was designed as follows:
S P = x min ( x ) max ( x ) min ( x ) + ε × 0.8 + 0.2
S N = max ( x ) x max ( x ) min ( x ) + ε × 0.8 + 0.2
where SP is the normalization formula for positive indicators and SN is the negative indicator normalization formula (lower values yield higher scores). x represents the raw metric value, min(x) is the minimum value of that metric, max(x) is the maximum value of that metric, and ε is a small constant (e−10) to avoid division-by-zero errors.
The normalized score standard prioritizes values closest to 1.0 as optimal. Scoring results are shown in Table 7.
Following extensive verification through more than five iterations, the findings shown in Figure 13 were obtained. The figure demonstrates that the ConvLSTM-PSO-XGBoost model significantly outperforms the other nine models.

5. Conclusions

This paper proposes a ConvLSTM-PSO-XGBoost fusion model that outperforms its individual components across all metrics. Evaluation encompassed feature processing, model optimization, model assessment, and predictive performance. Results demonstrate an accuracy rate of 94.4%, confirming the fusion model’s capability to accurately predict maritime 5G signals with superior performance compared to standalone models.
However, the current study has the following limitations.
Limited dataset coverage: data were collected only in the Zhoushan Sea area of China, excluding offshore environments.
No consideration of multi-base station handover: actual vessel navigation involves base station switching, necessitating the introduction of a handover prediction mechanism.
Lack of extreme weather testing: data from extreme scenarios such as typhoons (wind speeds > 33 m/s) were not included.
Future research directions include the following:
Integrating meteorological data (e.g., wave spectra, barometric pressure changes) to enhance prediction robustness.
Introducing attention mechanisms to focus on critical features (e.g., environmental parameters during occlusion).
Developing lightweight models (e.g., model compression techniques) for edge computing scenarios.
This study proposes a variable-weight fusion model combining ConvLSTM and PSO-XGBoost for predicting 5G signal strength at sea. Validation using field data from the East China Sea demonstrates superior overall performance compared to existing mainstream methods. The findings not only provide a novel approach for marine 5G signal prediction, but also offer a reliable data source for marine channel modeling and communication.

Author Contributions

K.Y.: Writing—original draft, Conceptualization, Supervision, Resources, Project administration. J.D.: Writing—original draft, Software, Formal analysis, Data curation. L.Q.: Writing—review and editing, Methodology, Visualization, Funding acquisition. B.Z.: Software, Formal analysis. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Fund Project of Key Laboratory of Marine Environmental Survey Technology and Application, Ministry of Natural Resources [grant number MESTA-2024-B005]; General Scientific Research Fund of Zhejiang Provincial Education Department [grant number Y202560032]; “Pioneer Leading Goose + X” Science and Technology Program of Zhejiang Province [grant number 2025C02018]; Discipline Construction Program of Zhejiang Province [grant number 1106406022102]; The National College Students' Innovation and Entrepreneurship Training Program, [grant number 202510340060]; Zhoushan Science and Technology R&D Project [grant number 2024C03007]; and Ningbo Natural Science Foundation [grant number 2023J090].

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Comparison Between Standardized and Raw Data Following Data Processing

Table A1. Comparison between standardized and raw data following data processing.
Table A1. Comparison between standardized and raw data following data processing.
FeatureStateMeanStandard DeviationMinimumMaximumMedian
SurgeRaw−0.0440.013−0.0980.01−0.044
Standardized5.0921.000−4.0894.0930.001
SwayRaw−0.0510.013−0.097−0.002−0.052
Standardized4.3681.000−3.4863.821−0.025
HeaveRaw1.0120.0180.9411.0811.013
Standardized−1.9691.000−3.8363.6630.020
RollRaw−2.8860.540−4.839−0.428−2.884
Standardized−1.3191.000−3.6154.5520.004
PitchRaw2.4530.4640.8184.0262.461
Standardized−1.7561.000−3.5173.3830.017
YawRaw74.5861.63070.76382.52474.586
Standardized−5.4311.000−2.3454.868−0.001
Roll angleRaw0.01560.314−1.0991.2210.000
Standardized1.8991.000−3.5433.832−0.049
Pitch angleRaw−0.0110.505−1.8921.8920.000
Standardized1.3051.000−3.7193.7650.022
Heave angleRaw0.0030.132−0.5490.6710.000
Standardized2.1361.000−4.1635.032−0.025
LM fieldRaw20.2244.27411.23027.53922.168
Standardized0.0001.000−2.1041.7110.454
AM filedRaw7.2611.2373.3210.2547.812
Standardized−6.0771.000−3.1842.4170.444
VM fieldRaw−43.6810.937−46.973−37.793−43.457
Standardized−1.2151.000−3.5106.2780.238
STRaw35.1161.09734.138.934.5
Standardized6.5331.000−0.9263.449−0.561
SSRaw16.9018.9440.00091.12920.651
Standardized−2.2791.000−1.8898.3000.419
DisRaw10.83311.423046.2864.980
Standardized7.6001.000−0.948493.103−0.512

References

  1. Alqurashi, F.S.; Trichili, A.; Saeed, N.; Ooi, B.S.; Alouini, M.-S. Maritime Communications: A Survey on Enabling Technologies, Opportunities, and Challenges. IEEE Internet Things J. 2023, 10, 3525–3547. [Google Scholar] [CrossRef]
  2. Duangsuwan, S.; Klubsuwan, K. Underwater Drone-Enabled Wireless Communication Systems for Smart Marine Communications: A Study of Enabling Technologies, Opportunities, and Challenges. Drones 2025, 9, 784. [Google Scholar] [CrossRef]
  3. Yang, K.; Zhou, N.; Røste, T.; Neskvern, H.; Eriksen, J. LTE massive MIMO (Pre-5G) test for land-to-boat scenarios in Oslo fjord. J. Phys. Conf. Ser. 2019, 1357, 012004. [Google Scholar] [CrossRef]
  4. Zhao, Y.; Hu, J.; Yang, K.; Cui, S. Deep Reinforcement Learning Aided Intelligent Access Control in Energy Harvesting Based WLAN. IEEE Trans. Veh. Technol. 2020, 69, 14078–14082. [Google Scholar] [CrossRef]
  5. Yang, H.; Lin, K.; Xiao, L.; Zhao, Y.; Xiong, Z.; Han, Z. Energy Harvesting UAV-RIS-Assisted Maritime Communications Based on Deep Reinforcement Learning Against Jamming. IEEE Trans. Wirel. Commun. 2024, 23, 9854–9868. [Google Scholar] [CrossRef]
  6. Park, J.; Kang, M.; Lee, Y.; Jung, J.; Choi, H.-T.; Choi, J. Multiple Autonomous Surface Vehicles for Autonomous Cooperative Navigation Tasks in a Marine Environment: Development and Preliminary Field Tests. IEEE Access 2023, 11, 36203–36217. [Google Scholar] [CrossRef]
  7. Wang, J.; Zhou, H.; Li, Y.; Sun, Q.; Wu, Y.; Jin, S.; Quek, T.Q.S.; Xu, C. Wireless Channel Models for Maritime Communications. IEEE Access 2018, 6, 68070–68088. [Google Scholar] [CrossRef]
  8. Yang, K.; Lin, J.; Ding, J.; Zheng, B.; Qin, L. An Integrated Safety Monitoring and Pre-Warning System for Fishing Vessels. J. Mar. Sci. Eng. 2025, 13, 1049. [Google Scholar] [CrossRef]
  9. Parsons, J.D. The Mobile Radio Propagation Channel, 2nd ed.; John Wiley & Sons: Chichester, UK, 2000. [Google Scholar]
  10. Saunders, S.R.; Aragón-Zavala, A. Antennas and Propagation for Wireless Communication Systems, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2007. [Google Scholar]
  11. Yang, C.; Wang, J.; Ma, J. Exploration of X-Band Communication for Maritime Applications in the South China Sea. IEEE Antennas Wirel. Propag. Lett. 2022, 21, 481–485. [Google Scholar] [CrossRef]
  12. Reyes-Guerrero, J.C. Experimental Broadband Channel Characterization in a Sea Port Environment at 5.8 GHz. IEEE J. Ocean. Eng. 2016, 41, 509–514. [Google Scholar] [CrossRef]
  13. Ren, X.-L.; Chen, A.-X. Solving the VRP Using Transformer-Based Deep Reinforcement Learning. In Proceedings of the 2023 International Conference on Machine Learning and Cybernetics (ICMLC), Adelaide, Australia, 9–11 July 2023; pp. 365–369. [Google Scholar]
  14. Szeliski, R. Computer Vision: Algorithms and Applications; Springer Nature: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
  15. Danenas, P.; Skersys, T. Exploring Natural Language Processing in Model-To-Model Transformations. IEEE Access 2022, 10, 116942–116958. [Google Scholar] [CrossRef]
  16. Meng, Z.; Zhao, S.; Chen, H.; Hu, M.; Tang, Y.; Song, Y. The Vehicle Testing Based on Digital Twins Theory for Autonomous Vehicles. IEEE J. Radio Freq. Identif. 2022, 6, 710–714. [Google Scholar] [CrossRef]
  17. Chandio, A.A.; Asikuzzaman, M.; Pickering, M.R.; Leghari, M. Cursive Text Recognition in Natural Scene Images Using Deep Convolutional Recurrent Neural Network. IEEE Access 2022, 10, 10062–10078. [Google Scholar] [CrossRef]
  18. Li, S.; Deng, W. Deep Facial Expression Recognition: A Survey. IEEE Trans. Affect. Comput. 2022, 13, 1195–1215. [Google Scholar] [CrossRef]
  19. Ahmed, I.; Jeon, G.; Piccialli, F. From Artificial Intelligence to Explainable Artificial Intelligence in Industry 4.0: A Survey on What, How, and Where. IEEE Trans. Ind. Inform. 2022, 18, 5031–5042. [Google Scholar] [CrossRef]
  20. Sánchez, I.; Vallejo, F.; Játiva, P.P.; Dehghan Firoozabadi, A. An Analysis of WiFi Coverage Modeling for a Hotspot in the Parish of Checa Employing Deterministic and Empirical Propagation Models. Appl. Sci. 2024, 14, 11120. [Google Scholar] [CrossRef]
  21. Palacios, P.; Azurdia-Meza, C.; Salazar, I.; Zabala-Blanco, D.; Roman Cañizares, M. Analysis of Spectrum Detection and Decision Using Machine Learning Algorithms in Cognitive Mobile Radio Networks. In Quality, Reliability, Security and Robustness in Heterogeneous Systems. QShine 2020. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering; Springer: Cham, Switzerland, 2021. [Google Scholar] [CrossRef]
  22. Hao, J.; Ho, T.K. Machine Learning Made Easy: A Review of Scikit-Learn Package in Python Programming Language. J. Educ. Behav. Stat. 2019, 44, 348–361. [Google Scholar] [CrossRef]
  23. Gould, S.; Hartley, R.; Campbell, D. Deep Declarative Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 3988–4004. [Google Scholar] [CrossRef]
  24. Abadi, M. TensorFlow: A System for Large-Scale Machine Learning. In Proceedings of the 12th USENIX Conference on Operating Systems Design and Implementation, Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  25. Jiang, H.; He, Z.; Ye, G.; Zhang, H. Network Intrusion Detection Based on PSO-XGboost Model. IEEE Access 2020, 8, 58392–58401. [Google Scholar] [CrossRef]
  26. Peréz-Wences, C.; Loo-Yau, J.R.; Moreno, P.; Reynoso-Hernández, J.A. Digital Predistortion of RF Power Amplifier Using the NARX-SVR Model. IEEE Microw. Wirel. Technol. Lett. 2023, 33, 475–478. [Google Scholar] [CrossRef]
  27. Chaudhary, P.; Manoj, B.R.; Chauhan, I.; Bhatnagar, M. Channel Estimation Using Linear Regression with Bernoulli–Gaussian Noise. Appl. Sci. 2025, 15, 10590. [Google Scholar] [CrossRef]
  28. Jangir, G.; Purohit, G.; Joshi, N.; Yadav, A. Evaluating Machine Learning Algorithms on BCI Dataset. In Proceedings of the 2024 1st International Conference on Sustainability and Technological Advancements in Engineering Domain (SUSTAINED), Faridabad, India, 13–14 December 2024; pp. 295–300. [Google Scholar]
  29. Wang, X.; Wang, S.; Liang, X.; Zhao, D.; Huang, J.; Xu, J.; Dai, B.; Miao, Q. Deep Reinforcement Learning: A Survey. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 5064–5078. [Google Scholar] [CrossRef]
  30. Mahmud, M.; Kaiser, M.S.; Hussain, A.; Vassanelli, S. Applications of Deep Learning and Reinforcement Learning to Biological Data. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2063–2079. [Google Scholar] [CrossRef]
  31. Oroza, C.A.; Zhang, Z.; Watteyne, T.; Glaser, S.D. A Machine-Learning-Based Connectivity Model for Complex Terrain Large-scale Low-power Wireless Deployments. IEEE Trans. Cogn. Commun. Netw. 2017, 3, 576–584. [Google Scholar] [CrossRef]
  32. Peng, L.; Yang, K.; Wu, J.; Wen, C.; Peng, T.; Li, X. Machine Learning Methods Comparison for Maritime Wireless Signal Strength Prediction. Eng. Appl. Artif. Intell. 2025, 158, 111357. [Google Scholar] [CrossRef]
  33. Juwaied, A.; Jackowska-Strumillo, L. DL-HEED: A Deep Learning Approach to Energy-Efficient Clustering in Heterogeneous Wireless Sensor Networks. Appl. Sci. 2025, 15, 8996. [Google Scholar] [CrossRef]
  34. Rigas, S.; Tzouveli, P.; Kollias, S. An End-to-End Deep Learning Framework for Fault Detection in Marine Machinery. Sensors 2024, 24, 5310. [Google Scholar] [CrossRef]
  35. Han, X.; Lin, B.; Wu, N.; Wang, P.; Na, Z.; Zhang, M. Design of a Turbo-based Deep Semantic Autoencoder for Marine Internet of Things. Internet Things 2024, 28, 101393. [Google Scholar] [CrossRef]
  36. Hussain, A.; Hussain, T.; Ullah, I.; Muminov, B.; Khan, M.Z.; Alfarraj, O.; Gafar, A. CR-NBEER: Cooperative-Relay Neighboring-Based Energy Efficient Routing Protocol for Marine Underwater Sensor Networks. J. Mar. Sci. Eng. 2023, 11, 1474. [Google Scholar] [CrossRef]
  37. Shlezinger, N.; Eldar, Y.C.; Boyd, S.P. Model-Based Deep Learning: On the Intersection of Deep Learning and Optimization. IEEE Access 2022, 10, 115384–115398. [Google Scholar] [CrossRef]
  38. Liu, Z.; Chen, L.; Zhou, X.; Jiao, Z.; Guo, G.; Chen, R. Machine Learning for Time-of-Arrival Estimation Using 5G Signals in Indoor Positioning. IEEE Internet Things J. 2023, 10, 9782–9795. [Google Scholar] [CrossRef]
  39. Li, L.; Ren, H.; Cheng, Q.; Xue, K.; Chen, W.; Debbah, M.; Han, Z. Millimeter-Wave Networking in the Sky: A Machine Learning and Mean Field Game Approach for Joint Beamforming and Beam-Steering. IEEE Trans. Wirel. Commun. 2020, 19, 6393–6408. [Google Scholar] [CrossRef]
  40. Abimannan, S.; El-Alfy, E.-S.M.; Chang, Y.-S.; Hussain, S.; Shukla, S.; Satheesh, D. Ensemble Multifeatured Deep Learning Models and Applications: A Survey. IEEE Access 2023, 11, 107194–107217. [Google Scholar] [CrossRef]
  41. Meti, A.; Deepti, B.; Athreya, K.J.; Anagha, B.C.; Mohan, R. Throughput Estimation of OBSS WLANs Using Ensemble Graph Neural Networks. In Proceedings of the 2024 9th International Conference on Computer and Communication Systems (ICCCS), Xi’an, China, 19–22 April 2024; pp. 943–949. [Google Scholar]
  42. Gharghan, S.K.; Nordin, R.; Ismail, M.; Abd Ali, J. Accurate Wireless Sensor Localization Technique Based on Hybrid PSO-ANN Algorithm for Indoor and Outdoor Track Cycling. IEEE Sens. J. 2016, 16, 529–541. [Google Scholar] [CrossRef]
  43. Zhan, Z.-H.; Zhang, J.; Li, Y.; Chung, H.S.-H. Adaptive Particle Swarm Optimization (PDF). IEEE Trans. Syst. Man Cybern. 2009, 39, 1362–1381. [Google Scholar] [CrossRef]
  44. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.-Y.; Wong, W.-K.; Woo, W.-C. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. In NIPS’15: Proceedings of the 29th International Conference on Neural Information Processing Systems—Volume 1; Montreal, QC, Canada; 7–12 December 2015; MIT Press: Cambridge, MA, USA; pp. 802–810.
  45. Kong, L.; Nian, H. Fault Detection and Location Method for Mesh-Type DC Microgrid Using Pearson Correlation Coefficient. IEEE Trans. Power Deliv. 2021, 36, 1428–1439. [Google Scholar] [CrossRef]
Figure 1. PSO flowchart.
Figure 1. PSO flowchart.
Jmse 14 00377 g001
Figure 2. PSO-XGBoost flowchart.
Figure 2. PSO-XGBoost flowchart.
Jmse 14 00377 g002
Figure 3. ConvLSTM network. ∗ denotes the convolution operation. The arrow shows the sequence of operations within the neural network.
Figure 3. ConvLSTM network. ∗ denotes the convolution operation. The arrow shows the sequence of operations within the neural network.
Jmse 14 00377 g003
Figure 4. ConvLSTM-PSO-XGBoost flowchart. ∗ denotes the convolution operation. The arrow → shows the sequence of opera-tions within the neural network. Another type of arrow Jmse 14 00377 i001 shows how data moves through these stages.
Figure 4. ConvLSTM-PSO-XGBoost flowchart. ∗ denotes the convolution operation. The arrow → shows the sequence of opera-tions within the neural network. Another type of arrow Jmse 14 00377 i001 shows how data moves through these stages.
Jmse 14 00377 g004
Figure 5. Test area and 5G terminal for signal collected.
Figure 5. Test area and 5G terminal for signal collected.
Jmse 14 00377 g005
Figure 6. Pearson correlation heatmap.
Figure 6. Pearson correlation heatmap.
Jmse 14 00377 g006
Figure 7. Training curve of ConvLSTM.
Figure 7. Training curve of ConvLSTM.
Jmse 14 00377 g007
Figure 8. Comparison of 16 fusion methods.
Figure 8. Comparison of 16 fusion methods.
Jmse 14 00377 g008
Figure 9. ConvLSTM-PSO-XGBoost prediction.
Figure 9. ConvLSTM-PSO-XGBoost prediction.
Jmse 14 00377 g009
Figure 10. Model evaluation plot.
Figure 10. Model evaluation plot.
Jmse 14 00377 g010
Figure 11. Ablation study plot.
Figure 11. Ablation study plot.
Jmse 14 00377 g011
Figure 12. Feature importance plot.
Figure 12. Feature importance plot.
Jmse 14 00377 g012
Figure 13. Model comparison.
Figure 13. Model comparison.
Jmse 14 00377 g013
Table 1. PSO-Optimized XGBoost parameters.
Table 1. PSO-Optimized XGBoost parameters.
ParameterRangePhysical Meaning
max_depth[3, 10]Tree depth, controls model complexity
learning_rate[0.01, 0.3]Learning rate, balancing iteration step size
n_estimators[100, 500]Number of trees, affecting fitting capability
min_child_weight[1, 10]Minimum sample weight sum for leaf nodes
subsample[0.5, 1]Sample subsampling ratio to prevent overfitting
Table 2. Model hyperparameters.
Table 2. Model hyperparameters.
Model NameParameter
PSO-XGBoostn_estimators = 201, learning_rate = 0.03093
ConvLSTMfilters = 32, 16, 8, kernel_size = (3, 3), dropout = 0.3
Support Vector Regressionkernel = ‘rbf’, C = 1, gamma = ‘scale’, epsilon = 0.01
KNeighbors Regressorn_neighbors = 7, weights = ‘distance’, algorithm = ‘auto’, leaf_size: 30
Linear Regressionintercept_: −69.0949790794979
Lasso Regression alpha = 0.001, max_iter = 1000, tol = 0.0001
Ridge Regressionalpha = 0.001, max_iter = 1000, tol = 0.0001
ElasticNet Regressionalpha = 0.001, l1_ratio = 0.1, max_iter = 1000
Decision Tree Regressioncriterion: absolute_error, max_dept: 10, min_samples_leaf: 4, min_samples_split: 10
Table 3. Performance evaluation of 10 models.
Table 3. Performance evaluation of 10 models.
ModelR2RMSEMAEMAPE
LR0.6740.5590.4608.669
Lasso0.6730.5600.4608.688
Ridge0.6720.5610.4618.712
ElasticNet0.6730.5600.4608.687
DTR0.8770.3440.2274.043
KNR0.7160.5220.4097.796
SVR0.7980.4400.3506.465
Proposed0.9440.2330.1522.745
Table 4. Results of ablation experiments.
Table 4. Results of ablation experiments.
ModelR2RMSEMAEMAPE
PSO-XGBoost0.8830.3440.2594.835
ConvLSTM0.8650.3610.2615.174
XGBoost0.7950.4430.3536.807
LSTM0.7950.4440.3636.783
ConvLSTM-PSO-XGBoost0.9440.2330.1522.745
Table 5. The importance scores of the features.
Table 5. The importance scores of the features.
FeatureScore
LM0.0919
ST0.0903
Dis0.0873
AM0.0840
Heave Angle0.0823
SS0.0803
VM0.0780
Pitch Angle0.0722
Roll Angle0.0661
Roll0.0610
Yaw0.0559
Heave0.0498
Surge0.0446
Sway0.0372
Pitch0.0193
Table 6. Scores of the model under various conditions.
Table 6. Scores of the model under various conditions.
DistanceSpeedR2RMSEMAEMAPE
20100.9490.3440.2572.601
40180.9010.3830.2872.067
60250.8610.4170.3125.741
40100.9240.4010.3003.110
40180.8850.4470.3344.922
40250.8530.4870.3645.827
60100.8820.4580.3434.963
60180.8410.5100.3825.897
60250.8140.5560.4167.422
Table 7. Model scoring results.
Table 7. Model scoring results.
ModelR2RMSEMAEMAPE
PSO-XGBoost0.8200.7300.7240.720
ConvLSTM0.7670.6880.7180.674
Linear Regression0.2040.2030.2030.206
Lasso Regression0.2020.2020.2020.203
Ridge Regression0.2000.2000.2000.200
ElasticNet Regression0.2020.2020.2020.203
Decision Tree Regression0.8030.7300.8170.826
KNeighbors Regressor0.3290.2950.3350.323
Support Vector Regression0.5710.4950.4870.501
ConvLSTM-PSO-XGBoost1111
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ding, J.; Yang, K.; Qin, L.; Zheng, B. Prediction Model for Maritime 5G Signal Strength Based on ConvLSTM-PSO-XGBoost Algorithm. J. Mar. Sci. Eng. 2026, 14, 377. https://doi.org/10.3390/jmse14040377

AMA Style

Ding J, Yang K, Qin L, Zheng B. Prediction Model for Maritime 5G Signal Strength Based on ConvLSTM-PSO-XGBoost Algorithm. Journal of Marine Science and Engineering. 2026; 14(4):377. https://doi.org/10.3390/jmse14040377

Chicago/Turabian Style

Ding, Jianjun, Kun Yang, Li Qin, and Bing Zheng. 2026. "Prediction Model for Maritime 5G Signal Strength Based on ConvLSTM-PSO-XGBoost Algorithm" Journal of Marine Science and Engineering 14, no. 4: 377. https://doi.org/10.3390/jmse14040377

APA Style

Ding, J., Yang, K., Qin, L., & Zheng, B. (2026). Prediction Model for Maritime 5G Signal Strength Based on ConvLSTM-PSO-XGBoost Algorithm. Journal of Marine Science and Engineering, 14(4), 377. https://doi.org/10.3390/jmse14040377

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop