MKAIS: A Hybrid Mamba–KAN Neural Network for Vessel Trajectory Prediction
Abstract
1. Introduction
- Designed a novel fusion framework that combines the efficient long-sequence modeling capability of the structured state space model Mamba with the strong representational power of KAN, enabling joint modeling of the spatial states of ship trajectories.
- Adopted a feature-separate embedding strategy for AIS data, where longitude, latitude, speed, and course are independently modeled and embedded at the input stage. At the output stage, the four predicted variables are also computed with separate loss functions, achieving fine-grained supervision and optimization at the feature level.
- Conducted comparative experiments on real-world datasets against several mainstream models, including LSTM, BiGRU, Attention_LSTM, Transformer, and Mamba, demonstrating the proposed model’s superior prediction accuracy across multiple metrics such as the MAE, RMSE, and Haversine distance.
2. Methodology
2.1. Problem Formulation
2.2. The Proposed MKAIS Model
- Embedding Layer: This component transforms the four-dimensional AIS features into a higher-dimensional latent space. The resulting representation is combined with positional encoding, and the fused embedding serves as the input to the subsequent MKAN module.
- MKAN Module: This module is composed of two key parts: Mamba and KAN. The state space model (SSM) in Mamba is leveraged to capture intricate dependencies in vessel trajectory sequences and extract global contextual information. Meanwhile, KAN enhances the model’s nonlinear representation power. By substituting traditional activation functions with trainable kernel functions, KAN is capable of adaptively learning complex input–output mappings, thereby offering higher flexibility and accuracy in modeling the nonlinear and dynamic behaviors of vessel trajectories. In addition, due to its lightweight design, KAN can be seamlessly combined with other modules, further boosting the overall prediction performance.
- Prediction Layer: This layer applies linear projection combined with nonlinear activation to transform the dynamic variations and spatial characteristics captured by the MKAN module into the output domain, thereby facilitating accurate vessel trajectory prediction.
2.2.1. Embedding Layer
2.2.2. The MKAN Module
Mamba Component
KAN Component
Prediction Layer
Training
3. Experiment
3.1. Experimental Setup
3.1.1. Dataset
3.1.2. Evaluation Metrics
3.1.3. Baseline Models
- (1)
- LSTM [32]: the long short-term memory network (LSTM) is an improved architecture of recurrent neural networks (RNNs). It introduces memory cells to mitigate the long-term dependency challenges faced by traditional RNNs in vessel trajectory prediction (VTP).
- (2)
- BiGRU [33]: the bidirectional gated recurrent unit (Bi-GRU) model processes sequential data in both forward and backward directions with its gated recurrent structure; this bidirectional architecture enables a more comprehensive capture of contextual dependencies in vessel trajectories, allowing it to effectively learn from both past and future states for enhanced prediction accuracy.
- (3)
- Attention_LSTM [34]: based on the LSTM, this model incorporates the Transformer attention mechanism to enhance the capability of modeling complex nonlinear trajectories in AIS data, thereby overcoming the limitations of traditional approaches.
- (4)
- Transformer [27]: this model applies the Transformer architecture to AIS data, effectively addressing the limitations of traditional models (e.g., LSTM) in handling nonlinear trajectory information.
- (5)
- TrAISformer [4]: captures long-term dependencies in AIS data within a high-dimensional feature space and addresses data heterogeneity and multimodal patterns by employing a modified loss function.
- (6)
- Mamba [12]: incorporates the Mamba architecture into vessel trajectory prediction to enhance forecasting accuracy and speed up inference, effectively overcoming performance limitations found in models like the Transformer.
3.1.4. Parameter Settings
3.1.5. Hyperparameter Settings
4. Results
4.1. Experimental Results and Analysis
Analysis of Short-Term Trajectory Prediction Results
- LSTM: The LSTM model exhibits relatively poor prediction performance across different time intervals, particularly in terms of the Haversine distance. Specifically, the prediction errors are 1.5508, 2.5709, and 3.9328, which are significantly higher than those of MKAIS. This deficiency arises from the LSTM’s limited capability in handling long-term dependencies and dynamic features, leading to suboptimal performance in complex ship trajectory prediction tasks.
- Bi-GRU: The Bi-GRU model demonstrates performance comparable to the LSTM baseline but fails to achieve significant improvements. Specifically, its MAE values of 0.6254, 1.7936, and 3.6111 for 1, 2, and 3 h predictions show only marginal enhancement over LSTM’s 0.6246, 1.8279, and 3.6211. In terms of the RMSE, Bi-GRU records 0.6577, 2.9798, and 6.5840 across the three time intervals, even slightly underperforming LSTM in the 1 h prediction (0.6425 vs. 0.6577). The Haversine distance metrics of 1.5639, 2.5583, and 3.9333 further confirm this trend, showing minimal deviation from the LSTM baseline. These results indicate that while the bidirectional architecture provides additional contextual information, it offers limited advantages in handling the complex temporal dependencies present in vessel trajectory prediction, particularly in long-term forecasting scenarios.
- Attention_LSTM: The Attention_LSTM model demonstrates improvements over LSTM across different time intervals; however, certain limitations remain. Specifically, its RMSE values are 0.5969, 2.6611, and 6.5150, which, although slightly better than those of LSTM, still lag behind TrAISformer and MKAIS. These results indicate that, despite the enhancements brought by the attention mechanism, the performance of Attention_LSTM in long-term prediction remains suboptimal, with substantial deviations from the actual trajectories, suggesting that further optimization is required.
- Transformer: The Transformer model outperforms Attention_LSTM across different time intervals, though there remains room for further improvement. Specifically, the MAE values of the Transformer are 0.5930, 1.6315, and 3.3615 for 1, 2, and 3 h predictions, respectively, showing better accuracy compared to LSTM (0.6246, 1.8279, 3.6211) and Attention_LSTM (0.6085, 1.7721, 3.5974). However, its performance is still inferior to that of MKAIS (0.5652, 1.4979, 2.9397) and TrAISformer (0.5884, 1.6252, 3.2611). In terms of the Haversine distance, the Transformer yields values of 1.0661, 2.2150, and 3.6479, indicating moderate performance with noticeable deviations in long-term predictions. Overall, the Transformer demonstrates a certain advantage in accuracy, but it has not yet reached the optimal level.
- TrAISformer: The TrAISformer model demonstrates superior performance across different time intervals compared to all other models, particularly excelling in long-term prediction accuracy. Specifically, the MAE values of TrAISformer are 0.5884, 1.6252, and 3.2611 for 1, 2, and 3 h predictions, respectively, showing a significant reduction compared with LSTM (0.6246, 1.8279, 3.6211) and Attention_LSTM (0.6085, 1.7721, 3.5974), thereby indicating a clear improvement in accuracy. Nevertheless, there remains room for improvement relative to MKAIS. In terms of the Haversine distance, TrAISformer achieves 1.0289, 2.1393, and 3.6089, which, while favorable, are still less accurate than MKAIS (0.8566, 1.8770, 3.1819) in long-term predictions. This suggests that although TrAISformer exhibits outstanding accuracy, its predictive capability remains suboptimal under certain conditions.
- Mamba: The Mamba model exhibits stable performance across all evaluation metrics, yet still leaves room for further optimization. Its MAE values for 1, 2, and 3 h predictions are 0.5876, 1.6320, and 3.2570, respectively, showing improvements over LSTM (0.6246, 1.8279, 3.6211) and Attention_LSTM (0.6085, 1.7721, 3.5974), while remaining higher than those of TrAISformer (0.5884, 1.6252, 3.2611) and MKAIS (0.5652, 1.4979, 2.9397). In terms of the RMSE, Mamba achieves values of 0.5218, 2.4572, and 6.2074, outperforming LSTM and Attention_LSTM but still falling short of TrAISformer (0.5104, 2.4572, 6.2554). Regarding the Haversine distance, Mamba records values of 1.0278, 2.1589, and 3.6325, which are slightly higher than those of TrAISformer (1.0289, 2.1393, 3.6089). Although the Mamba model demonstrates relatively balanced performance, its accuracy in certain cases—particularly in long-term prediction—remains somewhat inferior to that of more advanced models.
- LSTM: For the LSTM model, the Haversine distance within 10 h increased significantly, from 1.5508 km to 21.2634 km. As the prediction horizon extends, the error continues to grow, with a particularly sharp increase observed in predictions beyond 3 h. This indicates that while LSTM performs relatively well in short-term predictions (1 h and 2 h), its accuracy declines in long-term forecasts (especially beyond 3 h), leading to substantial deviations from the true trajectories. Compared with other models, the error growth of LSTM is more pronounced, particularly when contrasted with TrAISformer and MKAIS, where the LSTM shows considerably lower accuracy in long-term predictions. Overall, the LSTM demonstrates weak capability in handling long-term dependencies, requiring further improvements to reduce Haversine distance errors in extended forecasting.
- Bi-GRU: The Bi-GRU model exhibits performance highly similar to the LSTM baseline throughout the 10 h prediction horizon, with the Haversine distance increasing from 1.5639 km to 21.3369 km. In short-term predictions (1–3 h), its errors (1.5639 km, 2.5583 km, 3.9333 km) show negligible improvement over LSTM (1.5508 km, 2.5709 km, 3.9328 km), indicating limited advantages from the bidirectional architecture. As the prediction horizon extends beyond 3 h, the error growth pattern closely mirrors that of LSTM, with both models demonstrating parallel accumulation of deviations. Particularly in medium to long-term predictions (5–10 h), the performance gap between Bi-GRU and LSTM remains minimal, suggesting that the bidirectional processing mechanism provides insufficient enhancement in capturing complex temporal dependencies for vessel trajectory prediction. While the Bi-GRU slightly outperforms LSTM at certain time points (e.g., 7 h prediction: 12.2141 km vs. 12.2271 km), the overall improvement is marginal and fails to demonstrate substantial advantages over the conventional unidirectional architecture.
- Attention_LSTM: The Attention_LSTM model exhibits relatively stable Haversine distance performance within the 10 h horizon; however, the error still shows an upward trend as the prediction time increases. The distance rises from 1.4785 km to 21.4780 km, which marks an improvement over the LSTM (1.5508 km to 21.2634 km). Nonetheless, in long-term predictions (beyond 3 h), the error still increases significantly, with the 10 h prediction error remaining comparatively high relative to the other models. The advantage of Attention_LSTM over LSTM mainly stems from the incorporation of the attention mechanism, which enhances the model’s ability to capture long-term dependencies. However, despite this improvement, the model still lacks sufficient accuracy in long-term predictions, especially after the 3 h horizon, where the error rises rapidly. Compared with TrAISformer and MKAIS, Attention_LSTM still has room for improvement, particularly in long-term dependency modeling.
- Transformer: The Transformer model demonstrates relatively superior performance in terms of the Haversine distance within the 10 h horizon, increasing from 1.0661 km to 21.5362 km. Although the error grows with longer prediction times, its growth rate is smaller compared to LSTM and Attention_LSTM. At the 1 h and 2 h intervals, the Transformer achieves significantly lower prediction errors than LSTM and Attention_LSTM, indicating higher accuracy in short-term predictions. However, as the prediction time extends to 10 h, its error gradually approaches that of the other models. Particularly in long-term predictions (beyond 3 h), the error still shows a notable increase. Despite this, the Transformer performs well in modeling long-term dependencies and effectively captures sequential relationships. Nevertheless, further optimization is required in longer time-horizon prediction tasks, particularly to address the issue of error accumulation.
- TrAISformer: The TrAISformer model demonstrates overall strong performance in 10 h predictions, though some limitations remain. Its Haversine distance increases from 1.0289 km to 18.5502 km, showing a relatively smaller error growth. However, as the prediction horizon extends, errors still accumulate, with a more noticeable increase in long-term predictions (e.g., at 10 h). Compared with MKAIS (0.8566 km to 15.3848 km), TrAISformer still has room for improvement in long-term prediction accuracy, and its short-term precision is slightly inferior. While TrAISformer effectively captures long-term dependencies and mitigates error accumulation, its short-term accuracy lags behind some other models, suggesting potential shortcomings in capturing local features over shorter intervals. Overall, TrAISformer shows clear advantages in long-horizon forecasting but still requires further optimization in short-term precision and in controlling error accumulation for very long predictions.
- Mamba: The Mamba model exhibits relatively stable performance in terms of the Haversine distance over a 10 h prediction horizon, increasing from 1.0278 km to 20.8097 km. Its errors are smaller than those of LSTM and Attention_LSTM, particularly in short-term predictions (1 h and 2 h), where it achieves higher accuracy. However, as the prediction horizon extends, errors gradually accumulate, with a more pronounced increase beyond 3 h. Compared with TrAISformer and MKAIS, Mamba shows lower accuracy in long-term predictions, especially at the 10 h mark where its Haversine distance is larger. While Mamba demonstrates strong short-term performance, its ability to capture long-term dependencies is relatively weaker, leading to faster error accumulation. Relative to more advanced models such as TrAISformer, Mamba still requires further improvements in long-horizon forecasting, particularly in mitigating error growth.
4.2. Ablation Study
- A variant of the model incorporating five Mamba layers (denoted as Mamba-5) was employed to assess its performance within the encoder.
- A model consisting of five KAN layers (KAN-5) was designed to investigate its performance without the involvement of Mamba.
- The errors of Mamba-5 and KAN-5 exhibit similar trends over the time steps, both gradually increasing as time progresses. However, compared to MKAIS, their overall errors are higher, with noticeable deviations in the later stages (e.g., at 9 and 10 h). During the initial hours (e.g., 1 to 6 h), Mamba-5 and KAN-5 perform similarly, but their errors grow significantly over time, indicating that these two models cannot effectively control error accumulation in long-term predictions.
- MKAIS consistently outperforms Mamba-5 and KAN-5 at all time steps, particularly in long-term predictions (10 h), demonstrating its advantage in capturing long-term dependencies and mitigating error accumulation.
- By integrating the Mamba module with the causal self-attention mechanism, MKAIS is able to maintain sensitivity to short-term details while providing stable long-term predictions for complex vessel trajectories, effectively preventing rapid error accumulation.
5. Conclusions
- MKAIS achieved state-of-the-art performance across multiple prediction horizons, outperforming LSTM, Attention-LSTM, Transformer, TrAISformer, and Mamba in terms of the MAE, RMSE, and Haversine distance.
- The feature-separate embedding strategy proved effective in enhancing the model’s interpretability and learning efficiency for distinct motion attributes.
- The combination of Mamba and KAN offers a promising architecture for sequence modeling tasks that require both long-term dependency capture and high-fidelity representation of complex nonlinear dynamics.
- Integration into Decision-Support Frameworks: We will explore the integration of MKAIS as a core prediction module within larger maritime intelligent systems. Specifically, its high-accuracy predictions will be used to drive collision risk assessment algorithms and COLREGs-compliant path planning tools, providing actionable recommendations for officers on the bridge or in Vessel Traffic Service (VTS) centers.
- Modeling Multi-Agent Interactions and Environment: A key next step is to extend MKAIS to a multi-agent setting that explicitly reasons about the mutual influences between vessels. Concurrently, we will incorporate real-time environmental data (e.g., wind, waves) to move from pure trajectory prediction to intention-aware and context-aware motion forecasting.
- Validation in Simulated and Real-World Environments: We plan to validate the enhanced model through collaborations, utilizing high-fidelity maritime simulators for controlled testing and seeking access to industry datasets that encompass richer operational contexts.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Liu, J.; Shi, G.; Zhu, K. Vessel trajectory prediction model based on AIS sensor data and adaptive chaos differential evolution support vector regression (ACDE-SVR). Appl. Sci. 2019, 9, 2983. [Google Scholar] [CrossRef]
- Zhang, M.; Huang, L.; Wen, Y.; Zhang, J.; Huang, Y.; Zhu, M. Short-term trajectory prediction of maritime vessel using k-nearest neighbor points. J. Mar. Sci. Eng. 2022, 10, 1939. [Google Scholar] [CrossRef]
- Zhang, X.; Liu, G.; Hu, C.; Ma, X. Wavelet analysis based hidden Markov model for large ship trajectory prediction. In Proceedings of the 2019 Chinese Control Conference (CCC), Guangzhou, China, 27–30 July 2019; IEEE: New York, NY, USA, 2019; pp. 2913–2918. [Google Scholar]
- Nguyen, D.; Fablet, R. A transformer network with sparse augmented data representation and cross entropy loss for ais-based vessel trajectory prediction. IEEE Access 2024, 12, 21596–21609. [Google Scholar] [CrossRef]
- Guo, S.; Liu, C.; Guo, Z.; Feng, Y.; Hong, F.; Huang, H. Trajectory prediction for ocean vessels base on K-order multivariate Markov chain. In Proceedings of the International Conference on Wireless Algorithms, Systems, and Applications, Tianjin, China, 20–22 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 140–150. [Google Scholar]
- Perera, L.P.; Soares, C.G. Ocean vessel trajectory estimation and prediction based on extended Kalman filter. In Proceedings of the Second International Conference on Adaptive and Self-Adaptive Systems and Applications, Lisbon, Portugal, 21–26 November 2010; Citeseer: Princeton, NJ, USA, 2010; pp. 14–20. [Google Scholar]
- Zhu, F. Ship short-term trajectory prediction based on RNN. Proc. J. Phys. Conf. Ser. 2021, 2025, 012023. [Google Scholar] [CrossRef]
- Zhang, J.; Wang, H.; Cui, F.; Liu, Y.; Liu, Z.; Dong, J. Research into ship trajectory prediction based on an improved LSTM network. J. Mar. Sci. Eng. 2023, 11, 1268. [Google Scholar] [CrossRef]
- Kalman, R.E. A new approach to linear filtering and prediction problems. J. Fluids Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef]
- Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar] [CrossRef]
- Dao, T.; Gu, A. Transformers are ssms: Generalized models and efficient algorithms through structured state space duality. arXiv 2024, arXiv:2405.21060. [Google Scholar] [CrossRef]
- Suo, Y.; Ding, Z.; Zhang, T. The mamba model: A novel approach for predicting ship trajectories. J. Mar. Sci. Eng. 2024, 12, 1321. [Google Scholar] [CrossRef]
- Xiong, C.; Li, J.; Wu, X.; Liu, D.; Wang, Q.; Gao, R.; Ruan, J. Enhancing vessel trajectory prediction with bi-state space and squeeze excitation attention. Ocean Eng. 2025, 336, 121688. [Google Scholar] [CrossRef]
- Ye, H.; Wang, W.; Zhang, X. GC-MT: A Novel Vessel Trajectory Sequence Prediction Method for Marine Regions. Information 2025, 16, 311. [Google Scholar] [CrossRef]
- Huang, Y.; Cheng, Y.; Wang, K. Trajectory mamba: Efficient attention-mamba forecasting model based on selective ssm. In Proceedings of the Computer Vision and Pattern Recognition Conference, Nashville, TN, USA, 10–17 June 2025; pp. 12058–12067. [Google Scholar]
- Lin, Y.; Liu, Y.; Zhou, Z.; Wen, H.; Zheng, E.; Guo, S.; Lin, Y.; Wan, H. Ptrajm: Efficient and semantic-rich trajectory learning with pretrained trajectory-mamba. arXiv 2024, arXiv:2408.04916. [Google Scholar]
- Ren, Y.; Liu, L.; Lan, Z.; Cui, Z.; Yu, H. MLB-Traj: Map-Free Trajectory Prediction with Local Behavior Query for Autonomous Driving. IEEE Internet Things J. 2025, 12, 28556–28571. [Google Scholar] [CrossRef]
- Xiao, Z.; Hu, H.; Xu, G.; He, J. Tame: Temporal audio-based mamba for enhanced drone trajectory estimation and classification. In Proceedings of the ICASSP 2025—2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India, 6–11 April 2025; IEEE: New York, NY, USA, 2025; pp. 1–5. [Google Scholar]
- Zhang, W.; Zhu, X.; Zhang, J.; Guo, Y.; Tao, J.; Zhu, M. MCFM: Multi Channel-Frequency Mamba-Based Model for Flight Trajectory Prediction. In Proceedings of the International Conference on Pattern Recognition, Kolkata, India, 1–5 December 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 417–430. [Google Scholar]
- Liu, B.; Chen, C.; Li, J.; Yu, G.; Song, H.; Liu, X.; Cui, J.; Zhang, H. TrackingMiM: Efficient Mamba-in-Mamba Serialization for Real-time UAV Object Tracking. arXiv 2025, arXiv:2507.01535. [Google Scholar]
- Wu, H.; Yang, Y.; Xu, H.; Wang, W.; Zhou, J.; Zhu, L. Rainmamba: Enhanced locality learning with state space models for video deraining. In Proceedings of the 32nd ACM International Conference on Multimedia, Melbourne, Australia, 28 October–1 November 2024; pp. 7881–7890. [Google Scholar]
- Liu, Z.; Wang, Y.; Vaidya, S.; Ruehle, F.; Halverson, J.; Soljačić, M.; Hou, T.Y.; Tegmark, M. Kan: Kolmogorov-arnold networks. arXiv 2024, arXiv:2404.19756. [Google Scholar]
- Jiang, D.; Shi, G.; Li, N.; Ma, L.; Li, W.; Shi, J. TRFM-ls: Transformer-based deep learning method for vessel trajectory prediction. J. Mar. Sci. Eng. 2023, 11, 880. [Google Scholar] [CrossRef]
- Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
- Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
- Kingma, D.P. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Takahashi, K.; Zama, K.; Hiroi, N.F. Ship trajectory prediction using AIS data with TransFormer-based AI. In Proceedings of the 2024 IEEE Conference on Artificial Intelligence (CAI), Singapore, 25–27 June 2024; IEEE: New York, NY, USA, 2024; pp. 1302–1305. [Google Scholar]
- Park, J.; Jeong, J.; Park, Y. Ship trajectory prediction based on bi-LSTM using spectral-clustered AIS data. J. Mar. Sci. Eng. 2021, 9, 1037. [Google Scholar] [CrossRef]
- Rao, S.J. Regression modeling strategies: With applications to linear models, logistic regression, and survival analysis. J. Am. Stat. Assoc. 2003, 98, 257–258. [Google Scholar] [CrossRef]
- Judge, G.G.; Griffiths, W.E.; Hill, R.C.; Lütkepohl, H.; Lee, T.C. The Theory and Practice of Econometrics; John wiley & sons: Hoboken, NJ, USA, 1991. [Google Scholar]
- Inman, J. Navigation and Nautical Astronomy: For the Use of British Seamen; F. and J. Rivington: London, UK, 1849. [Google Scholar]
- Tang, H.; Yin, Y.; Shen, H. A model for vessel trajectory prediction based on long short-term memory neural network. J. Mar. Eng. Technol. 2022, 21, 136–145. [Google Scholar] [CrossRef]
- Wang, C.; Ren, H.; Li, H. Vessel trajectory prediction based on AIS data and bidirectional GRU. In Proceedings of the 2020 International Conference on Computer Vision, Image and Deep Learning (CVIDL), Chongqing, China, 10–12 July 2020; IEEE: New York, NY, USA, 2020; pp. 260–264. [Google Scholar]
- Liu, T.; Ma, J. Ship navigation behavior prediction based on AIS data. IEEE Access 2022, 10, 47997–48008. [Google Scholar] [CrossRef]










| Category | Model | Long-Sequence Handling Capability | Resource Consumption | AIS Modeling Capability |
|---|---|---|---|---|
| Traditional Neural Networks | RNN, BP Neural Network | Weak | Low | Low |
| Recurrent Neural Network Variants | LSTM, GRU | Weak | Medium | Medium |
| Attention Mechanism Networks | Transformer, Informer | Medium | High | High |
| State Space Models | Mamba, S4 | Medium | Low | Medium |
| Item | Description |
|---|---|
| Dataset | ct_dma |
| Data Source | https://www.dma.dk/safety-at-sea/navigational-information/ais-data |
| Temporal Coverage | 1 January to 31 March 2019 |
| Geographical Range | Latitude: 55.5°–58.0°, Longitude: 10.3°–13.0° |
| Total Samples | 11,888 |
| Training Set | 1 January–10 March |
| Validation Set | 11 March–20 March |
| Testing Set | 21 March–31 March |
| Parameter | Value |
|---|---|
| Batch Size | 16 |
| Historical Sequence Length (P) | 30 time steps (5 h) |
| Learning Rate | |
| Training Epochs | 20 |
| Prediction Horizons | 6, 12, 18 steps (1, 2, 3 h) |
| Number of Encoding Layers | 5 |
| Mamba SSM State Dimension | 16 |
| KAN Grid Size | 5 |
| KAN Spline Order | 3 |
| Model | MAE | RMSE | Haversine | ||||||
|---|---|---|---|---|---|---|---|---|---|
| 1 h | 2 h | 3 h | 1 h | 2 h | 3 h | 1 h | 2 h | 3 h | |
| LSTM | 0.6246 | 1.8279 | 3.6211 | 0.6425 | 2.9815 | 6.5990 | 1.5508 | 2.5709 | 3.9328 |
| Bi-GRU | 0.6254 | 1.7936 | 3.6111 | 0.6577 | 2.9798 | 6.5840 | 1.5639 | 2.5583 | 3.9333 |
| Attention_LSTM | 0.6085 | 1.7721 | 3.5974 | 0.5969 | 2.6611 | 6.5150 | 1.4785 | 2.4798 | 3.9340 |
| Transformer | 0.5930 | 1.6315 | 3.3615 | 0.5004 | 2.4772 | 6.2716 | 1.0661 | 2.2150 | 3.6479 |
| TrAISFormer | 0.5884 | 1.6252 | 3.2611 | 0.5104 | 2.4572 | 6.2554 | 1.0289 | 2.1393 | 3.6089 |
| Mamba | 0.5876 | 1.6320 | 3.2570 | 0.5218 | 2.4574 | 6.2074 | 1.0278 | 2.1589 | 3.6325 |
| MKAlS | 0.5652 | 1.4979 | 2.9397 | 0.4897 | 2.3489 | 5.6773 | 0.8566 | 1.8770 | 3.1819 |
| Improved (%) | 3.81% | 7.83% | 9.74% | 2.13% | 4.41% | 10.03% | 16.65% | 12.26% | 11.83% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Xiong, C.; Li, J.; Zhuang, Y.; Wu, X.; Luo, M.; Wang, Q. MKAIS: A Hybrid Mamba–KAN Neural Network for Vessel Trajectory Prediction. J. Mar. Sci. Eng. 2025, 13, 2119. https://doi.org/10.3390/jmse13112119
Xiong C, Li J, Zhuang Y, Wu X, Luo M, Wang Q. MKAIS: A Hybrid Mamba–KAN Neural Network for Vessel Trajectory Prediction. Journal of Marine Science and Engineering. 2025; 13(11):2119. https://doi.org/10.3390/jmse13112119
Chicago/Turabian StyleXiong, Caiquan, Jiaming Li, Yuzhe Zhuang, Xinyun Wu, Mao Luo, and Qi Wang. 2025. "MKAIS: A Hybrid Mamba–KAN Neural Network for Vessel Trajectory Prediction" Journal of Marine Science and Engineering 13, no. 11: 2119. https://doi.org/10.3390/jmse13112119
APA StyleXiong, C., Li, J., Zhuang, Y., Wu, X., Luo, M., & Wang, Q. (2025). MKAIS: A Hybrid Mamba–KAN Neural Network for Vessel Trajectory Prediction. Journal of Marine Science and Engineering, 13(11), 2119. https://doi.org/10.3390/jmse13112119
