MS Mamba: Spectrum Forecasting Method Based on Enhanced Mamba Architecture
Abstract
1. Introduction
- Dual-SSM Mamba module with gated convolutional fusion: We design a Mamba module based on dual selective SSMs for core feature extraction, where a gated convolutional fusion mechanism integrates Mamba’s strength in long-term dependency modeling with dynamic convolution’s advantage in local feature extraction. This enables efficient extraction of multidimensional spatiotemporal–spectral features from spectrum data for precise forecasting.
- Multi-scale feature pyramid with adaptive prediction heads: We construct a feature pyramid that adapts to forecasting horizons and cooperates with prediction heads to improve both short- and long-term prediction accuracy. The pyramid dynamically matches each prediction head with appropriate levels of abstract features, effectively mitigating the adverse impact of non-stationary characteristics. A staged prediction scheme is further applied to avoid computational redundancy in reprocessing full sequences, thereby reducing resource consumption.
- Comprehensive experimental evaluation using real-world spectrum monitoring data, demonstrating significant advantages over benchmark models in both prediction accuracy and computational efficiency.
2. Related Work
2.1. Autoregressive Model-Based Spectrum Prediction
2.2. Hidden Markov Model-Based Spectrum Prediction
2.3. Deep Learning-Based Spectrum Prediction
3. Problem Formulation and Dataset Characterization
3.1. Problem Formulation
3.2. Dataset Characterization
4. Multi-Scale Mamba Model
4.1. Description of the MS-Mamba Model
4.2. Spatiotemporal Embedding Layer
4.3. The Mamba Module Composed of Dual-Selective SSM
4.4. Multi-Scale Predictive Feature Pyramid Module
Algorithm 1: The Detailed Procedures of the Proposed Multi-Scale Mamba Method |
1: Input: Sliding window spectral dataset 2: Output: Prediction results 3: Set feature embedding layer weights 4: Set time embedding layer weights 5: Initialize mamba blocks 6: Initialize feature pyramid 7: Initialize prediction heads weights 8: Set optimizer 9: for epoch e = 1 to T do 10: Forward propagation 11: 12: Calculate the loss 13: Backward propagation 14: 15: 16: Evaluate validation set: 17: Calculate MAE/RMSE per prediction length 18: Update learning rate 19: Save best model if validation loss improves 20: end 21: return |
5. Experiments and Discussion
5.1. Experimental Setup
5.2. Experimental Result Analysis on Predictive Performance
- (1)
- Convolutional feature extractors: ConvLSTM and ResNet.
- (2)
- Attention-based architectures: Transformer, Informer, Autoformer, and iTransformer.
- (3)
- Baseline: Vanilla Mamba without multi-scale pyramid or adaptive prediction heads.
5.3. Analysis of Model Efficiency
- (1)
- MS-Mamba achieves optimal prediction accuracy with minimal resource consumption.
- (2)
- It reduces memory usage by 17% and power by 70% versus iTransformer.
- (3)
- Transformers exhibit quadratic complexity in sequence length, causing excessive resource demands.
- (4)
- ConvLSTM, ResNet show low resource usage but inferior accuracy.
- (5)
- Compared to basic Mamba, MS-Mamba’s enhanced architecture increases memory by 9% and power by 22%, representing an acceptable overhead for significant accuracy gains. The selective state space model (SSM) enables linear-time computation through recursive state updates, outperforming self-attention mechanisms. Integrated multi-scale pyramids further eliminate redundant full-sequence processing. This co-design yields a computationally efficient spectrum prediction framework with balanced accuracy-throughput trade-offs.
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Li, F.; Shen, B.; Guo, J.; Lam, K.-Y.; Wei, G.; Wang, L. Dynamic Spectrum Access for Internet-of-Things Based on Federated Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2022, 71, 7952–7956. [Google Scholar] [CrossRef]
- Yucek, T.; Arslan, H. A Survey of Spectrum Sensing Algorithms for Cognitive Radio Applications. IEEE Commun. Surv. Tutor. 2009, 11, 116–130. [Google Scholar] [CrossRef]
- Ali, A.; Hamouda, W. Advances on Spectrum Sensing for Cognitive Radio Networks: Theory and Applications. IEEE Commun. Surv. Tutor. 2017, 19, 1277–1304. [Google Scholar] [CrossRef]
- Haykin, S. Cognitive Radio: Brain-Empowered Wireless Communications. IEEE J. Sel. Areas Commun. 2005, 23, 201–220. [Google Scholar] [CrossRef]
- Miuccio, L.; Panno, D.; Riolo, S. Dynamic Uplink Resource Dimensioning for Massive MTC in 5G Networks Based on SCMA. In Proceedings of the European Wireless 2019, 25th European Wireless Conference, Aarhus, Denmark, 2–4 May 2019; pp. 1–6. [Google Scholar]
- Sun, H.; Nallanathan, A.; Wang, C.-X.; Chen, Y. Wideband Spectrum Sensing for Cognitive Radio Networks: A Survey. IEEE Wirel. Commun. 2013, 20, 74–81. [Google Scholar] [CrossRef]
- Boulogeorgos, A.-A.A.; Chatzidiamantis, N.D.; Karagiannidis, G.K.; Georgiadis, L. Energy Detection under RF Impairments for Cognitive Radio. In Proceedings of the 2015 IEEE International Conference on Communication Workshop (ICCW), London, UK, 8–12 June 2015; pp. 955–960. [Google Scholar] [CrossRef]
- Xing, X.; Jing, T.; Cheng, W.; Huo, Y.; Cheng, X. Spectrum Prediction in Cognitive Radio Networks. IEEE Wirel. Commun. 2013, 20, 90–96. [Google Scholar] [CrossRef]
- Ding, G.; Jiao, Y.; Wang, J.; Zou, Y.; Wu, Q.; Yao, Y.-D.; Hanzo, L. Spectrum Inference in Cognitive Radio Networks: Algorithms and Applications. IEEE Commun. Surv. Tutor. 2018, 20, 150–182. [Google Scholar] [CrossRef]
- Suriya, M.; Sumithra, M.G. Study of Spectrum Prediction Techniques in Cognitive Radio Networks. In Proceedings of the 2022 8th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 25–26 March 2022; Volume 1, pp. 467–471. [Google Scholar]
- Pandian, P.; Selvaraj, C.; Bhalaji, N.; Arun Depak, K.G.; Saikrishnan, S. Machine Learning Based Spectrum Prediction in Cognitive Radio Networks. In Proceedings of the 2023 International Conference on Networking and Communications (ICNWC), Chennai, India, 5–6 April 2023; pp. 1–6. [Google Scholar]
- Wang, L.; Hu, J.; Zhang, C.; Jiang, R.; Chen, Z. Deep Learning Models for Spectrum Prediction: A Review. IEEE Sens. J. 2024, 24, 28553–28575. [Google Scholar] [CrossRef]
- Wen, Z.; Luo, T.; Xiang, W.; Majhi, S.; Ma, Y. Autoregressive Spectrum Hole Prediction Model for Cognitive Radio Systems. In Proceedings of the IEEE International Conference on Communications Workshops (ICC Workshops), Beijing, China, 21 May 2008; pp. 154–157. [Google Scholar]
- Gorcin, A.; Celebi, H.; Qaraqe, K.A.; Arslan, H. An Autoregressive Approach for Spectrum Occupancy Modeling and Prediction Based on Synchronous Measurements. In Proceedings of the 2011 IEEE 22nd International Symposium on Personal, Indoor and Mobile Radio Communications, Toronto, ON, Canada, 11–14 September 2011; pp. 705–709. [Google Scholar]
- Ghosh, C.; Cordeiro, C.; Agrawal, D.P.; Rao, M.B. Markov Chain Existence and Hidden Markov Models in Spectrum Sensing. In Proceedings of the 2009 IEEE International Conference on Pervasive Computing and Communications, Galveston, TX, USA, 9–13 March 2009; pp. 1–6. [Google Scholar]
- Li, Y.; Dong, Y.; Zhang, H.; Zhao, H.; Shi, H.; Zhao, X. Spectrum Usage Prediction Based on High-Order Markov Model for Cognitive Radio Networks. In Proceedings of the 2010 10th IEEE International Conference on Computer and Information Technology, Bradford, UK, 29 June–1 July 2010; pp. 2784–2788. [Google Scholar]
- Pan, G.; Wu, Q.; Zhou, B.; Li, J.; Wang, W.; Ding, G.; Yau, D.K.Y. Spectrum Prediction With Deep 3D Pyramid Vision Transformer Learning. IEEE Trans. Wirel. Commun. 2025, 24, 509–525. [Google Scholar] [CrossRef]
- Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. Proc. AAAI Conf. Artif. Intell. 2021, 35, 11106–11115. [Google Scholar] [CrossRef]
- Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting. In Proceedings of the Advances in Neural Information Processing Systems, Online, 4–14 December 2021; Curran Associates, Inc.: Sydney, NSW, Australia, 2021; Volume 34, pp. 22419–22430. [Google Scholar]
- Liu, Y.; Hu, T.; Zhang, H.; Wu, H.; Wang, S.; Ma, L.; Long, M. iTransformer: Inverted Transformers Are Effective for Time Series. Forecasting Preprint. 2024. Available online: https://arxiv.org/abs/2310.06625 (accessed on 8 July 2025).
- Gu, A.; Dao, T. Mamba: Linear-Time Sequence Modeling with Selective State Spaces 2024. Preprint. 2024. Available online: https://arxiv.org/abs/2312.00752 (accessed on 14 July 2025).
- Cai, X.; Zhu, Y.; Wang, X.; Yao, Y. MambaTS: Improved Selective State Space Models for Long-Term Time Series Forecasting. Preprint. 2024. Available online: https://arxiv.org/abs/2402.06274 (accessed on 14 July 2025).
- Luo, Y.; Wang, Y. A Statistical Time-Frequency Model for Non-Stationary Time Series Analysis. IEEE Trans. Signal Process. 2020, 68, 4757–4772. [Google Scholar] [CrossRef]
- Ozyegen, O.; Mohammadjafari, S.; Kavurmacioglu, E.; Maidens, J.; Bener, A.B. Experimental Results on the Impact of Memory in Neural Networks for Spectrum Prediction in Land Mobile Radio Bands. IEEE Trans. Cogn. Commun. Netw. 2020, 6, 771–782. [Google Scholar] [CrossRef]
- Wang, Z.; Kong, F.; Feng, S.; Wang, M.; Yang, X.; Zhao, H.; Wang, D.; Zhang, Y. Is Mamba Effective for Time Series Forecasting? Neurocomputing 2025, 619, 129178. [Google Scholar] [CrossRef]
- Akbar, I.A.; Tranter, W.H. Dynamic Spectrum Allocation in Cognitive Radio Using Hidden Markov Models: Poisson Distributed Case. In Proceedings of the Proceedings 2007 IEEE Southeast Conference, Richmond, VA, USA, 22–25 March 2007; pp. 196–201. [Google Scholar]
- Park, C.-H.; Kim, S.-W.; Lim, S.-M.; Song, M.-S. HMM Based Channel Status Predictor for Cognitive Radio. In Proceedings of the 2007 Asia-Pacific Microwave Conference, Bangkok, Thailand, 11–14 December 2007; pp. 1–4. [Google Scholar]
- Tumuluru, V.K.; Wang, P.; Niyato, D. Channel Status Prediction for Cognitive Radio Networks. Wirel. Commun. Mob. Comput. 2010, 10, 829–842. [Google Scholar] [CrossRef]
- Chen, Z.; Qiu, R.C. Prediction of Channel State for Cognitive Radio Using Higher-Order Hidden Markov Model. In Proceedings of the IEEE Southeast Conference 2010 (SoutheastCon), Charlotte, NC, USA, 18–21 March 2010; pp. 276–282. [Google Scholar]
- Wang, X.; Peng, T.; Zuo, P.; Wang, X. Spectrum Prediction Method for ISM Bands Based on LSTM. In Proceedings of the 2020 5th International Conference on Computer and Communication Systems (ICCCS), Shanghai, China, 22–24 May 2020; pp. 580–584. [Google Scholar]
- Shawel, B.S.; Woldegebreal, D.H.; Pollin, S. Convolutional LSTM-Based Long-Term Spectrum Prediction for Dynamic Spectrum Access. In Proceedings of the 2019 27th European Signal Processing Conference (EUSIPCO), Coruña, Spain, 2–6 September 2019; pp. 1–5. [Google Scholar]
- Gao, Y.; Zhao, C.; Fu, N. Joint Multi-Channel Multi-Step Spectrum Prediction Algorithm. In Proceedings of the 2021 IEEE 94th Vehicular Technology Conference (VTC2021-Fall), Norman, OK, USA, 27–30 September 2021; pp. 1–5. [Google Scholar]
- Pan, G.; Wu, Q.; Ding, G.; Wang, W.; Li, J.; Xu, F.; Zhou, B. Deep Stacked Autoencoder-Based Long-Term Spectrum Prediction Using Real-World Data. IEEE Trans. Cogn. Commun. Netw. 2023, 9, 534–548. [Google Scholar] [CrossRef]
- Zhang, X.; Guo, L.; Ben, C.; Peng, Y.; Wang, Y.; Shi, S.; Lin, Y.; Gui, G. A-GCRNN: Attention Graph Convolution Recurrent Neural Network for Multi-Band Spectrum Prediction. IEEE Trans. Veh. Technol. 2024, 73, 2978–2982. [Google Scholar] [CrossRef]
Parameters | Symbols | Values |
---|---|---|
Dataset split ratio | D | 4:1 |
Look-back window | Wl | 60 min |
Time step | Wr | 60–300 min |
Time resolution | Tr | 60 min |
Frequency resolution | F | 20 MHz |
The number of hidden layer | H | 128 |
Batch size | B | 64 |
Epoches | e | 50 |
Learning rate | 0.001 |
Model | Parameters (M) | FLOPs (G) | Inference Time (ms/Prediction) |
---|---|---|---|
MS-Mamba | 0.67 | 0.02 | 0.112 |
Mamba | 0.16 | 0.01 | 0.105 |
Informer | 1.6 | 0.09 | 0.028 |
Autoformer | 1.7 | 0.08 | 0.024 |
iTransformer | 2.23 | 0.17 | 0.115 |
Transformer | 2.27 | 0.12 | 0.017 |
Resnet | 1.3 | 0.02 | 0.018 |
ConvLSTM | 1.88 | 0.1 | 0.425 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, D.; Xu, D.; Hu, G.; Zhang, W. MS Mamba: Spectrum Forecasting Method Based on Enhanced Mamba Architecture. Electronics 2025, 14, 3708. https://doi.org/10.3390/electronics14183708
Liu D, Xu D, Hu G, Zhang W. MS Mamba: Spectrum Forecasting Method Based on Enhanced Mamba Architecture. Electronics. 2025; 14(18):3708. https://doi.org/10.3390/electronics14183708
Chicago/Turabian StyleLiu, Dingyin, Donghui Xu, Guojie Hu, and Wang Zhang. 2025. "MS Mamba: Spectrum Forecasting Method Based on Enhanced Mamba Architecture" Electronics 14, no. 18: 3708. https://doi.org/10.3390/electronics14183708
APA StyleLiu, D., Xu, D., Hu, G., & Zhang, W. (2025). MS Mamba: Spectrum Forecasting Method Based on Enhanced Mamba Architecture. Electronics, 14(18), 3708. https://doi.org/10.3390/electronics14183708