Next Article in Journal
A Movement Classification of Polymyalgia Rheumatica Patients Using Myoelectric Sensors
Next Article in Special Issue
Two-Layer Edge Intelligence for Task Offloading and Computing Capacity Allocation with UAV Assistance in Vehicular Networks
Previous Article in Journal
Machine Learning-Assisted Speech Analysis for Early Detection of Parkinson’s Disease: A Study on Speaker Diarization and Classification Techniques
Previous Article in Special Issue
Device-Free Tracking through Self-Attention Mechanism and Unscented Kalman Filter with Commodity Wi-Fi
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Long-Term Joint Temporal–Spectral Network for Spectrum Prediction

School of Electronic and Communication Engineering, Sun Yat-sen University, Shenzhen 518107, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(5), 1498; https://doi.org/10.3390/s24051498
Submission received: 27 January 2024 / Revised: 21 February 2024 / Accepted: 22 February 2024 / Published: 26 February 2024

Abstract

:
Spectrum prediction is a promising technique to release spectrum resources and plays an essential role in cognitive radio networks and spectrum situation generating. Traditional algorithms normally focus on one-dimensional or predict spectrum values in a slot-by-slot manner and thus cannot fully perceive the spectrum states in complex environments and lack timeliness. In this paper, a deep learning-based prediction method with a simple structure is developed for temporal–spectral and multi-slot spectrum prediction simultaneously. Specifically, we first analyze and construct spectrum data suitable for the model to simultaneously achieve long-term and multi-dimensional spectrum prediction. Then, a hierarchical spectrum prediction system is developed that takes advantage of the advanced Bi-ConvLSTM and the seq2seq framework. The Bi-ConvLSTM captures time–frequency characteristics of spectrum data, and the seq2seq framework is used for long-term spectrum prediction. Furthermore, the attention mechanism is used to address the limitations of the seq2seq framework that compresses all inputs into fixed-length vectors, resulting in information loss. Finally, the experimental results have shown that the proposed model has a significant advantage over the benchmark schemes. Especially, the proposed spectrum prediction model achieves 6.15%, 0.7749, 1.0978, and 0.9628 in MAPE, MAE, RMSE, and R 2 , respectively, which is better than all the baseline deep learning models.

1. Introduction

The electromagnetic spectrum, as a critical national strategic resource all over the world, has drawn more and more attention [1,2]. With the development of information technology, the spectrum demand also explodes, but spectrum utilization is low at the same time. As pointed out by the Federal Communications Commission (FCC), numerous allocated spectrum resources are idle to a large extent in time and space [3,4], and the average utilization rate does not exceed 5% at any time and place [5]. In addition, as the spectrum sensing capability of monitoring devices is limited, spectrum monitoring data are very sparse or even scarce in multiple dimensions, such as time, space, and the frequency domain. So, it is difficult to form a detailed and comprehensive spectrum situation, which results in great challenges regarding the full use of the spectrum. Therefore, how to fully grasp the spectrum situation and further improve spectrum utilization are essential to solve the shortage of spectrum resources in the complex electromagnetic environment.
Cognitive radio networks (CRNs) are believed to be an important and promising method to relieve the shortage, whose key technologies include spectrum sensing, spectrum decision, spectrum sharing, spectrum prediction, and spectrum shifting. Figure 1 shows the working process of CRNs; firstly, spectrum sensing, as a key technology of CRNs, can obtain the spectrum state effectively [6,7] to improve spectrum utilization through a large number of continuous detections.
Secondly, the spectrum decision selects the best hole for access and realizes the spectrum shifting according to hole characteristics by spectrum sensing. Then, the spectrum decision provides channel capacity information for spectrum sharing. However, as users increase, the time and energy consumed by spectrum detection increase sharply, which could cause spectrum holes. To solve the above problem, spectrum prediction, as another key technology of CRNs and spectrum situation generation [8], has abstracted increasing attention.
Spectrum prediction utilizes historical spectrum sensing information to mine the occupancy pattern of each channel in the time domain and the potential correlation between channel states and then makes a prediction of channel occupancy in the future time slots. It can quickly find the spectrum holes, guide the spectrum sensing in the next moment and the subsequent spectrum decisionmaking, reduce the amount of spectrum switching, and improve spectrum utilization.
The research on spectrum prediction has made great progress during the past few years. Several traditional spectrum prediction techniques [9], such as regression model [10], Markov model [11], Bayesian inference [12], support vector machine [13], artificial neural network [14], and matrix/tensor completion [15,16], have been proposed for spectrum prediction. Although the traditional methods can improve spectrum utilization to some extent, they are unable to meet the actual needs.
Recently, deep learning has shown its promising capability in spectrum prediction, with high precision, strong robustness, and adaptability. Recurrent neural network (RNN)-based methods (Long Short-Term Memory (LSTM) and gated recurrent unit network (GRU)) [17] are capable of mining underlying temporal correlations among spectrum data. In [18], LSTM was employed to simultaneously predict the Radio Spectrum State (RSS) for two time slots, which requested a large amount of computing resources and suffers from very long training time. To solve the above problems, Ling Yu et al. [19] introduced the Taguchi method and LSTM for time domain spectrum prediction, effectively reducing the time and computational power requirements. Xue Wang et al. [20] used the Back Propagation-LSTM (BP-LSTM) network model for spectrum prediction, which has better prediction performance than BP, LSTM, and GRU. Nevertheless, the spectrum states are interrelated in the time and frequency domains. MA Aygül et al. [21,22] exploited correlation over time and frequency of spectrum data through a 2D-LSTM with better performance than 1D-LSTM, but this method had a weak ability to extract frequency relationships.
To address the aforementioned issues, hybrid networks have good performance in extracting temporal–spectral features of spectrum data by taking advantage of different networks. Lixing Yu et al. [23] combined convolutional neural network (CNN) with the GRU for temporal–spectral spectrum prediction with high prediction accuracy. In [24], a scheme was formulated with CNN-LSTM for multi-dimensional spectrum prediction. Xiaojin Ding et al. [25] combined CNN and bidirectional LSTM (BiLSTM) for predicting the state of multi-channels with higher accuracy than LSTM, BiLSTM, and CNN-LSTM. Graph convolution network (GCN) has a stronger ability to extract spatial features than CNN. Han Zhang et al. [26] designed a graph network model combining GCN and LSTM for multi-channel spectrum prediction that had better predictive performance compared with other methods. To improve prediction accuracy, Xile Zhang et al. [27] proposed a multi-band spectrum prediction method based on attention graph convolutional recurrent neural networks (A-GCRNN), which applied temporal correlation and frequency band correlation to spectrum prediction tasks. These methods have good performance for multi-dimensional spectrum prediction but predict spectrum values in a slot-by-slot manner, thus lacking timeliness.
Overall, simultaneously achieving multi-dimensional and long-term spectrum prediction is a great challenge. Multi-dimensional and long-term studies on effective prediction algorithms are relatively few. Ling Yu et al. [28] constructed a temporal–spectral residual network for multi-slot high-frequency (HF) band prediction from an image inference, but the structure of this method is complex, and selecting data with different time trends is a major challenge. In this study, to solve the above challenging problems, we propose a model with a simple structure for long-term and temporal–spectral spectrum prediction simultaneously. Firstly, we propose a spectrum matrices construction method that contains values of multiple channels at different time slots. Secondly, the seq2seq model is used for long-term spectrum prediction. It performs better in multi-step prediction by considering sequence dependencies between output labels. In addition, it also has a flexible framework. Thirdly, we apply a Bi-ConvLSTM as an encoder to mine the correlations of spectrum values across different channels and extract temporal features within a certain time window. Bi-ConvLSTM is an improved network based on LSTM, which not only enables temporal modeling but also characterizes spectral characteristics [29]. Although the decoder in seq2seq is greatly affected by the length of the intermediate vector, it can be relieved by introducing the attention mechanism. The attention mechanism can solve the limitation of the seq2seq structure, which encodes the constructed spectrum data into a sequence of vectors. In addition, we notice that the correlation between the spectrum states is not equal. Thus, the attention mechanism is also used to assign weights for different spectrum points, which improves the model prediction accuracy. Finally, we utilize Bi-ConvLSTM and dense networks as a decoder, which can adaptively focus on certain parts of spectrum states.
Our main contributions in this paper are the following:
  • We propose a spectrum data construction method with a simple structure to make multi-dimensional and long-term spectrum predictions simultaneously. Different from the existing spectrum prediction in a slot-by-slot manner, the proposed approach is more efficient and can predict multi-slot spectrum states ahead of multiple spectrum points.
  • We combine Bi-ConvLSTM and seq2seq to construct the proposed networks that can achieve both temporal–spectral and long-term spectrum prediction.
  • Validated on real-world datasets, the experimental results show that our proposed spectrum prediction model achieves 6.15%, 0.7749, 1.0978, and 0.9628 in mean absolute percentage error (MAPE), mean absolute error (MAE), root mean square error (RMSE), and R-squared ( R 2 ), respectively, which is better than all the baseline deep learning models. Furthermore, the designed model is robust against missing spectrum data.
The remainder of this paper is organized as follows. Section 2 compares the differences between the proposed prediction model and traditional prediction models. Section 3 presents the structure of the long-term joint temporal–spectral network. Section 4 shows the experimental results of prediction on real-world spectrum datasets. The Section 5 draws some conclusions.

2. Problem Formulation

As shown in Figure 2, most traditional prediction models learn the inherent relationship ( T 1 ) consecutive column vectors { χ 1 , χ 2 , …, χ T 1 } to predict the spectrum values χ T at time T, and each column vector represents spectrum states of F points. Window with fixed length moves forward slot by slot over time and spectrum values. Subsequently, states and qualities of spectrum points in different time slots can be predicted. Considering the actual demand and forecasting timeliness, the prediction model needs to simultaneously predict the values with acceptable error within a relatively long period. Figure 3 shows the principle of the proposed spectrum prediction model; the input and output of the prediction model are { χ t n , χ t n + 1 , …, χ t }, { χ ^ t n + 1 , χ ^ t n + 2 ,…, χ ^ t + 1 }, respectively.
Spectrum data do not exist independently and have closely intrinsic relationships in various dimensions. Therefore, our study utilizes the correlation of spectrum data in the time–frequency domain to research spectrum prediction. The F T grid diagram is constructed with a spectrum value of m spectrum points to every time slot for the proposed spectrum prediction model, as shown in Figure 4. The rows and columns of the grid diagram represent spectrum point m and time slot t, respectively. There is a correlation between rows and columns in the graph in terms of frequency and time, respectively, and adjacent time slots and channels have a stronger correlation. The spectrum state of m spectrum point in each T time slot is χ . The process of long-term joint temporal–spectral spectrum prediction can be represented as follows:
χ t n , χ t n + 1 , , χ t 2 , χ t 1 , χ t χ t + 1

3. Deep Long-Term Joint Temporal–Spectral Network

In this section, we first analyze the proposed model. To achieve time–frequency joint prediction, the proposed model adopts Bi-ConvLSTM to bidirectionally extract temporal and spectral features of historical spectrum data, which can improve accuracy. In addition, the proposed model also adopts seq2seq framework so that the inputs and outputs are matrices to achieve multi-step prediction. The attention mechanism in proposed model assigns weights based on the correlations between different frequency points. Therefore, the proposed model can achieve long-term and temporal–spectral spectrum prediction simultaneously and improve the spectrum prediction performance. We conduct spectrum prediction by developing a hierarchical deep learning framework that consists of an input layer, an encoder layer, a decoder layer, and an output layer, as shown in Figure 5. The specific processing and forecasting process is as follows.
(1) Input layer: Spectrum prediction models can be roughly divided into binary prediction and power level prediction [30]. Different spectrum prediction types have different meanings in specific scenarios. In this study, power spectral density (PSD) is chosen to predict. First, we construct the spectrum matrix for the proposed model. Specifically, denote by χ t all spectrum points’ values of time length T. The current and previous spectrum values are represented as { χ t n , χ t n + 1 , …, χ t }.
(2) Encoder layer: Second, the constructed spectrum data of the input layer { χ t n , χ t n + 1 , …, χ t } are sent to the encoder layer. The encoder layer consists of a Bi-ConvLSTM network and an attention mechanism, which encodes the input data into a vector sequence. The Bi-ConvLSTM contains two convolutional layers and two max-pooling layers in our proposed model. Different from other methods, bidirectional extracts regional temporal and spectral correlation features simultaneously and improves the proposed model’s performance. The structure of the Bi-ConvLSTM and its internal structure are shown in Figure 6 and Figure 7, respectively.
Bi-ConvLSTM: Bi-ConvLSTM achieves forward hidden layer state h t and backward hidden layer state h t that are opposite along the time axis by forward convLSTM and backward convLSTM. The forward convLSTM and backward convLSTM obtain information about the spectrum data along the time axis in the past and future, respectively. The key equations that define the Bi-ConvLSTM network for a given input χ t are provided as follows:
h t = C o n v L S T M ( h t 1 , χ t , c t 1 )
h t = C o n v L S T M ( h t + 1 , χ t , c t + 1 )
h t = [ h t , h t ]
f t = σ W χ f χ t + W h f h t 1 + W c f c t 1 + B f i t = σ W χ i χ t + W h i h t 1 + W c i c t 1 + B i c ˜ t = h W χ c χ t + W h c h t 1 + B c c t = i t c ˜ t + f t c t 1 o t = σ W χ o χ t + W h o h t 1 + W c o c t + B o h t = o t c t
where χ t represents input spectrum data. W χ f , W χ i , and W χ o are weight matrices of input, hidden state cell, and memory cell in forgotten gate, respectively. W h f , W h i , and W h f are weight matrices of input, hidden state cell, and memory cell in input gate, respectively. W c f , W c i , and W c o are the weight matrices of input, hidden state cell, and a memory cell in the output gate, respectively. B f , B i , B c , and B 0 are bias of forgotten gate, input gate, current state, and output gate, respectively. c t 1 and c ˜ t are long-term memory and current memory. h t is hidden state, which contains h t and h t . ⊙ and * are Hadamard products and convolutional operators, respectively. σ , h are s i g m o i d and t a n h activation functions, respectively.
Attention Mechanism: The relationships of each spectrum point are set equally in the traditional seq2seq model. However, the degree of correlation between channels is different. So, it is not appropriate to set an equal weight value for each set of input spectrum data. Therefore, the attention mechanism is employed to solve the above drawbacks and reserves more information of input spectrum data. The attention layer in this proposed structure consists of two parts for each input, the outputs of all Bi-ConvLSTM units in the encoder and the input state of a corresponding Bi-ConvLSTM in the decoder. In particular, the internal structure of the attention layer is shown in Figure 8. The alignment model e i j is calculated by Dot. s i is the hidden state of the neural network in the decoder at time i. h j is the j-th corresponding hidden vector of the input sequence in the encoder.
(3) Decoder layer: Then, the vectors generated by the encoder layer are sent to the decoder layer. The decoder layer consists of a Bi-ConvLSTM network and a dense network. It adaptively selects a subset of these vectors as the decoding result based on the intermediate vectors obtained by the encoder and the historical parameters in the network. Similarly, the Bi-ConvLSTM of the decoder layer contains two convolutional layers and two max-pooling layers.
A Dense Network: The dense network is employed to convert the vector dimension of output from the Bi-ConvLSTM of the decoder layer to the dimension of the final predicted value, which has several layers of fully connected neural nodes. Specifically, the dense network consists of three full layers of fully connected neural nodes in our proposed model.
(4) Output layer: Finally, the output of the dense layer is sent to an activation function to obtain a prediction result. The fina output is χ t + 1 , which donates future values of all spectrum points at multiple time slots.
Activation function and loss function: In particular, the prediction result at a time slot T and the real spectrum states at its next slot are sent to a loss function to predict spectrum values at the next time slot by updating the gradients calculated by the loss function. The gradients are sent to the Bi-ConvLSTM network and dense network for updating their weight matrices to decrease the distance between the predicted result and the ground truth. To find a final prediction result, we choose a t a n h function as the activation function that takes the output of the dense network as the input that maps the prediction result into a vector of elements between −1 and 1. In addition, we choose the mean squared error (MSE) as the loss function of the proposed model, which is widely applied in deep learning applications.
Training algorithm: The proposed spectrum prediction model is trained by minimizing MSE between the true value and the predicted value as
L ( w , b ) = χ t + 1 χ ^ t + 1 2
where w , b are all learn-able parameters in the proposed spectrum prediction model.
Algorithm 1 outlines the proposed model training process. Firstly, training sequences are constructed from the original spectrum data (lines 1–3). Then, forward propagation and back-propagation are repeatedly applied to train the model (lines 4–14).
Algorithm 1 The proposed model training algorithm.
  • Input: Historical spectrum data sequence
  • Output: Trained long-term joint temporal–spectral network
  • //Construct the dataset
  • 1: construct spectrum matrix χ t n , χ t n + 1 , , χ t 2 , χ t 1 , χ t from historical spectrum data
  • 2: D ϕ
  • 3: divide set D into train set D t r a i n and test set D t e s t
  • 4: //Train the proposed prediction model
  • 5: initialize all learn-able parameters in long-term joint temporal–spectral network
  • 6: Repeat
  • 7:      randomly select a batch of instances from D t r a i n
  • 8:      find w , b by minimizing the objective (6)
  • 9: Until the training epochs are met
  • 10: //Test the proposed prediction model
  • 11: for each sample in D t e s t  do
  • 12:      fed into the trained proposed model
  • 13:      output the prediction results of that sample
  • 14: end for

4. Experiments

4.1. Settings

Dataset Description: In this section, we design some comparative experiments to evaluate the performance of the proposed spectrum prediction method. The datasets used in our spectrum prediction experiments are gathered from RWTH Aachen University in Germany [31]. These datasets are available at https://github.com/chengrunmeng/Aachen-spectrum-data-part accessed on 2 July 2023. Specifically, the data are about spectrum states measured at the residential area in Maastricht. The datasets consist of two sub-bands with 770 MHz and 3750 MHz central frequencies and are named dataset 1 and dataset 2, respectively. Each sub-band has a bandwidth of 1500 MHz and a frequency resolution of 200 kHz.
Hyper-parameters: The settings of the proposed model are as follows. ConvLSTM2D 1 is composed of 5 identical convolution kernels with the size 5 × 5, while ConvLSTM2D 2 consists of 10 kernels with the size 3 × 3. Particularly, the convolution kernels slide over full rows of the input data matrix to extract spectral correlation features at a multi-time slot. The number of hidden units is 128. Adam is used as the optimizer. The number of epochs and initial learning rate are set as 200 and 0.001, respectively. The batch size is 32. The length of time is 30, which is the most recent data that we consider as input to predict the values of all the spectrum points at the 30 time slots. We conduct extensive experiments to test the proposed model by varying the time length ranging from 10 to 50 and find that setting 30 can achieve the best performance. MSE is used as a loss function for proposed model training and testing. The prediction time slots are 30. In addition, 90% of the spectrum data are used to train the model and treat the remaining 10 % as the test set.
Baselines and Evaluation Metrics: For comparison, 10 prediction models based on deep learning are considered as baselines, which are LSTM [18], GRU [32], BiLSTM [33], CNN [34], CNN-LSTM [35], CNN-BiLSTM [25], CNN-BiLSTM-attention [36], seq2seq-LSTM [37], seq2seq-LSTM-attention [38], and seq2seq-Bi-ConvLSTM [39]. In order to fairly compare the performance of models, the input information and output information of all prediction models are all the same and the baseline models use the multi-step principle to output results. Moreover, to evaluate the performance of the designed model, MAPE, MAE, RMSE, and R 2 are used to evaluate the performance. Especially, the calculation process is as follows:
M A P E = 100 % N k = 1 N y ^ k y k y k
M A E = 1 N k = 1 N y k y ^ k
R M S E = 1 N k = 1 N ( y k y ^ k ) 2
R 2 = 1 k = 1 N ( y k y ^ k ) 2 k = 1 N ( y k y ¯ k ) 2
where y k is the ground truth, y ^ k is the prediction value, and y ¯ k is the average value of the ground truth.

4.2. Experimental Results and Discussion

Effect of The Time Length: Choosing the appropriate time length is crucial for improving model performance.
So, we first explore the effect of time length on the performance of the proposed model and select the most appropriate values as parameters. Figure 9 shows the impact of the time length on the performance of the proposed model. In this experiment example, we vary the time length and measure the accuracy achieved by the designed model. The value of the accuracy metric increases with the increase in the time length. When the time length is 30, the accuracy is highest. Then, the value is decreasing when the time length grows. This result shows that the best time length is about 30 for our designed model and the used dataset.
Baseline Model Comparison: Figure 10 shows the performance of the designed model regarding dataset 1 and dataset 2, respectively. From Figure 10, we can see that the designed model can achieve better performance than CNN-BiLSTM-attention and seq2seq-LSTM-attention on two spectrum datasets. Furthermore, from the above result, it can also be seen that the proposed model has good generalization ability. The reason for choosing the above two models for comparison is as follows. According to the results of the literature [25], CNN-BiLSTM-attention achieves better performance for time–frequency spectrum prediction than LSTM, BiLSTM, GRU, and CNN-LSTM. In addition, seq2seq-LSTM-attention has good performance on long-term time series prediction.
Table 1 demonstrates the performance improvement by comparing the designed model with ten baseline models. Firstly, the spectral and temporal correlation features are hidden in the spectrum data. We observe that the learning models with the CNN are better than that only considering the temporal correlations. Secondly, hybrid models (such as CNN-LSTM) perform better than single models (such as LSTM, GRU, BiLSTM, and CNN), which is because hybrid models take advantage of different networks. Then, the models with bidirectional structures (such as BiLSTM and CNN-BiLSTM) perform better than one-directional models (such as LSTM and CNN-LSTM), which is due to the fact that the models with bidirectional structures can extract bidirectional features of spectrum data. While these models perform better, they are more complex in structure and more difficult to train. In addition, the models with an attention mechanism (such as CNN-BiLSTM-attention and seq2seq-LSTM-attention) perform better than the models without an attention mechanism (such as CNN-BiLSTM and seq2seq-LSTM), which is because the attention mechanism focuses on the correlation between spectrum values of different frequency points at each moment in time to assign weights for the vector set. Finally, the proposed model, advanced Bi-ConvLSTM and seq2seq structure, can achieve the best performance regarding MAE, MAPE, RMSE, and R 2 . Take the MAE as an example: LSTM, GRU, BiLSTM, CNN, CNN-LSTM, CNN-BiLSTM, CNN-BiLSTM-attention, seq2seq-LSTM, seq2seq-LSTM-attention, seq2seq-Bi-ConvLSTM, and seq2seq-Bi-ConvLSTM-attention achieve MAE values of 0.8982, 0.8847, 0.8796, 0.8679, 0.8533, 0.8498, 0.8345, 0.8556, 0.8436, 0.8064, and 0.7749, respectively. So, the proposed model can effectively reduce the amount of spectrum switching and improve the spectrum utilization through simultaneous multi-step and joint time–frequency prediction.
The Stability of the Proposed Model: Figure 11 demonstrates the stability of the designed model when using it in datasets with missing data. In particular, we use the spectrum data with different missing rates to evaluate the performance of the designed model.
Firstly, the performance of all the prediction models decreases as the rate of missing data increases, which is because, the greater the rate of missing data, the more information is lost. Secondly, we can find that the designed model can achieve the highest accuracy when the spectrum datasets are incomplete with missing data. The proposed model, CNN-BiLSTM-attention, and seq2seq-LSTM-attention can achieve MAE values of 0.7864, 0.8543, and 0.8876, respectively, when the missing ratio is 5% in dataset 1. Similarly, the proposed model also performs best on dataset 2 with different missing rates. In particular, the decreasing performance slope of the proposed model is the lowest when the missing rate is increasing from 5% to 10% and 15%. This also shows the strong stability of the designed model.

5. Conclusions

In this paper, we have investigated the problem of spectrum occupancy prediction and proposed a deep learning prediction model achieving both long-term and time–frequency spectrum prediction. To be more specific, we take advantage of advanced Bi-ConvLSTM to extract temporal–spectral relationships in spectrum data and the seq2seq framework for long-term spectrum prediction. Then, the attention mechanism is used to solve the seq2seq limitation. Subsequently, we evaluate the prediction performance of the conventional LSTM, GRU, BiLSTM, CNN, CNN-LSTM, CNN-BiLSTM, CNN-BiLSTM-attention, seq2seq-LSTM, seq2seq-LSTM-attention, and seq2seq-Bi-ConvLSTM for comparison purposes. Moreover, the experimental results demonstrate that the proposed model performs better and also is robust to missing data in spectrum training data. Finally, the proposed model has been applied to our program and will be adopted in real-world scenarios.

Author Contributions

Conceptualization, methodology, and formal analysis, L.W.; writing—original draft preparation, L.W.; formal analysis, R.J.; writing—review and editing, J.H.; supervision, J.H. and Z.C.; project administration, J.H. and Z.C.; funding acquisition, J.H. and Z.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by the National Key R&D Program of China under Grant 2021YFA0716600, the Shenzhen Fundamental Research Program under Grant No. JCYJ20180307151430655, and the Shenzhen Science and Technology Program under Grant No. KQTD20190929172704911.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are openly available at https://github.com/chengrunmeng/Aachen-spectrum-data-part.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ding, G.; Wang, J.; Wu, Q.; Yao, Y.D.; Li, R.; Zhang, H.; Zou, Y. On the limits of predictability in real-world radio spectrum state dynamics: From entropy theory to 5G spectrum sharing. IEEE Commun. Mag. 2015, 53, 178–183. [Google Scholar] [CrossRef]
  2. Akyildiz, I.F.; Lo, B.F.; Balakrishnan, R. Cooperative spectrum sensing in cognitive radio networks: A survey. Phys. Commun. 2011, 4, 40–62. [Google Scholar] [CrossRef]
  3. Eze, J.; Zhang, S.; Liu, E.; Eze, E. Cognitive radio-enabled Internet of Vehicles: A cooperative spectrum sensing and allocation for vehicular communication. IET Netw. 2018, 7, 190–199. [Google Scholar] [CrossRef]
  4. Hu, F.; Chen, B.; Zhu, K. Full spectrum sharing in cognitive radio networks toward 5G: A survey. IEEE Access 2018, 6, 15754–15776. [Google Scholar] [CrossRef]
  5. Tandra, R.; Sahai, A. Fundamental limits on detection in low SNR under noise uncertainty. In Proceedings of the 2005 International Conference on Wireless Networks, Communications and Mobile Computing, Maui, HI, USA, 13–16 June 2005; Volume 1, pp. 464–469. [Google Scholar]
  6. Jia, M.; Zhang, X.; Sun, J.; Gu, X.; Guo, Q. Intelligent resource management for satellite and terrestrial spectrum shared networking toward B5G. IEEE Wirel. Commun. 2020, 27, 54–61. [Google Scholar] [CrossRef]
  7. Ghasemi, A.; Sousa, E.S. Collaborative spectrum sensing for opportunistic access in fading environments. In Proceedings of the First IEEE International Symposium on New Frontiers in Dynamic Spectrum Access Networks, 2005. DySPAN 2005, Baltimore, MD, USA, 8–11 November 2005; pp. 131–136. [Google Scholar]
  8. Lan, K.; Zhao, H.; Zhang, J.; Long, C.; Luo, M. A spectrum prediction approach based on neural networks optimized by genetic algorithm in cognitive radio networks. In Proceedings of the 10th International Conference on Wireless Communications, Networking and Mobile Computing (WiCOM 2014), Beijing, China, 26–28 September 2014; pp. 131–136. [Google Scholar]
  9. Ding, G.; Jiao, Y.; Wang, J.; Zou, Y.; Wu, Q.; Yao, Y.D.; Hanzo, L. Spectrum Inference in Cognitive Radio Networks: Algorithms and Applications. IEEE Commun. Surv. Tutor. 2018, 20, 150–182. [Google Scholar] [CrossRef]
  10. Akbar, I.A.; Tranter, W.H. Dynamic spectrum allocation in cognitive radio using hidden Markov models: Poisson distributed case. In Proceedings of the 2007 IEEE SoutheastCon, Richmond, VA, USA, 22–25 March 2007; pp. 196–201. [Google Scholar]
  11. Zhao, Y.; Luo, S.; Yuan, Z.; Lin, R. A New Spectrum Prediction Method for UAV Communications. In Proceedings of the 2019 IEEE 5th International Conference on Computer and Communications (ICCC), Chengdu, China, 6–9 December 2019; pp. 826–830. [Google Scholar]
  12. Chen, X.; Yang, J.; Ding, G. Minimum Bayesian risk based robust spectrum prediction in the presence of sensing errors. IEEE Access 2018, 6, 29611–29625. [Google Scholar] [CrossRef]
  13. Pandian, P.; Selvaraj, C.; Bhalaji, N.; Arun Depak, K.G.; Saikrishnan, S. Machine Learning based Spectrum Prediction in Cognitive Radio Networks. In Proceedings of the 2023 International Conference on Networking and Communications (ICNWC), Chennai, India, 5–6 April 2023; pp. 1–6. [Google Scholar]
  14. Naikwadi, M.H.; Patil, K.P. A Survey of Artificial Neural Network based Spectrum Inference for Occupancy Prediction in Cognitive Radio Networks. In Proceedings of the 2020 4th International Conference on Trends in Electronics and Informatics (ICOEI)(48184), Tirunelveli, India, 15–17 June 2020; pp. 903–908. [Google Scholar]
  15. Ge, C.; Wang, Z.; Zhang, X. Robust long-term spectrum prediction with missing values and sparse anomalies. IEEE Access 2019, 7, 16655–16664. [Google Scholar] [CrossRef]
  16. Sun, J.; Wang, J.; Ding, G.; Shen, L.; Yang, J.; Wu, Q.; Yu, L. Long-term spectrum state prediction: An image inference perspective. IEEE Access 2018, 6, 43489–43498. [Google Scholar] [CrossRef]
  17. M, S.; G, S.M. Study of Spectrum Prediction Techniques in Cognitive Radio Networks. In Proceedings of the 2022 8th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 25–26 March 2022; Volume 1, pp. 467–471. [Google Scholar]
  18. Nandakumar, R.; Ponnusamy, V.; Mishra, A.K. LSTM Based Spectrum Prediction for Real-Time Spectrum Access for IoT Applications. Intell. Autom. Soft Comput. 2023, 35, 2805–2819. [Google Scholar] [CrossRef]
  19. Yu, L.; Chen, J.; Ding, G.; Tu, Y.; Yang, J.; Sun, J. Spectrum prediction based on Taguchi method in deep learning with long short-term memory. IEEE Access 2018, 6, 45923–45933. [Google Scholar] [CrossRef]
  20. Wang, X.; Chen, Q.; Yu, X. Research on Spectrum Prediction Technology Based on B-LTF. Electronics 2023, 12, 247. [Google Scholar] [CrossRef]
  21. Aygül, M.A.; Nazzal, M.; Sağlam, M.İ.; da Costa, D.B.; Ateş, H.F.; Arslan, H. Efficient spectrum occupancy prediction exploiting multidimensional correlations through composite 2D-LSTM models. Sensors 2020, 21, 135. [Google Scholar] [CrossRef] [PubMed]
  22. Aygül, M.A.; Nazzal, M.; Ekti, A.R.; Görçin, A.; da Costa, D.B.; Ateş, H.F.; Arslan, H. Spectrum occupancy prediction exploiting time and frequency correlations through 2D-LSTM. In Proceedings of the 2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Antwerp, Belgium, 25–28 May 2020; pp. 1–5. [Google Scholar]
  23. Yu, L.; Guo, Y.; Wang, Q.; Luo, C.; Li, M.; Liao, W.; Li, P. Spectrum availability prediction for cognitive radio communications: A DCG approach. IEEE Trans. Cogn. Commun. Netw. 2020, 6, 476–485. [Google Scholar] [CrossRef]
  24. Shawel, B.S.; Woldegebreal, D.H.; Pollin, S. Convolutional LSTM-based long-term spectrum prediction for dynamic spectrum access. In Proceedings of the 2019 27th European Signal Processing Conference (EUSIPCO), A Coruna, Spain, 2–6 September 2019; pp. 1–5. [Google Scholar]
  25. Ding, X.; Feng, L.; Zou, Y.; Zhang, G. Deep learning aided spectrum prediction for satellite communication systems. IEEE Trans. Veh. Technol. 2020, 69, 16314–16319. [Google Scholar] [CrossRef]
  26. Zhang, H.; Tian, Q.; Han, Y. Multi channel spectrum prediction algorithm based on GCN and LSTM. In Proceedings of the 2022 IEEE 96th Vehicular Technology Conference (VTC2022-Fall), London, UK, 26–29 September 2022; pp. 1–5. [Google Scholar]
  27. Zhang, X.; Guo, L.; Ben, C.; Peng, Y.; Wang, Y.; Shi, S.; Lin, Y.; Gui, G. A-GCRNN: Attention Graph Convolution Recurrent Neural Network for Multi-Band Spectrum Prediction. IEEE Trans. Veh. Technol. 2024, 73, 2978–2982. [Google Scholar] [CrossRef]
  28. Yu, L.; Chen, J.; Zhang, Y.; Zhou, H.; Sun, J. Deep spectrum prediction in high frequency communication based on temporal-spectral residual network. China Commun. 2018, 15, 25–34. [Google Scholar] [CrossRef]
  29. Gao, Y.; Yang, L.T.; Yang, J.; Wang, H.; Zhao, Y. Attention U-Net Based on Bi-ConvLSTM and Its Optimization for Smart Healthcare. IEEE Trans. Comput. Soc. Syst. 2023, 10, 1966–1974. [Google Scholar] [CrossRef]
  30. Zuo, P.; Peng, T.; Wang, X.; You, K.; Jing, H.; Guo, W.; Wang, W. Spectrum prediction for frequency bands with high burstiness: Analysis and method. In Proceedings of the 2020 IEEE 91st Vehicular Technology Conference (VTC2020-Spring), Antwerp, Belgium, 25–28 May 2020; pp. 1–7. [Google Scholar]
  31. Wellens, M. Empirical Modelling of Spectrum Use and Evaluation of Adaptive Spectrum Sensing in Dynamic Spectrum Access Networks. Ph.D. Thesis, RWTH Aachen University, Aachen, Germany, 2010. [Google Scholar]
  32. Tusha, A.; Kaplan, B.; Cırpan, H.A.; Qaraqe, K.; Arslan, H. Intelligent Spectrum Occupancy Prediction for Realistic Measurements: GRU based Approach. In Proceedings of the 2022 IEEE International Black Sea Conference on Communications and Networking (BlackSeaCom), Sofia, Bulgaria, 6–9 June 2022; pp. 179–184. [Google Scholar]
  33. Ben, C.; Peng, Y.; Zhang, X.; Wang, Y.; Guo, L.; Lin, Y.; Gui, G. Multi-Band Spectrum Prediction Method via Singular Spectrum Analysis and A-BiLSTM. In Proceedings of the 2023 IEEE 23rd International Conference on Communication Technology (ICCT), Wuxi, China, 20–22 October 2023; pp. 645–650. [Google Scholar]
  34. Ren, X.; Mosavat-Jahromi, H.; Cai, L.; Kidston, D. Spatio-Temporal Spectrum Load Prediction Using Convolutional Neural Network and ResNet. IEEE Trans. Cogn. Commun. Netw. 2022, 8, 502–513. [Google Scholar] [CrossRef]
  35. Li, L.; Xie, W.; Zhou, X. Cooperative Spectrum Sensing Based on LSTM-CNN Combination Network in Cognitive Radio System. IEEE Access 2023, 11, 87615–87625. [Google Scholar] [CrossRef]
  36. Yilihamu, D.; Ablimit, M.; Hamdulla, A. Speech Language Identification Using CNN-BiLSTM with Attention Mechanism. In Proceedings of the 2022 3rd International Conference on Pattern Recognition and Machine Learning (PRML), Chengdu, China, 22–24 July 2022; pp. 308–314. [Google Scholar]
  37. Anneswa, G.; Sneha, K.; Jacobus, V.D.M. Spectrum Usage Analysis And Prediction using Long Short-Term Memory Networks. In Proceedings of the 24th International Conference on Distributed Computing and Networking, Kharagpur, India, 4–7 January 2023; pp. 270–279. [Google Scholar]
  38. Cui, Y.; Yin, B.; Li, R.; Du, Z.; Ding, M. Short-time Series Load Forecasting By Seq2seq-LSTM Model. In Proceedings of the 2020 IEEE 9th Joint International Information Technology and Artificial Intelligence Conference (ITAIC), Chongqing, China, 11–13 December 2020; Volume 9, pp. 517–521. [Google Scholar]
  39. Ke, X.; Guoqiang, Z.; Zhaoyang, D.; Kang, Z.; Kaizhu, H. Self-supervised Generative Learning for Sequential Data Prediction. Appl. Intell. 2023, 53, 20675–20689. [Google Scholar]
Figure 1. The working process of CRNs.
Figure 1. The working process of CRNs.
Sensors 24 01498 g001
Figure 2. The principle of traditional spectrum prediction models.
Figure 2. The principle of traditional spectrum prediction models.
Sensors 24 01498 g002
Figure 3. The principle of proposed spectrum prediction model.
Figure 3. The principle of proposed spectrum prediction model.
Sensors 24 01498 g003
Figure 4. The construction of spectrum data for the proposed prediction model.
Figure 4. The construction of spectrum data for the proposed prediction model.
Sensors 24 01498 g004
Figure 5. The construction of long-term joint temporal–spectral spectrum prediction models.
Figure 5. The construction of long-term joint temporal–spectral spectrum prediction models.
Sensors 24 01498 g005
Figure 6. The structure of Bi-ConvLSTM.
Figure 6. The structure of Bi-ConvLSTM.
Sensors 24 01498 g006
Figure 7. Internal structure of the ConvLSTM.
Figure 7. Internal structure of the ConvLSTM.
Sensors 24 01498 g007
Figure 8. The internal structure of the attention layer.
Figure 8. The internal structure of the attention layer.
Sensors 24 01498 g008
Figure 9. The effect of the time length of the designed model.
Figure 9. The effect of the time length of the designed model.
Sensors 24 01498 g009
Figure 10. Performance comparison with the existing schemes.
Figure 10. Performance comparison with the existing schemes.
Sensors 24 01498 g010
Figure 11. The stability against missing data for spectrum prediction.
Figure 11. The stability against missing data for spectrum prediction.
Sensors 24 01498 g011
Table 1. The comparison of performance on test set.
Table 1. The comparison of performance on test set.
MethodsMAEMAPERMSE R 2
LSTM0.89828.23%1.65120.8223
BiLSTM0.87968.07%1.57280.8473
GRU0.88478.17%1.58510.8398
CNN0.86797.93%1.56950.8491
CNN-LSTM0.85337.67%1.55880.8587
CNN-BiLSTM0.84987.45%1.55370.8619
CNN-BiLSTM-attention0.83457.25%1.42180.9065
Seq2seq-LSTM0.85567.47%1.56180.8578
Seq2seq-LSTM-attention0.84367.36%1.54650.8783
Seq2seq-Bi-ConvLSTM0.80646.64%1.23550.9362
Seq2seq-Bi-ConvLSTM-attention0.77496.15%1.09780.9628
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, L.; Hu, J.; Jiang, R.; Chen, Z. A Deep Long-Term Joint Temporal–Spectral Network for Spectrum Prediction. Sensors 2024, 24, 1498. https://doi.org/10.3390/s24051498

AMA Style

Wang L, Hu J, Jiang R, Chen Z. A Deep Long-Term Joint Temporal–Spectral Network for Spectrum Prediction. Sensors. 2024; 24(5):1498. https://doi.org/10.3390/s24051498

Chicago/Turabian Style

Wang, Lei, Jun Hu, Rundong Jiang, and Zengping Chen. 2024. "A Deep Long-Term Joint Temporal–Spectral Network for Spectrum Prediction" Sensors 24, no. 5: 1498. https://doi.org/10.3390/s24051498

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop