Next Article in Journal
Advanced Monocular Outdoor Pose Estimation in Autonomous Systems: Leveraging Optical Flow, Depth Estimation, and Semantic Segmentation with Dynamic Object Removal
Next Article in Special Issue
Solving Integer Ambiguity Based on an Improved Ant Lion Algorithm
Previous Article in Journal
MR Elastography Using the Gravitational Transducer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

XAI GNSS—A Comprehensive Study on Signal Quality Assessment of GNSS Disruptions Using Explainable AI Technique

by
Arul Elango
1,* and
Rene Jr. Landry
2
1
Vignan’s Foundation for Science, Technology and Research, Guntur 522213, Andhra Pradesh, India
2
LASSENA—Laboratory of Space Technologies, Embedded Systems, Navigation and Avionics, École de Technologie Supérieure (ETS), Montreal, QC H3C-1K3, Canada
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(24), 8039; https://doi.org/10.3390/s24248039
Submission received: 9 October 2024 / Revised: 28 November 2024 / Accepted: 13 December 2024 / Published: 17 December 2024
(This article belongs to the Special Issue Signal Processing for Satellite Navigation and Wireless Localization)

Abstract

:
The hindering of Global Navigation Satellite Systems (GNSS) signal reception by jamming and spoofing attacks degrades the signal quality. Careful attention needs to be paid when post-processing the signal under these circumstances before feeding the signal into the GNSS receiver’s post-processing stage. The identification of the time domain statistical attributes and the spectral domain characteristics play a vital role in analyzing the behaviour of the signal characteristics under various kinds of jamming attacks, spoofing attacks, and multipath scenarios. In this paper, the signal records of five disruptions (pure, continuous wave interference (CWI), multi-tone continuous wave interference (MCWI), multipath (MP), spoofing, pulse, and chirp) are examined, and the most influential features in both the time and frequency domains are identified with the help of explainable AI (XAI) models. Different Machine learning (ML) techniques were employed to assess the importance of the features to the model’s prediction. From the statistical analysis, it has been observed that the usage of the SHapley Additive exPlanations (SHAP) and local interpretable model-agnostic explanations (LIME) models in GNSS signals to test the types of disruption in unknown GNSS signals, using only the best-correlated and most important features in the training phase, provided a better classification accuracy in signal prediction compared to traditional feature selection methods. This XAI model reveals the black-box ML model’s output prediction and provides a clear explanation of the specific signal occurrences based on the individual feature contributions. By using this black-box revealer, we can easily analyze the behaviour of the GNSS ground-station signals and employ fault detection and resilience diagnosis in GNSS post-processing.

1. Introduction

The positioning-based applications available today that rely on satellite navigation systems, including the widely used Global Positioning System (GPS), Russian GLONASS, European GALILEO, Chinese BEIDOU, Japanese QZSS and Indian IRNSS, have become integral parts of various aspects of modern life, including transportation, agriculture, telecommunications, and emergency services. These systems depend on a collection of satellites orbiting the globe to provide accurate positioning, navigation, and timing information to users worldwide. However, they are susceptible to interference, which can disrupt the signals and compromise their reliability. Satellite navigation interference refers to deliberate or unintentional disruptions of signals between satellites and ground-based receivers. Interference can take various forms, including jamming, which is the deliberate transmission of strong radio signals on the same frequencies used by satellite navigation systems to overpower and disrupt genuine satellite signals. The jamming devices are sometimes used for illegal purposes, for example, the usage of cheap commercial hardware jammers, such as cigarette lighters, available on the market. Other illegitimate GNSS creations, like signals erupting from high-power transmitters, also degrade the quality of the signal by completely blocking the signal reception and not allowing the GNSS satellites to compute the user position. Another kind of attack is spoofing, which is a more sophisticated form of interference where attackers generate fake satellite signals that mimic authentic ones. Spoofing can deceive receivers into calculating inaccurate positions. Signal blockage may be natural or deliberate; for example, obstructions caused by human intervention and obstacles created by the construction of tall buildings or terrain can block or reflect satellite signals, causing multipath errors that lead to inaccuracies in positioning. Multipath interference signals bouncing off nearby surfaces before reaching the receiver can create timing errors and inaccuracies.
In the event of GNSS catastrophic failures, continuous monitoring and taking precautionary measures to detect unauthorized jamming and spoofing are required to prevent failures in GNSS data transmission at ground stations [1,2,3]. Early detection and rapid responses to counteract the impact of disruptions are required for seamless data reception and transmission [4,5]. Recently, some solutions have involved implementing security protocols and GNSS signal encryption techniques in navigation data to resist these kinds of interruptions, which could make it more difficult to disrupt or manipulate the signals. In the GNSS-denied situation, alternative navigation techniques, like the use of inertial measurement units (IMUs), a mixed-mode operation that involves using accelerometers and gyroscope fusion, may help to measure the acceleration and angular velocity, which enhances the resilience and redundancy. Subsequently, computed features from the data serve as input to the ML and deep learning (DL) techniques for classifying and detecting the types of interruption. However, there is no proper explanation in the model output that describes how the model is responding to the changes in the signal and what factors influence the identification of key discriminative factors in distinguishing different classes of disruptions.
Taking into consideration all the above shortcomings in signal detection using traditional ML techniques, this study proposes a novel framework for analyzing the GNSS signal characteristics. The proposed approach not only enables the accurate classification of GNSS signal disruptions but also provides valuable insights into the underlying factors driving these disruptions. The main contributions of the work are summarized as follows:
  • In this paper, we recorded a synthetic GNSS dataset under various disruptions at different Jamming-to-Signal Ratio (JSR) levels using the Skydel simulator. The dataset was used to evaluate the performances of different ML models on signal prediction/classification tasks.
  • We extracted and compared a set of attributes (time domain and frequency domain) using the statistical parameters to identify the strongly and weakly correlated features, as visualized in the correlation plot.
  • We utilized the most influential features from the global (SHAP) and local explanation (LIME) XAI models in the form of feature rankings, a summary plot, and a forced plot with detailed explanations, using only the key features for predicting and classifying the signal disruption resulted in improved metrics.

2. Literature Review

There are some existing approaches based on the artificial intelligence applied in GNSSs that are discussed in articles that describe the utilization of different ML/DL models at different stages, such as the RF front-end stage, pre-correlation stage, and post-correlation stage [6,7,8,9,10,11,12,13,14]. The different ML algorithms used in the literature, such as the distance-based representational model, autoregressive integrated moving average (ARIMA), and statistical measures (mean, standard deviation, variance, median, quantile, kurtosis, skewness, etc.), have been shown that they have several shortcomings and that they cannot be used to model multivariate time series data. Some of the traditional statistical approaches used in [15] provided information about the trends and outliers in the signal, with the help of the correlation in the data on 30% of the quartile error values. Many categories of anomalies and several ML algorithms listed in the literature (supervised, semi-supervised, and unsupervised) were used to detect the GNSS abnormalities in the signal. A very detailed description of the type of data, the categories of anomalies, and the evaluation criteria are provided in [16] in the context of Internet of Things (IOT) data, which paved the way for analyzing the suitability of the selected deep learning model. Similarly, a survey of ML models and a variety of ML-based abnormality detection schemes illustrated in [17] created an interest among readers in choosing the best model for anomaly detection.
Some of the studies listed here used real- and complex-valued long short-term memory (LSTM) algorithms that predicted the next time steps based on loss values [18,19]. A hyper approach developed based on LSTM and a one-class support vector machine-based anomaly detection method has been implemented by Elsayed et al. [20] for detecting attacks in the network of an unbalanced dataset. The method employed has the advantage of being able to operate on high-dimensional datasets by reducing the processing time. Using this hybrid model, a high detection rate was obtained for securing the Software Defined Network (SDN) from malicious traffic. An LSTM network layer has been used to denoise and forecast the interruption in the GNSS permanent station for early warning monitoring on different time series data [21,22]. However, the data used in this work rely on post-processing observations. A recurrent neural network (RNN)-based autoencoder network constructed by Wu et al. [23] is a good choice for learning the bonds among time-correlated data. This autoencoder network could detect interference with F1-scores of 0.93, 0.90, and 0.83, respectively, under three communication scenarios: a 16-quadrature amplitude modulated (QAM) signal with superimposed quadrature phase shift keying (QPSK), a frequency modulated (FM) signal, and a clean FM signal contaminated by a direct sequence spread spectrum (DSSS). This model only considers anomaly detection in generic modulation schemes.
An overview of the use of LSTM models in GNSS jamming and spoofing and their limitations is presented next. Spoofing detection is carried out with the help of a general adversarial network (GAN) by taking the cross-ambiguity function (CAF) as an acquisition output for both authentic and spoofed signals [24]. The acquisition of satellite signals needs to be performed as a function of receiver signal processing. A jamming detection method using an LSTM-based model was introduced by [25] for satellite signal reception in the presence of alpha-state noise. The pure satellite signal of interest that is not a jamming instance is used for LSTM training. Here, the myriad filter was employed to suppress the alpha-state noise. A detection rate of 98% was achieved for a JSR of 7 dB. The random motion between the transmitted and received GPS signals differs from the live sky-generated Doppler patterns that are used to find the presence of abnormal behaviour in the Doppler frequency or malicious attackers to analyze the spoofing threats in GNSSs. In another study [26], a training dataset obtained from the daily live record data uploaded by NASA was used to train the LSTM models for spoofing detection. The architecture implemented consisted of two LSTM layers with 32 neurons, with one dropout layer and one dense layer, for predicting the spoofing attacks. The Doppler shift detected from the post-processing of the GPS signal is used for training. The hybrid learning framework of GNSS observations using anomaly detection via the hierarchical density-based spatial clustering of applications with noise (HDBSCAN) technique has been used by Xia et al. [27]. The ComNav K508 GNSS OEM receiver was used to log the Receiver INdependent EXchange (RINEX) data that are used for training purposes, and the NovAtel ProPak6 receiver was used to collect the test set for real-time anomaly detection. Anomaly detection in the transponder frequency spectra of satellite communications systems with high-dimensional time series data using LSTM networks is implemented by Gunn et al. [28]; this provides a lower reconstruction error, but the time-frequency approach for collecting the images for each snapshot increases the data handling and the data storage complexity. A machine learning-based identification of the clean, non-line of sight (NLOS), and multipath signals in a GNSS is presented in [29]. The variations in the received power level, the pseudorange residue, and changes in the variance between the pseudorange and delta pseudorange obtained from the National Electrical Manufacturers Association (NEMA) and RINEX are used to train the machine learning algorithm.
To handle the time synchronization attacks, a three-layer multi-layer perceptron neural network was implemented in [30] to nullify jamming in stationary GPS receivers. The authors prepared the clock offset dataset of 200 samples used as an input to the model for training, with different threshold conditions that lead to spoofing detection and positioning error reduction. Spoofing detection in GNSS signals with the aid of the LSTM prediction model has been implemented by Dasgupta et al. [31] using a publicly available real-world driving dataset. The data captured from the controlled area network (CAN), GNSS, and IMU sensors with different parameters such as acceleration, steering wheel angle, speed, and distance are used to detect spoofing attacks in autonomous vehicles.
To the best of our knowledge, most of the methods listed above used either traditional statistical methods [32,33] or spectrogram images, post-processed data, correlation outputs, and data collected at the navigation level to train an ML/DL model to identify possible trends in the signal quality and abnormalities in GNSS data. However, there is still a need to determine the reason behind these changes in the signal behaviour, and the cause of abnormalities detected at the RF signal level has to be identified before the signal is fed into the receiver signal processing chain. Nevertheless, the aforementioned literature studies used several ML/DL models and collected the most influential features by suitably classifying or predicting the anomalous GNSS signals with proper justification. However, their AI models failed to provide the reasons behind the classification/prediction and how the classification decisions were made. Also, it is hard to know on what basis the AI model is functioning and what aspects the model uses to determine the result. Because of the lack of transparency and trustworthiness of these AI models, they are called black-box models, and therefore, it is necessary to know the causes and effects of signal disruption or what would happen if the signal attributes changed.
To solve this problem, XAI methods are used. These methods are very popular in the AI community and very well-known for making decisions and understanding or detecting the root cause of a correct/wrong prediction or classification in terms of all aspects [34,35,36,37,38]. XAI algorithms such as LIME and SHAP gained more attention because they can provide an explanation of the prediction results of a variety of applications [39,40,41,42,43]. The pre-ad hoc and post-ad hoc AI models mentioned in the literature, such as the saliency map [44], Grad-CAM [45], DeepLIFT [46], and LRP [47], are applied in ML algorithms to provide more informativeness and faithfulness in medical diagnostics and decision-making processes in safety-critical autonomous systems [48]. To avoid potential failures in the GNSS positioning and timing information, the XAI black-box revealer can be applied to retain the resilience of Position, Navigation, and Timing (PNT)-dependent seamless services and operations.

3. GNSS Signal Disruption Simulation

The GNSS signal provides the position, velocity, and time (PVT) of the end user anywhere on the globe using the satellite measurements of at least four satellites. For this purpose, right-circularly polarized (RHCP) waves are continuously transmitted over three carrier frequencies, L1, L2, and L5 (1575.42 MHz, 1227.6 MHz, and 1176.45 MHz), respectively. The L1 and L2 carrier signals are modulated with binary phase shift keying (BPSK) modulation and pseudorandom noise (PRN) codes known as the C/A code (coarse and acquisition code) and the encrypted P(Y) code (precision code) [49]. The C/A codes are applicable for dual-purpose use in terms of civilian and military applications. In contrast, the P(Y) code is exclusively used by the US military and under the United States Department of Defense’s authorization [49].

GNSS Signal Quality Monitoring Setup

In this paper, the GNSS signals are recorded using a Skydel simulated satellite signal, represented in the form of in-phase (I) and quadrature-phase (Q) components. These signals are combined using a 2:1 combiner with various types of signal disruption generated by the simulator’s advanced interference tool environment. The RTL-SDR RF front end is utilized to digitize the signals at a specified sampling rate. To record the files, we can specify the sampling frequency, centre frequency, duration of the recording, etc. The default settings are a centre frequency of 1575.42 MHz, a sampling rate of 2.9 MSPS, 1024 samples per frame, and a duration of 300 s. The file format is int16 in the baseband frequency, separated into two .dat files, where each one is int8. The order of the signal recording is as follows: the first file is the in-phase (I) samples, and the second one is the quadrature (Q) samples. The total number of samples is 32,676, acquired in one millisecond, with a bandwidth of 2 MHz [50].
The digitized signals are then processed through GNSS receiver modules, including acquisition, tracking, navigation solution, and position calculation. The experimental setup for the GNSS signal under different disruptions is depicted in Figure 1a,b.
The L1 GNSS signal is expressed by
S L 1 = A p · P t · D t · cos 2 π f 1 t + φ + A c · C t · D t · s i n ( 2 π f 1 t + φ ) ,
where S L 1 , A p , and P(t) denote the L1 signal frequency, the amplitude, and the phase of the P(Y) code, respectively. D(t) corresponds to the navigation message, f 1 is the frequency of the L1 carrier, φ is the initial phase, and A c   and C(t) are the amplitude and phase of the C/A code, respectively. The nominal GNSS signal in the time domain, its spectrum, and its spectrogram recorded through an RTL SDR are depicted in Figure 2a.
Continuous Wave Interference (CWI)
The CWI-affected GNSS signal is given by
C W I = exp j 2 π f c w t ,
where f c w and t represent the centre frequency and period of interference, respectively. The CWI power level is varied from 10 to 60 dB and the centre frequency is set as 1575.42 MHz. Plots of the signal simulation in the time and frequency domains, its spectrum, and the spectrogram recorded for 25 dB can be seen in Figure 2b.
Pulse Interference
The pulse type of interference signal is denoted by
P I = P j     p τ t k = 1 K δ t k f r k . exp j 2 π f j k t ,
where P j and p τ t are the signal power and the pulse signal of the duty cycle ( τ ) , respectively. f r j k is the pulse repetition frequency, δ is the Dirac pulse, and is the convolution operator. Here, the pulse repetition frequency is 10,000 Hz, the duty cycle is 20%, and the signal power is fixed at 25 dB to record the signals in both the time and frequency domains, as shown in Figure 2c.
Chirp Interference (CI)
The mathematical expression for CI is given as follows:
C I = exp ( 2 π k 2 t 2 + 2 π f 0 t ) .
A bandwidth of 10 MHz and a sweeping time of 100 µs are set to record all events under chirp interference, as can be seen from Figure 2d.

4. Statistical Quality Assessment of GNSS Signal

In this study, we conducted a comprehensive analysis of GNSS signals by computing both time and frequency domain attributes. The time domain features included the mean, standard deviation, median, mean absolute deviation, root mean square error, 25th and 75th percentiles, inter-percentile range, skewness, kurtosis, entropy, and maximum-to-mean ratio. These time domain features were computed to capture the statistical properties of the signal. These metrics offer clear information about the patterns, central tendency, variability, distribution of power and shape of the spectrum in a specific band of interest, and overall complexity of the signal.

4.1. Time Domain Features

The features in the time domain representation are identified based on the statistical parameters computed below:
  • Mean:
The mean value of the samples of the GNSS signal is calculated as
M e a n j = 1 n i = 1 n X i j
The mean is calculated as the mean value of the ith sample, where   X i j is the jth feature (or data point) of the ith sample, and n is the number of features (or data points) in each sample.
2.
Median:
Similarly, the median value is computed as
M e d i a n j = M e d i a n X i 1 , X i 2 , X i n .
3.
Standard Deviation:
S t d j = 1 n i = 1 n X i j μ j 2
4.
Mean Absolute Deviation:
M A D j = i = 1 n X i j μ j n
where n is the total number of data points in the dataset.
5.
Root Mean Square Error:
R M S E j = 1 n i = 1 n X i j 2
6.
25th Percentile:
P 25 = p e r c e n t i l e X , 25
7.
75th Percentile:
P 75 = p e r c e n t i l e X , 75
8.
Inter-percentile Range:
I P R = P 75 P 25 .
9.
Skewness:
S k e w n e s s j = i = 1 n ( X i j X ^ ) 3 n σ 3 .
10.
Kurtosis:
K u r t o s i s j = i = 1 n ( X i j X ^ ) 4 n σ 4 .
11.
Entropy:
H j = i P x i j l o g 2 P x i j .
12.
Maximum-to-Mean Ratio (MTMR):
M T M R j = max x i j μ .
Moreover, frequency domain features were calculated to examine spectral characteristics. These include the normalized spectrum bandwidth, normalized spectrum kurtosis (NSK), normalized spectrum flatness (NSF), the ratio of the variance to the squared mean of the normalized spectrum (RVSM), single-frequency energy aggregation, and the ratio of the maximum peak to the second maximum peak of the normalized spectrum (RMPS). These features provide valuable information regarding the distribution of the signal power across different frequency components, a measure of information, the shape of the spectrum concerning the centre frequency, and a measure of the spectral shape and concentration. The spectral domain features are listed in Section 4.2.

4.2. Frequency Domain Features

By finding the spectral content of the signal using the fast Fourier transform (FFT), the following attributes were calculated:
  • Normalized Spectrum Bandwidth (NSBW):
N S B W = f m a x f m i n f c ,
where fc is the centre frequency.
2.
Normalized Spectrum Kurtosis (NSK):
N S K = k = f 1 f 2 f k S c 4 P k k = f 1 f 2 P k s 4 ,
where S c is the spectral centroid. It is given by
S c = i = 1 m f k m i i = 1 m m i
where f 1   and f 2 are the bounds of the frequency band, fk is the frequency, and P k is the normalized power and m-magnitude of the bin number.
3.
Normalized Spectrum Flatness (NSF):
N S F = e x p ( 1 N k = 1 N l n X ( k ) 1 N k = 1 N X ( k ) ,
where X ( k ) is the magnitude of the k-th frequency bin in the spectrum. N is the total number of frequency bins.
4.
Ratio of the Variance to the Squared Mean of the Normalized Spectrum (RVSM):
R V S M = σ ( X ) k = 1 N X ( k ) 2 ,
where X is the normalized spectrum, σ ( X ) is the variance of the normalized spectrum, and N is the total number of elements in the normalized spectrum.
5.
Single-Frequency Energy Aggregation (SFEA):
The summation of the squares of the Fourier transform coefficients results in an energy aggregation given by
E t o t a l = i = 1 n X i 2
6.
Ratio of the Maximum Peak to the Second Maximum Peak of the Normalised Spectrum (RMPS):
First, find the index in the spectrum X m a x of the maximum peak in X:   i m a x = a r g m a x ( X ) and remove the maximum peak from X, denoted as X’:   X = X / X i m a x . Then, the second peak is calculated as m a x ( X ).
When computing the feature correlation using the Pearson correlation coefficient, the closer the value is to +1, the better we can deduce from the plot that the feature being considered is the feature most correlated with the target variable (response). The correlation coefficient is defined as
r a , b = i = 1 n ( a i a ~   ) ( b i b ~ ) i = 1 n ( a i a ~ ) 2 i = 1 n ( b i b ~ ) 2 ,
where the two attribute samples are denoted by a i and b i , and the mean values of the samples are given as a ~ and b ~ . The correlation coefficient represents the degree of positive or negative correlation between two variables, with its value ranging from −1.0 to +1.0. the darker colours signify weaker correlations, while the lighter shades denote stronger correlations. Figure 3a,b show the correlation maps of the weakly and strongly correlated features of the deteriorated GNSS signal in both domains.
Figure 4a illustrates the power levels of different types of interference versus six frequency domain features. Notably, as the power level increases from 40 to 60 dB, there is a sudden rise in the normalized spectral kurtosis value. For chirp-type interference, the normalized spectral flatness, RVSM, and RMPS values are slightly elevated at mid-range interference power levels.
In contrast, the frequency domain feature, depicted in Figure 4b, shows a spurious peak in the mean value for pulse-type interference around a 30 dB interference power level. Additionally, features such as the Root Mean Square (RMS), 75th percentile, inter-percentile range, entropy, and skewness exhibit a reduction in values compared to other types of interference. The kurtosis value for pulse-type interference varies significantly, whereas it remains constant for other types of interference.
The next step was to decide on the best choice among two signal domain attributes, primarily depending on how well each set of features could distinguish between the different signal types. The statistical analysis has been made for both time and frequency domain attributes of various GNSS signal categories. We conducted a p-test to evaluate the normality of the signals. The null hypothesis of the p-test of the residuals follows a normal distribution. Usually, the hypothesis is rejected if a p-value is less than 0.05. To visualize the results, a box-and-whisker plot of the p-values is plotted in Figure 5 for 25.6 s of data for each signal.
The black line of the box-and-whisker plot represents the median p-value that indicates the degree of normality. Based on the level of significance, the p-value is set to 0.05, denoted by a red dashed line. This confirms a high likelihood chance that the distribution is not normal below this value. The p-value is greater than 0.05, revealing that the residuals tend to follow a normal distribution. It is observed, based on the analysis, that the time domain attributes generally behave as normal distributions, and it is very difficult to find significant deviations. Therefore, the time domain attribute may not be an ideal choice for analyzing GNSS signal disruption using ML techniques. On the other hand, the p-values in the frequency domain for most GNSS signal categories were below 0.05 except for pulse, indicating deviations from normality and providing more promising information for analysis.

5. XAI Analysis for GNSS Signal Disruption Classification

5.1. Data Preparation

The dataset recorded from the Skydel simulator containing I and Q raw data was processed based on user-defined specifications from the RF front end. The samples were then prepared in both domains: the time domain and the frequency domain (by computing the FFT). As described in the previous section, the statistical features were generated and pre-processed (removing the missing values and duplicate records). The resulting dataset was subsequently partitioned into training and testing datasets.

5.2. ML Algorithm Training and Testing

The proposed model architecture for GNSS signal abnormality classification is depicted in Figure 6. In this work, five types of ML black-box models were employed to classify different types of signal disruptions in GNSS signals. Alongside these, a model explanation framework using LIME and SHAP was developed in parallel, with the sophisticated capabilities of this model explanation being feature contributions, an importance plot, a SHAP force plot, LIME model explanations, the best-suited features only being selected for training, and the redundant features and least important features being eliminated. The evaluation metrics are compared with traditional feature selection and elimination methods Principle Component Analysis (PCA, backward and forward selection methods). The redundant and unimportant features are neglected, and only the most important features are chosen for training. Then, based on the evaluation metrics (accuracy, precision, recall, F1-score, and Receiver Operating Characteristic ROC), the best-performing ML model was selected. Finally, knowing the cause of not predicting a particular class and the significance of the most trustworthy features for predicting the undesirable event would be helpful in analyzing the catastrophic failures in the GNSS ground-station data collection system so that the end user can take precautionary measures based on where exactly the signal quality is degraded and what factors contribute to its abnormal behaviour in the earlier stage before it is subjected to the post-processing stage. The ML models alone failed to capture the relationship between the features when it came to predicting abnormal behaviours in adverse conditions under jamming and spoofing signal attacks. However, this XAI framework helped to identify the irregularities, making it possible to investigate and respond quickly based on the explanations associated with the features in safety-critical operations.
The collected GNSS data have a duration of 40 s. A training-validation-testing split of 64-16-20% is chosen in this work. Each millisecond of data consists of 32,676 samples. We have used 25.6 s of data for training, which is 80% of 32 s of data, and 6.4 s of data are used for the validation phase. The remaining 20% of the total data, with a duration of 8 s, are used for the testing phase.

5.3. Local Explanations Results—LIME Technique

The prediction outputs of various ML algorithms are analyzed using the LIME model, which generates the local explanations by approximating the ML model’s output with an interpretable result based on the class prediction probabilities. The explanations for the predictions of the DT, KNN, AdaBoost, RF, and SVM models using the frequency domain features are shown in Figure 7a. For a specific sample in the test dataset, the model correctly predicted the class as spoofing, as the most important features for this sample are ‘normalized spectrum kurtosis’ and ‘single-frequency energy aggregation’ have more positive values that are the reason why the model correctly predicted the sample as spoofing. From the results, it is evident that the total value of features confirming that it is a spoofing category is greater than that of non-spoofing features. Therefore, the model correctly predicted this sample as spoofing. On the other hand, using the AdaBoost algorithm, the prediction probability of the testing instance appears to be 33% for MP and MCWI; this ambiguity led to inaccurate results despite the sample belonging to the spoofing category. The performance metrics of ML models using the LIME framework are given in Table 1a. Among all models, the SVM has the highest prediction results, and AdaBoost showed poor performance.
The prediction results of choosing all 12 features in the time domain using LIME are shown in Figure 7b. The RF model has the highest prediction accuracy of 71%, followed by DT and SVM. These models correctly predicted the class as MCWI, whereas other ML models could not be able to give the correct prediction. However, compared to frequency domain attributes, the performance of the ML model is significantly lower because most of the time, domain characteristics of the signal make it very difficult to distinguish the temporal relationships across the categories such as the multipath, spoofing, and clean signals, as these signal characteristics are highly similar in the time domain.
Table 2a shows the ML models’ performances based on the predicted label on the nth instance for the corresponding test sample using all six features in the dataset using the LIME algorithm. Here, there is difficulty in predicting the clean signal, as can be seen from the outputs of all the ML models. The AdaBoost model could not predict the MP, MCWI, clean, and spoofing signals correctly and failed in predicting all instances. RF has superior performance in predicting all kinds of interference, except chirp. Compared to all models, the SVM provided better metrics.
Table 2b gives a quantitative assessment of the ML model performance on LIME, which is essential for comparing the effectiveness across various scenarios. When it comes to the time domain features, the DT and RF models performed exceptionally well; in fact, the robustness and reliability of the model results can be seen in several cases with a perfect score of 100%. However, looking into the KNN and SVM models showed significant variability in some conditions; they also achieved very low accuracy, recall, and F1-score values. AdaBoost completely failed to provide the correct signal prediction for the MP and spoofing signals but performed well in other categories.

5.4. Local Explanations Results—SHAP Technique

SHAP is a black-box mathematical model revealer that can uncover the hidden operations in the outcome of any ML model. The explanation of the output of the ML model is governed by the game theory concept; it involves finding each feature’s contribution to the prediction. This method provides interpretability and explainability in both formats: global and local contexts. The Shapley value, which is fundamental to this approach, reflects the average expected marginal contribution of a feature across all possible feature combinations.
The impact of the most significant feature and its contribution to the prediction is governed by the following equation [34]:
φ j v a l = S { x 1 , . . x p } \ { x j } S ! p S 1 ! p !   v a l   S { x j } v a l ( S )
The term φ j   ∈ R represents the Shapley value for feature j. S ∈ {0, 1} denotes the subset of features included in the model, x is the vector of feature values for the instance being explained, p is the total number of features, val(S) is the prediction for the feature values in set S, marginalized over the features not included in S.
The XAI method is applied to all ML models, and based on the SHAP algorithm, the prediction of each class is computed; the influences of these individual features are plotted in ascending order in Figure 8a. For the frequency domain features, the normalized spectral kurtosis has provided the dominant value, followed by single-frequency energy aggregation, and the least important features are identified as the ratio of the minimum peak to the maximum normalized peak and the normalized spectral bandwidth. Similarly, for the time domain attributes, entropy and the root mean square error have played a significant role in predicting the output classes. The less important features are the skewness, median, maximum-to-mean ratio, and mean absolute deviation, which contributed very little to the prediction. A feature ranking based on the relative importance is given in Table 3a.
The SHAP value of the RF model’s output explains how the time and frequency domain features impact the output of the model. In Figure 8b, the impacts of the features on the GNSS signal disruption classes are stacked to create a feature importance plot. The summary plot for multiclass classification illustrates what the XAI model has learned from these features.
While analyzing the frequency domain features, except for KNN, RF, and other ML algorithms, the prediction of the model is determined by more than three features. For a clean signal, spectral flatness contributes the most. In a multipath signal, all features contribute equally except the normalized spectrum bandwidth and the ratio of the maximum peak to the second maximum peak of the normalized spectrum. For a pulse signal, spectral kurtosis is the most significant contributor. The least important feature is identified as the normalized spectrum bandwidth for the spoofing category. The plot shows that the normalized spectrum bandwidth has a greater contribution compared to other features. In both the CWI and MCWI cases, all features contribute equally, except the normalized spectrum bandwidth.
Table 3b shows the feature ranking based on the SHAP algorithm. The majority of features are ranked between fourth and tenth; therefore, there is no combined effect of picking the best-matched features after fine-tuning the ML models’ hyperparameters (DT: maximum depth from 4–50, minimum leaf samples from 5–20; RF: maximum depth from 5–50, number of estimators from 50–150; KNN: no. of neighbours from 4–10, leaf size from 5–50; SVM: regularization parameters 1–10, radial basis function, linear kernel; AdaBoost: depth from 5–50). Selecting the dominating attributes collectively is unreliable for the time domain features since every machine learning model generates a distinct rating.
Based on the order of attributes, the time domain feature importance plot is shown; we can see that the class CWI hardly uses the median, skewness, mean, max.-to-mean ratio, and mean absolute deviation. Additionally, the pulse and chirp classes use the same features. Kurtosis contributed significantly to the MCWI signal, which is why confusion arises between these classes relatively often. In order to separate the MP and clean signal types in a better way, new features need to be generated that are uniquely dedicated to distinguishing these classes. The normalized spectrum kurtosis has contributed significantly more, followed by single-frequency energy aggregation. For testing on all ML models, the ratio of the maximum peak to the second maximum peak of the normalized spectrum did not show much importance at all. For the time domain features, the mean, median, and maximum-to-mean ratio also do not contribute to the model’s prediction. A comparison of all models in both the time and frequency domains is depicted in Figure 9a,b.
The mean, median, 75th percentile and maximum-to-mean ratio are the least important features, and most of the attributes for ML models in the time domain are not utilized to predict the output. The mean absolute deviation is the only feature used in KNN. In the frequency domain, the normalized spectrum kurtosis is used by most of the ML models, as shown earlier in the feature importance plot; also, the ratio of the maximum peak to the second maximum peak of the normalized spectrum has not been considered much in forecasting the results of the model.
In this work, initially, the global SHAP values for all ML models are calculated; then, the feature importance is computed by multiplying the SHAP value with the model accuracy for each feature. Finally, we determine the overall average ranking across models. The frequency domain feature values are chosen from the initial value for the top k features. The metrics of all ML models show that a suitable performance can be obtained by selecting only the top four features and removing two unimportant features, resulting in the same performance, as can be seen in Table 4a. There is not much improvement in choosing more than four attributes; therefore, the last two features did not considerably improve the model performance.
Similarly, the time domain features are also used to evaluate the model’s performance, and there is no change in the accuracy value after increasing the feature values beyond K = 6; the RF model gives a good performance compared to the other models. Table 4b shows the effectiveness of top features in classifying the types of disruption. For the most important features, using traditional feature engineering, the model behaviour is not interpreted with respect to the types of disruption.
The features are ordered by their importance for each category of jamming instances using the summary plot in order to jointly analyze the feature importance and feature effects. Each point represents a Shapley value for a feature and an instance. The position on the y-axis corresponds to the attributes, and the x-axis specifies the Shapley value, with the colours indicating the feature values from low (blue) to high (red). A positive value pushes the prediction value higher, and negative values lower the prediction. The summary plots in Figure 10a show that the mean absolute deviation, being the least important feature, has low Shapley values for CWI and chirp interference. For MP and spoofing, the following features have no impact: the mean, mean absolute deviation, maximum-to-mean ratio, median, and skewness.
In Figure 10b, the summary plot illustrates that the least significant feature is the RMPS, with low Shapley values, for the CWI and spoofing categories. The chirp interference summary plot indicates that single-frequency energy aggregation and normalized spectrum kurtosis have a more neutral effect on the model’s predictions, and features like the normalized spectrum bandwidth and the ratio of the variance to the squared mean of the normalized spectrum have a significant impact. The dispersion of points also shows how different features have an impact on various cases.
The outcome of a SHAP model with a clear explanation for a specific GNSS signal type is visualized in Figure 11a,b in the form of a force plot. An illustration of a signal model prediction is displayed here, identifying the individual features’ contributions and providing an error analysis of signal disruption. The base value is fixed as 0.142, and the predicted value comes closer by 0.12 for a pulse signal, where the standard deviation had higher values, and the rest of the features had much lower values. For chirp and MCWI, kurtosis is a common feature that pushed the model to a lower score. In the case of MP, all features pushed in the positive direction, but kurtosis played a dominant role in correct prediction. Conversely, for the spoofing signals, most of the features are located below the base value of 0.148, pushing the model toward a lower prediction value of 0.05. The force plot for chirp interference reveals that it significantly pushes the prediction higher with a contribution of +0.15, while kurtosis reduces it by −0.0235, resulting in a final prediction of 0.1735. Similarly, in the MCWI case, kurtosis also contributed very little. In contrast, most of the features help to predict the multipath instance, and none of the features are important for predicting the spoofing instance.
The pulse signal prediction is based on all strongly correlated features that are above the base value of 0.142, whereas for the CWI signal, only two features (single-frequency energy aggregation and normalized spectral kurtosis) lie above the base value of 0.1425. The clean and chirp signals are similar to each other, with the normalized spectral kurtosis pushing the model to a higher score. Most of the features contributed to MP prediction since all predicted values are above the base value to yield the correct occurrence.
The One-vs-the-Rest (OvR) multiclass technique, commonly referred to as one-vs-all, is used to compute the area under the curve (AUC) for all machine learning models. To do this, an ROC curve must be computed for every GNSS signal class. In each iteration, one class is considered the positive class, and the other is regarded as the negative class. In the combined way, the classes are framed, and to consider every class equally, the metrics for each class are determined separately, and finally, we average them. By using these steps, the false positive and true positive rates of each class are aggregated based on the following equations:
T P R = i T P i T P i + F N i ,
F P R = i F P i F P i + T N i .
The KNN performs best in detecting clean signals by plotting the ROC curve with a score of 96, whereas other ML models such as RF, SVM, and AdaBoost performs closely around 90 while the DT lags with 42. For CWI detection, RF shows superior performance with 83, the SVM provides marginal performance around 55, and AdaBoost is the lowest at 33. The model’s ability is determined based on the area under the curve and how the model distinguishes between different classification categories, as shown by the ROC curve in Figure 12. The clean signal has occupied the highest area for most of the ML algorithms. For the AdaBoost, RF, and SVM algorithms, the model assigned a lower probability to the positive values, flipping the labels of the negative class and resulting in a lower area for the categories of MP and spoofing. The ROC plots for all categories are shown in Figure 12.
As per the frequency domain features, the classification of signal disruptions is evaluated in the form of a confusion matrix, as shown in Figure 13. The AdaBoost algorithm completely misclassifies the clean and multipath signals. The occurrence of this false classification between the MP and clean signals is due to similar kinds of signal characteristics in the frequency domain features, as the model should be able to capture any dissimilarities in the spectra of both disruptions. This instance is handled using the DT, Random Forest, KNN and SVM algorithms with more than 40% error.
Table 5 shows the results of feature selection using SHAP compared with other traditional feature selection methods; here, the optimal ‘k’ feature value is chosen as 4, based on the top features that are more trustworthy in classifying the instances correctly. Using the SVM method, SHAP has shown the leading performance when classifying the different GNSS signal disruptions. Other than the chirp signal, most of the signals are correctly classified, but when other feature selection/elimination methods are applied, the clean and spoofing signals are not properly classified. The superiority of the SHAP-based novel feature selection method can be seen in Table 5. The DT and KNN ML models, for most of the labels, classify the signal correctly, but when traditional methods are applied, the performance metrics are lower. For RF and KNN, our feature selection methods show the best performance, along with other feature reduction methods. Finally, for AdaBoost, the overall accuracy decreases under our methods; however, other performance metrics (precision, recall, and F1-score) increase for detecting the abnormalities in the signal classification.

6. Discussion

Major Findings and Future Scope

The study discussed in this paper primarily focused on the effectiveness and interpretability of employing ML algorithms and using XAI techniques for GNSS signal classification and detection tasks. The results demonstrated that ML models performed at the average level for signal classification; however, the decision-making process of the model was entirely dependent on interpretability and finding the inner relationship between the features. By incorporating the LIME and SHAP algorithms into feature importance and model predictions, improved transparency and easier model understanding can be achieved. Furthermore, the research findings demonstrated that the model’s prediction effectively leveraged domain knowledge and created causal relationships, leading to an enhanced comprehension of the attributes influencing abnormal behaviour in the data. Overall, the results showed that anomaly detection can be performed in a better way, and the system has the ability to grasp different behaviours in the data when the ML algorithms are combined with explainability techniques.
The major challenges in classifying the GNSS signals captured in a real-time scenario were that it was completely different from the simulation environment, and the signal reception could be completely deteriorated by narrow-band and wide-band interference [5]. Furthermore, in an urban short-delay MP environment, from analyzing the time domain characteristics, it is very difficult to understand the inner relationship between the signal attributes in the MP scenario. Similarly, the frequency domain characteristics also failed to capture the variations in the spectrum of the pure GNSS and MP categories. Analyzing characteristics from both domains may not be sufficient to identify the abnormalities in the signal. Based on the spectrogram/scalogram image features, it may be possible to find the differences concerning time and frequency observations by using advanced XAI techniques such as saliency maps or the GradCAM and EigenCAM models, which would be helpful in analyzing the variations in the signal behaviour effectively.

7. Conclusions

Our research enabled an accurate classification of GNSS signal disruptions using XAI techniques on top of ML models. The significance of time and frequency domain features was considered to enhance the understanding of GNSS signal behaviour under different conditions. Moreover, accurate positioning can be achieved effectively by utilizing XAI autonomous signal quality monitoring in GNSS data processing. At an earlier stage, the root causes of an abnormality can be easily found without post-processing the signal, that is, before feeding it to the receiver. The techniques used emphasized the most discriminating part of the disruption, and the application of the XAI technique ensured transparency and interpretability in our classification, thereby taking care of trustworthiness in model development and providing confidence in the results of signal quality analysis and prediction. Overall, this study provided a robust framework for interpreting and analyzing GNSS signals. The prediction results using the LIME model show that the SVM has performed well overall, but the prediction of clean signals remains a major concern, as most of the time, these signals are incorrectly classified. Similarly, the classification results using the SHAP algorithm indicate that chirp and clean signals have the problem of misclassification, particularly between these categories. To address these challenges, careful attention needs to be paid to designing a GNSS monitoring setup equipped with sophisticated ML models that can provide more robust signal prediction/classification capabilities.

Author Contributions

The two authors equally contributed to the XAI-based GNSS Signal Quality Assessment, Methodology—A.E.; Validation—R.J.L.; Investigation—A.E.; Writing, review and Editing—A.E., R.J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data that support the findings of this study are available upon reasonable request from the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sokolova, N.; Morrison, A.; Diez, A. Characterization of the GNSS RFI Threat to DFMC GBAS Signal Bands. Sensors 2022, 22, 8587. [Google Scholar] [CrossRef] [PubMed]
  2. Kaasalainen, S.; Mäkelä, M.; Ruotsalainen, L.; Malmivirta, T.; Fordell, T.; Hanhijärvi, K.; Wallin, A.E.; Lindvall, T.; Nikolskiy, S.; Olkkonen, M.K.; et al. Reason—Resilience and Security of Geospatial Data for Critical Infrastructures. In Proceedings of the CEUR Workshop Proceedings, Luxembourg, 3–4 December 2020; Volume 2880. [Google Scholar]
  3. Abraha, K.E.; Frisk, A.; Wiklund, P. GNSS Interference Monitoring and Detection Based on the Swedish CORS Network SWEPOS. J. Géod. Sci. 2024, 14, 20220157. [Google Scholar] [CrossRef]
  4. Lebrun, S.; Kaloustian, S.; Rollier, R.; Barschel, C. GNSS Positioning Security: Automatic Anomaly Detection on Reference Stations. In Critical Information Infrastructures Security; Percia, D., Mermoud, A., Maillart, T., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 60–76. [Google Scholar]
  5. Morrison, A.; Sokolova, N.; Gerrard, N.; Rødningsby, A.; Rost, C.; Ruotsalainen, L. Radio-Frequency Interference Considerations for Utility of the Galileo E6 Signal Based on Long-Term Monitoring by ARFIDAAS. J. Inst. Navig. 2023, 70, navi.560. [Google Scholar] [CrossRef]
  6. Jagiwala, D.; Shah, S.N. Possibilities of AI Algorithm Execution in GNSS. In Proceedings of the URSI Regional Conference on Radio Science (URSI-RCRS 2022), IIT Indore, India, 1–4 December 2022. [Google Scholar]
  7. Siemuri, A.; Kuusniemi, H.; Elmusrati, M.S.; Valisuo, P.; Shamsuzzoha, A. Machine Learning Utilization in GNSS—Use Cases, Challenges and Future Applications. In Proceedings of the 2021 International Conference on Localization and GNSS (ICL-GNSS), Tampere, Finland, 1–3 June 2021; pp. 1–6. [Google Scholar] [CrossRef]
  8. Ferre, R.M.; De La Fuente, A.; Lohan, E.S. Jammer Classification in GNSS Bands Via Machine Learning Algorithms. Sensors 2019, 19, 4841. [Google Scholar] [CrossRef] [PubMed]
  9. Elango, G.A.; Sudha, G.F.; Francis, B. Weak Signal Acquisition Enhancement in Software GPS Receivers—Pre-filtering Combined Post-correlation Detection Approach. Appl. Comput. Inform. 2017, 13, 66–78. [Google Scholar] [CrossRef]
  10. Jiang, Q.; Yan, S.; Yan, X.; Chen, S.; Sun, J. Data-Driven Individual–Joint Learning Framework for Nonlinear Process Monitoring. Control. Eng. Pract. 2020, 95, 104235. [Google Scholar] [CrossRef]
  11. Mohanty, A.; Gao, G. A Survey of Machine Learning Techniques for Improving Global Navigation Satellite Systems. EURASIP J. Adv. Signal Process. 2024, 2024, 73. [Google Scholar] [CrossRef]
  12. Reda, A.; Mekkawy, T. GNSS Jamming Detection Using Attention-Based Mutual Information Feature Selection. Discov. Appl. Sci. 2024, 6, 163. [Google Scholar] [CrossRef]
  13. van der Merwe, J.R.; Franco, D.C.; Feigl, T.; Rügamer, A. Optimal Machine Learning and Signal Processing Synergies for Low-Resource GNSS Interference Classification. IEEE Trans. Aerosp. Electron. Syst. 2024, 60, 2705–2721. [Google Scholar] [CrossRef]
  14. Fu, D.; Li, X.; Mou, W.; Ma, M.; Ou, G. Navigation jamming signal recognition based on long short-term memory neural networks. J. Syst. Eng. Electron. 2022, 33, 835–844. [Google Scholar] [CrossRef]
  15. Le Thi, N.; Männel, B.; Jarema, M.; Seemala, G.K.; Heki, K.; Schuh, H. Selection of an Optimal Algorithm for Outlier Detection in GNSS Time Series. In Proceedings of the EGU General Assembly 2021, Online, 19–30 April 2021. [Google Scholar]
  16. Al-Amri, R.; Murugesan, R.K.; Man, M.; Abdulateef, A.F.; Al-Sharafi, M.A.; Alkahtani, A.A. A Review of Machine Learning and Deep Learning Techniques for Anomaly Detection in IoT Data. Appl. Sci. 2021, 11, 5320. [Google Scholar] [CrossRef]
  17. Nassif, A.B.; Abu Talib, M.; Nasir, Q.; Dakalbab, F.M. Machine Learning for Anomaly Detection: A Systematic Review. IEEE Access 2021, 9, 78658–78700. [Google Scholar] [CrossRef]
  18. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  19. Savolainen, O.; Elango, A.; Morrison, A.; Sokolova, N.; Ruotsalainen, L. GNSS Anomaly Detection with Complex-Valued LSTM Networks. In Proceedings of the 2024 International Conference on Localization and GNSS (ICL-GNSS), Antwerp, Belgium, 25–27 July 2024; pp. 1–7. [Google Scholar] [CrossRef]
  20. Elsayed, M.S.; Le-Khac, N.A.; Dev, S.; Jurcut, A.D. DdoSNet: A Deep-Learning Model for Detecting Network Attacks. In Proceedings of the 2020 IEEE 21st International Symposium on “A World of Wireless, Mobile and Multimedia Networks”, Cork, Ireland, 31 August–3 September 2020; pp. 391–396. [Google Scholar]
  21. Loli Piccolomini, E.; Gandolfi, S.; Poluzzi, L.; Tavasci, L.; Cascarano, P.; Pascucci, A. Recurrent Neural Networks Applied to GNSS Time Series for Denoising and Prediction. In Proceedings of the 26th International Symposium on Temporal Representation and Reasoning (TIME 2019), Málaga, Spain, 16–19 October 2019. [Google Scholar]
  22. Shahvandi, M.K.; Soja, B. Modified Deep Transformers for GNSS Time Series Prediction. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 8313–8316. [Google Scholar] [CrossRef]
  23. Wu, Q.; Sun, Z.; Zhou, X. Interference Detection and Recognition Based on Signal Reconstruction Using Recurrent Neural Network. In Proceedings of the IEEE International Conference, Shanghai, China, 20–24 May 2019. [Google Scholar]
  24. Li, J.; Zhu, X.; Ouyang, M.; Li, W.; Chen, Z.; Fu, Q. GNSS Spoofing Jamming Detection Based on Generative Adversarial Network. IEEE Sensors J. 2021, 21, 22823–22832. [Google Scholar] [CrossRef]
  25. Liu, M.; Jin, L.; Shang, B. LSTM-based Jamming Detection for Satellite Communication with Alpha-Stable Noise. In Proceedings of the IEEE Wireless Communications and Networking Conference Workshops (WCNCW), Nanjing, China, 29 March 2021. [Google Scholar]
  26. Calvo-Palomino, R.; Bhattacharya, A.; Bovet, G.; Giustiniano, D. LSTM-based GNSS Spoofing Detection Using Low-Cost Spectrum Sensors. In Proceedings of the IEEE 21st International Symposium on “A World of Wireless, Mobile and Multimedia Networks”, Cork, Ireland, 31 August–3 September 2020. [Google Scholar]
  27. Xia, Y.; Pan, S.; Meng, X.; Gao, W.; Ye, F.; Zhao, Q.; Zhao, X. Anomaly Detection for Urban Vehicle GNSS Observation with a Hybrid Machine Learning System. Remote. Sens. 2020, 12, 971. [Google Scholar] [CrossRef]
  28. Gunn, L.; Smet, P.; Arbon, E.; McDonnell, M.D. Anomaly Detection in Satellite Communications Systems Using LSTM Networks. In Proceedings of the 2018 Military Communications and Information Systems (ICMCIS 2018), Warsaw, Poland, 22–23 May 2018. [Google Scholar]
  29. Hsu, L.T. GNSS Multipath Detection Using a Machine Learning Approach. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017. [Google Scholar]
  30. Orouji, N.; Mosavi, M.R. A Multi-layer Perceptron Neural Network to Mitigate the Interference of Time Synchronization Attacks in Stationary GPS Receivers. GPS Solut. 2021, 25, 84. [Google Scholar] [CrossRef]
  31. Dasgupta, S.; Rahman, M.; Islam, M.; Chowdhury, M. Prediction-based GNSS Spoofing Attack Detection for Autonomous Vehicles. In Proceedings of the Transportation Research Board 100th Annual Meeting, Washington, DC, USA, 5–29 January 2021. [Google Scholar]
  32. Wang, P.; Cetin, E.; Dempster, A.G.; Wang, Y.; Wu, S. Time Frequency and Statistical Inference Based Interference Detection Technique for GNSS Receivers. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 2865–2876. [Google Scholar] [CrossRef]
  33. Qin, W.; Dovis, F. Situational Awareness of Chirp Jamming Threats to GNSS Based on Supervised Machine Learning. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 1707–1720. [Google Scholar] [CrossRef]
  34. Tritscher, J.; Krause, A.; Hotho, A. Feature Relevance XAI in Anomaly Detection: Reviewing Approaches and Challenges. Front. Artif. Intell. 2023, 6, 1099521. [Google Scholar] [CrossRef]
  35. Abououf, M.; Singh, S.; Mizouni, R.; Otrok, H. Explainable AI for Event and Anomaly Detection and Classification in Healthcare Monitoring Systems. IEEE Internet Things J. 2023, 11, 3446–3457. [Google Scholar] [CrossRef]
  36. Rawal, A.; McCoy, J.; Rawat, D.B.; Sadler, B.M.; Amant, R.S. Recent Advances in Trustworthy Explainable Artificial Intelligence: Status, Challenges, and Perspectives. IEEE Trans. Artif. Intell. 2021, 3, 852–866. [Google Scholar] [CrossRef]
  37. Theissler, A.; Spinnato, F.; Schlegel, U.; Guidotti, R. Explainable AI for Time Series Classification: A Review, Taxonomy and Research Directions. IEEE Access 2022, 10, 100700–100724. [Google Scholar] [CrossRef]
  38. Elkhawaga, G.; Elzeki, O.M.; Abu-Elkheir, M.; Reichert, M. Why Should I Trust Your Explanation? An Evaluation Approach for XAI Methods Applied to Predictive Process Monitoring Results. IEEE Trans. Artif. Intell. 2024, 5, 1458–1472. [Google Scholar] [CrossRef]
  39. Sarker, I.H.; Janicke, H.; Mohsin, A.; Gill, A.; Maglaras, L. Explainable AI for Cybersecurity Automation, Intelligence and Trustworthiness in Digital Twin: Methods, Taxonomy, Challenges and Prospects. ICT Express 2024, 10, 935–958. [Google Scholar] [CrossRef]
  40. Laktionov, I.; Diachenko, G.; Rutkowska, D.; Kisiel-Dorohinicki, M. An Explainable AI Approach to Agrotechnical Monitoring and Crop Diseases Prediction in Dnipro Region of Ukraine. J. Artif. Intell. Soft Comput. Res. 2023, 13, 247–272. [Google Scholar] [CrossRef]
  41. Aslam, N.; Khan, I.U.; Alansari, A.; Alrammah, M.; Alghwairy, A.; Alqahtani, R.; Almushikes, M.; AL Hashim, M. Anomaly Detection Using Explainable Random Forest for the Prediction of Undesirable Events in Oil Wells. Appl. Comput. Intell. Soft Comput. 2022, 2022, 1558381. [Google Scholar] [CrossRef]
  42. Shams Khoozani, Z.; Sabri, A.Q.M.; Seng, W.C.; Seera, M.; Eg, K.Y. Navigating the Landscape of Concept-Supported XAI: Challenges, Innovations, and Future Directions. Multimed. Tools Appl. 2024, 83, 67147–67197. [Google Scholar] [CrossRef]
  43. Sharma, N.A.; Chand, R.R.; Buksh, Z.; Ali, A.B.M.S.; Hanif, A.; Beheshti, A. Explainable AI Frameworks: Navigating the Present Challenges and Unveiling Innovative Applications. Algorithms 2024, 17, 227. [Google Scholar] [CrossRef]
  44. Sledz, A.; Heipke, C. Thermal Anomaly Detection Based on Saliency Analysis from Multimodal Imaging Sources. ISPRS Ann. Photogramm. Remote. Sens. Spat. Inf. Sci. 2021, V-1-2021, 55–64. [Google Scholar] [CrossRef]
  45. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations From Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  46. Shrikumar, A.; Greenside, P.; Kundaje, A. Learning Important Features Through Propagating Activation Differences. In Proceedings of the International conference on machine learning, Long Beach, CA, USA, 10–15 June 2019. [Google Scholar]
  47. Lapuschkin, S.; Binder, A.; Montavon, G.; Müller, K.R.; Samek, W. The LRP Toolbox for Artificial Neural Networks. J. Mach. Learn. Res. 2016, 17, 1–5. [Google Scholar]
  48. Athavale, J.; Baldovin, A.; Graefe, R.; Paulitsch, M.; Rosales, R. AI and Reliability Trends in Safety-Critical Autonomous Systems on Ground and Air. In Proceedings of the 2020 50th Annual IEEE/IFIP International Conference on Dependable Systems and Networks Workshops (DSN-W), Valencia, Spain, 29 June–2 July 2020; pp. 74–77. [Google Scholar] [CrossRef]
  49. Kaplan, E.; Hegarty, C. Understanding GPS/GNSS: Principles and Applications, 3rd ed.; Artech House Publishers: London, UK, 2017. [Google Scholar]
  50. Orolia User Guide. GSG-8 Advanced GNSS/GPS Simulator. Available online: https://www.orolia.com/products/gnss-simulation/gsg-8-gnss-simulator (accessed on 24 March 2023).
Figure 1. (a) Skydel CSG-8 GNSS simulator setup for recording jamming and spoofing signals, (b) Various GNSS disruption recording setups and post-processing using GNSS SDR.
Figure 1. (a) Skydel CSG-8 GNSS simulator setup for recording jamming and spoofing signals, (b) Various GNSS disruption recording setups and post-processing using GNSS SDR.
Sensors 24 08039 g001
Figure 2. (a) Clean GNSS signal in the time domain and its spectrum in the frequency domain recorded using RTL SDR. (b) CWI GNSS signals in the time domain and its spectrum in the frequency domain recorded using RTL SDR. (c) Pulse interference GNSS signal in the time domain and its spectrum in the frequency domain recorded using RTL SDR. (d) Chirp interference GNSS signal in the time domain and its spectrum in the frequency domain recorded using RTL SDR.
Figure 2. (a) Clean GNSS signal in the time domain and its spectrum in the frequency domain recorded using RTL SDR. (b) CWI GNSS signals in the time domain and its spectrum in the frequency domain recorded using RTL SDR. (c) Pulse interference GNSS signal in the time domain and its spectrum in the frequency domain recorded using RTL SDR. (d) Chirp interference GNSS signal in the time domain and its spectrum in the frequency domain recorded using RTL SDR.
Sensors 24 08039 g002aSensors 24 08039 g002b
Figure 3. (a) Correlation maps for time domain features (i) Clean GNSS (ii) Chirp (iii) Pulse (iv) CWI (v) Spoofing (vi) Multipath (vii) MCWI. (b) Correlation maps for frequency domain features (i) Clean GNSS (ii) Chirp (iii) Pulse (iv) CWI (v) Spoofing (vi) Multipath (vii) MCWI.
Figure 3. (a) Correlation maps for time domain features (i) Clean GNSS (ii) Chirp (iii) Pulse (iv) CWI (v) Spoofing (vi) Multipath (vii) MCWI. (b) Correlation maps for frequency domain features (i) Clean GNSS (ii) Chirp (iii) Pulse (iv) CWI (v) Spoofing (vi) Multipath (vii) MCWI.
Sensors 24 08039 g003aSensors 24 08039 g003bSensors 24 08039 g003c
Figure 4. (a) Frequency domain features at different jamming power levels, (b) Time domain features under various jamming power levels.
Figure 4. (a) Frequency domain features at different jamming power levels, (b) Time domain features under various jamming power levels.
Sensors 24 08039 g004aSensors 24 08039 g004bSensors 24 08039 g004cSensors 24 08039 g004d
Figure 5. Boxplot comparison of different disruptions in GNSS signals based on p-values.
Figure 5. Boxplot comparison of different disruptions in GNSS signals based on p-values.
Sensors 24 08039 g005
Figure 6. The proposed XAI framework-based GNSS signal disruption classification model.
Figure 6. The proposed XAI framework-based GNSS signal disruption classification model.
Sensors 24 08039 g006
Figure 7. (a) Frequency domain LIME model decision for correct/incorrect predictions of GNSS signal for different ML models (i) Decision tree, (ii) AdaBoost, (iii) KNN, (iv) Random forest, and (v) SVM. (b) Time domain LIME model decision of correct/incorrect predictions of GNSS signal for different ML models: (i) Decision tree, (ii) AdaBoost, (iii) KNN, (iv) Random forest, and (v) SVM.
Figure 7. (a) Frequency domain LIME model decision for correct/incorrect predictions of GNSS signal for different ML models (i) Decision tree, (ii) AdaBoost, (iii) KNN, (iv) Random forest, and (v) SVM. (b) Time domain LIME model decision of correct/incorrect predictions of GNSS signal for different ML models: (i) Decision tree, (ii) AdaBoost, (iii) KNN, (iv) Random forest, and (v) SVM.
Sensors 24 08039 g007aSensors 24 08039 g007b
Figure 8. (a) Feature contributions based on frequency domain attributes: (i) decision tree, (ii) AdaBoost, (iii) KNN, (iv) random forest, and (v) SVM. (b) Plots of contributions of combinations of time domain features for different types of GNSS signal disruptions: (i) decision tree, (ii) AdaBoost, (iii) KNN, (iv) random forest, and (v) SVM.
Figure 8. (a) Feature contributions based on frequency domain attributes: (i) decision tree, (ii) AdaBoost, (iii) KNN, (iv) random forest, and (v) SVM. (b) Plots of contributions of combinations of time domain features for different types of GNSS signal disruptions: (i) decision tree, (ii) AdaBoost, (iii) KNN, (iv) random forest, and (v) SVM.
Sensors 24 08039 g008aSensors 24 08039 g008bSensors 24 08039 g008c
Figure 9. (a): Frequency domain feature importance comparison using different ML algorithms. (b): Time domain feature importance comparison using different ML algorithms.
Figure 9. (a): Frequency domain feature importance comparison using different ML algorithms. (b): Time domain feature importance comparison using different ML algorithms.
Sensors 24 08039 g009
Figure 10. (a) Waterfall summary plots (time domain), (i) Pulsed; (ii) CWI; (iii) MCWI; (iv) MP; (v) Clean; (vi) Chirp; (vii) Spoofing. (b) Waterfall summary plots (frequency domain), (i) Pulsed (ii) CWI; (iii) Clean; (iv) chirp; (v) MCWI; (vi) MP; (vii) Spoofing.
Figure 10. (a) Waterfall summary plots (time domain), (i) Pulsed; (ii) CWI; (iii) MCWI; (iv) MP; (v) Clean; (vi) Chirp; (vii) Spoofing. (b) Waterfall summary plots (frequency domain), (i) Pulsed (ii) CWI; (iii) Clean; (iv) chirp; (v) MCWI; (vi) MP; (vii) Spoofing.
Sensors 24 08039 g010aSensors 24 08039 g010bSensors 24 08039 g010c
Figure 11. (a) SHAP force plots for GNSS time domain features, From top: (i) pulse, (ii) CWI, (iii) clean, (iv) chirp, (v) MCWI, (vi) multipath, and (vii) spoofing signals. (b) SHAP force plots for GNSS frequency domain features. From top: (i) pulse, (ii) CWI, (iii) clean, (iv) chirp, (v) MCWI, (vi) multipath, and (vii) spoofing signals.
Figure 11. (a) SHAP force plots for GNSS time domain features, From top: (i) pulse, (ii) CWI, (iii) clean, (iv) chirp, (v) MCWI, (vi) multipath, and (vii) spoofing signals. (b) SHAP force plots for GNSS frequency domain features. From top: (i) pulse, (ii) CWI, (iii) clean, (iv) chirp, (v) MCWI, (vi) multipath, and (vii) spoofing signals.
Sensors 24 08039 g011aSensors 24 08039 g011b
Figure 12. ROC one-vs-rest multiclass plot of machine learning models tested under different GNSS signal disruptions.
Figure 12. ROC one-vs-rest multiclass plot of machine learning models tested under different GNSS signal disruptions.
Sensors 24 08039 g012
Figure 13. Confusion matrices (frequency domain features).
Figure 13. Confusion matrices (frequency domain features).
Sensors 24 08039 g013
Table 1. (a) Overall performance of the LIME model in the frequency domain using top six features. (b) Overall performance of LIME model on time domain using top 12 features.
Table 1. (a) Overall performance of the LIME model in the frequency domain using top six features. (b) Overall performance of LIME model on time domain using top 12 features.
(a)
AI ModelAccuracyPrecisionRecallF1-Score
DT74747574
RF75757575
KNN79797979
SVM81828181
AdaBoost29294333
(b)
AI ModelAccuracyPrecisionRecallF1-Score
DT66656465
RF71737273
KNN41414143
SVM1612166
AdaBoost57485750
Table 2. (a): Results of ML models’ prediction performances using frequency domain feature selection method on GNSS dataset under k = 6. (b): Results of ML models’ prediction performances using time domain feature selection method on GNSS dataset under k = 12.
Table 2. (a): Results of ML models’ prediction performances using frequency domain feature selection method on GNSS dataset under k = 6. (b): Results of ML models’ prediction performances using time domain feature selection method on GNSS dataset under k = 12.
(a)
AI ModelClassPrecisionRecallF1-Score
DTMCWI100100100
MP1009799
Chirp192320
Clean333
CWI100100100
Pulse100100100
Spoofing9810099
RFMCWI100100100
MP1009799
Chirp807080
Clean161717
CWI100100100
Pulse100100100
Spoofing9810099
KNNMCWI100100100
MP1009799
Chirp313533
Clean262324
CWI100100100
Pulse100100100
Spoofing9810099
SVMMCWI9110095
MP1009799
Chirp506857
Clean383335
CWI1007284
Pulse1009799
Spoofing9810099
AdaBoostMCWI000
MP000
Chirp5010067
Clean 000
CWI100100100
Pulse5010067
Spoofing000
(b)
AI ModelClassPrecisionRecallF1-score
DTMCWI10099100
MP9910099
Chirp100100100
Clean100100100
CWI100100100
Pulse100100100
Spoofing100100100
RFMCWI100100100
MP1009799
Chirp807080
Clean161717
CWI100100100
Pulse100100100
Spoofing9810099
KNNMCWI343936
MP274333
Chirp404241
Clean262526
CWI999195
Pulse423136
Spoofing301722
SVMMCWI000
MP3248
Chirp000
Clean2869
CWI1007284
Pulse611
Spoofing9810099
AdaBoostMCWI339950
MP000
Chirp41214
Clean000
CWI100100100
Pulse100100100
Spoofing000
Table 3. (a) Relative frequency domain feature importance (feature ranking). (b) Relative time domain feature importance (feature ranking).
Table 3. (a) Relative frequency domain feature importance (feature ranking). (b) Relative time domain feature importance (feature ranking).
(a)
FeatureDTRFKNNSVMAdaBoostAverage Rank
Normalized spectrum kurtosis1111211
Ratio of the variance to the squared mean of the normalized spectrum624312.63
Normalized spectrum flatness243442.833
Ratio of the maximum peak to the second maximum peak of the normalized spectrum565554.34
Normalized spectrum bandwidth356664.34
Single-frequency energy aggregation432232.32
(b)
FeatureDTRFKNNSVMAdaBoostAverageRank
Mean101032127.47
Median9923116.87
Standard deviation5581246.87
Mean absolute deviation 81211106.46
Root mean square error1274644
25th percentile966615.66
75th percentile674596.26
Inter-percentile range345785.45
Skewness889837.27
Kurtosis2311925.45
Entropy411010566
Maximum-to-mean ratio711121179.610
Table 4. (a) ML models’ performance results using different top ‘k’ frequency domain feature selection methods on GNSS dataset. (b) ML models’ performance results using different top ‘k’ time domain feature selection methods on GNSS dataset.
Table 4. (a) ML models’ performance results using different top ‘k’ frequency domain feature selection methods on GNSS dataset. (b) ML models’ performance results using different top ‘k’ time domain feature selection methods on GNSS dataset.
(a)
No. of FeaturesMetricDTRFKNNSVMAdaBoost
K = 1Accuracy0.45000.28570.47500.46070.4223
Precision0.45240.16670.51250.37940.3901
Recall0.45000.28570.47500.46070.4356
F1-score0.45040.18370.48520.37190.3421
K = 2Accuracy0.97500.71430.9786,0.82500.5682
Precision0.97630.56560.97980.85580.5321
Recall0.97500.71430.97860.82500.5789
F1-score0.97490.61720.97850.81580.5971
K = 3Accuracy0.98670.85710.97880.92500.6210
Precision0.92440.76670.94250.93840.6535
Recall0.94620.85710.95320.92500.7152
F1-score0.97640.80230.96420.92340.7543
K = 4Accuracy0.98990.96720.96350.99640.6458
Precision 10.99890.97850.99650.7342
Recall110.98260.99640.6735
F1-score1110.99640.7211
K = 5Accuracy10.992910.99640.6712
Precision10.993210.99650.7443
Recall10.992910.99640.6990
F1-score10.992910.99640.7124
K = 6Accuracy10.857110.92500.6876
Precision10.766710.93840.7342
Recall10.857110.92500.6735
F1-score10.802310.92340.7211
(b)
No. of FeaturesMetricDTRFKNNSVMAdaBoost
K = 1Accuracy 0.62500.57140.70360.42860.3246
Precision 0.62520.40480.70380.19170.0869
Recall0.62500.57140.70360.42860.3458
F1-score0.62430.45240.70260.26330.3675
K = 2Accuracy0.45240.57140.70360.45710.4412
Precision0.62520.40480.70380.24890.2357
Recall0.6250,0.57140.70360.45710.4043
F1-score0.62430.45240.70260.32050.3215
K = 3Accuracy0.7500,0.85360.77860.77860.6548
Precision0.74890.78220.77860.73130.5436
Recall0.75000.85360.77860.77860.4265
F1-score0.74940.80600.77850.726243,445
K = 4Accuracy0.7480,0.85360.77860.82500.6550
Precision0.74210.78220.77820.75760.5436
Recall0.74640.85360.77860.82500.4336
F1-score0.74390.80600.77830.77720.4987
K = 5Accuracy0.74800.83930.77860.77500.6548
Precision0.73940.76920.75880.77400.5436
Recall0.74640.83930.82860.77500.4265
F1-score0.74250.79160.78090.774143,445
K = 6Accuracy0.74640.84290.7750,0.82860.6753
Precision0.73940.77270.77400.7588,0.5964
Recall0.7464,0.84290.77500.82860.4413
F1-score0.74250.79520.77410.78090.4842
K = 7Accuracy0.74640.84290.77500.82860.6759
Precision0.73940.77270.77400.75880.61225
Recall0.74640.84290.77500.82860.6709
F1-score0.74250.79520.77410.78090.4007
K = 8Accuracy0.74640.84290.77500.82860.6986
Precision0.73940.77270.77400.75880.6543
Recall0.74640.84290.77500.8286,0.6709
F1-score0.74250.79520.77410.78090.4346
K = 9Accuracy0.74640.84290.77500.82860.6986
Precision0.73940.77270.77400.75880.6543
Recall0.74640.84290.77500.82860.6709
F1-score0.74250.79520.77410.78090.4346
K = 10Accuracy0.74640.84290.77500.82860.6986
Precision0.73940.77270.77400.75880.6543
Recall0.74640.84290.77500.82860.6709
F1-score0.74250.79520.77410.78090.4346
K = 11Accuracy0.74640.84290.77500.82860.6986
Precision0.73940.77270.77400.75880.6543
Recall0.74640.84290.77500.82860.6709
F1-score0.74250.79520.77410.78090.4346
K = 12Accuracy0.74640.84290.77500.82860.6986
Precision0.73940.77270.77400.75880.6543
Recall0.74640.84290.77500.82860.6709
F1-score0.74250.77410.78090.78620.4346
Table 5. Results per class for every feature selection method under k = 4.
Table 5. Results per class for every feature selection method under k = 4.
ModelClassSHAPPCABackward EliminationForward Selection
DTMetricAccPrecRecF1AccPrecRecF1AccPrecRecF1AccPrecRecF1
MCWI100100100100979597929795979297959792
MP951009799988991909889919098899190
Chirp24192320211921202317322325222623
Clean72767883821812121218151216222309
CWI100100100100988890939888909398889083
Pulse100100100100977681909776819097728079
Spoofing989810099979597929795979297959792
RFMCWI94979397988991909889919098899190
MP981009799959492918893928990918886
Chirp98807080727671766766636160656667
Clean19161717181317151213151212131311
CWI100100100100977681909776819097768190
Pulse100100100100979597929795979297959792
Spoofing989810099988991909889919098899190
KNNMCWI100100100100938587808164727277757572
MP991009799958887807565646061646663
Chirp37313533343224242931332930282933
Clean27262324323731241919242120273235
CWI100100100100959493909478737168727378
Pulse100100100100979493908991898879818081
Spoofing1009810099879795937175737470717271
SVMMCWI959110095756575797068697864615664
MP961009799988890828088817979887979
Chirp65506857977681909776819097768190
Clean34383335979597929795979297959792
CWI871007284988991909889919098899190
Pulse981009799787672706769737277747771
Spoofing979810099847475778681716973767772
AdaBoostMCWI25232623242524192221242327222426
MP45464515443237394241323339414034
Chirp535010067535451434435453248424335
Clean54564555545241445043474243413938
CWI56545645455142414448504652433944
Pulse49506567535456474556545054524646
Spoofing51565445594646433443344441434239
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Elango, A.; Landry, R.J. XAI GNSS—A Comprehensive Study on Signal Quality Assessment of GNSS Disruptions Using Explainable AI Technique. Sensors 2024, 24, 8039. https://doi.org/10.3390/s24248039

AMA Style

Elango A, Landry RJ. XAI GNSS—A Comprehensive Study on Signal Quality Assessment of GNSS Disruptions Using Explainable AI Technique. Sensors. 2024; 24(24):8039. https://doi.org/10.3390/s24248039

Chicago/Turabian Style

Elango, Arul, and Rene Jr. Landry. 2024. "XAI GNSS—A Comprehensive Study on Signal Quality Assessment of GNSS Disruptions Using Explainable AI Technique" Sensors 24, no. 24: 8039. https://doi.org/10.3390/s24248039

APA Style

Elango, A., & Landry, R. J. (2024). XAI GNSS—A Comprehensive Study on Signal Quality Assessment of GNSS Disruptions Using Explainable AI Technique. Sensors, 24(24), 8039. https://doi.org/10.3390/s24248039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop