4.1. BDS Broadcast Ephemeris Orbit Parameter Prediction
The subset of data was processed using methods such as differencing. As a predictive multivariate time-series anomaly detection model, the broadcast ephemeris data, excluding specific parameters like satellite clock bias and health indicators, predominantly display periodic time-series characteristics. This study integrated the findings from the literature and the evaluation results to forecast parameters, including , , , , , and clock bias. Due to the systemic relationships among the internal parameters of the broadcast ephemeris data, four parameters—, , , and —were included as related variables in the prediction process.
The prediction ranges were determined by capturing short- to medium-term trends and periodicity, specifically involving 7-step forecasts based on 22 steps, 10-step forecasts based on 40 steps, and 25-step forecasts based on 110 steps. With the exception of clock bias and , all other parameters met the aforementioned criteria, where epochs ranged from 40 to 120. The interdependencies among parameters allowed this study to infer positional and temporal corrections through orbital mechanics and time-adjustment formulas.
To forecast the broadcast ephemeris time series, a sliding-window strategy was employed, dividing the training and testing sets into several sliding time windows
T. The overall loss function was defined as the root mean square error (RMSE) between the true values and predicted values for future time windows. The error metrics displayed in the figures indicate excellent performance in the context of broadcast ephemeris predictions. Selected error results are presented in the accompanying figure (
Figure 16).
An analysis of the results indicated that for parameters such as () and (), the LSTM exhibited a lag in predicted values relative to the actual data within certain intervals. Longitudinal observations revealed that this lag primarily occurred during the descending phases of the data trends. We hypothesize that the sliding time window strategy resulted in a delayed response of the model when faced with rapidly changing trends. To address this issue, we introduced an attention mechanism and a time-embedding mechanism to extract richer temporal features, significantly enhancing the model’s predictive accuracy during periods of decline.
Table 4 presents the average prediction errors for the orbital parameters. Initially, the LSTM model demonstrated strong performance in processing orbital parameters with significant volatility and periodic patterns, achieving satisfactory prediction accuracy. The incorporation of the attention mechanism further enhanced the model’s predictive capabilities, significantly improving the fit between the predicted values and the actual measurements, with error rates consistently maintained below 10%. Additionally, the introduction of threshold control methods optimized the model’s predictive accuracy.
An analysis of the RMSE values for four models—LSTM, A-LSTM, TE-LSTM, and IF-TEA-LSTM—revealed their performance, as shown in
Figure 17. The LSTM model struggled to capture the nonlinear variations in BDS satellite orbital parameters, particularly for
, which is influenced by external factors such as multipath effects and space weather, resulting in relatively high RMSE values. However, the integration of the attention mechanism and time-embedding mechanism significantly improved feature extraction in the A-TE-LSTM model, leading to notable improvements in prediction accuracy.
Figure 17.
Prediction accuracy and performance of
under different reference frames. The six subplots (
a–
f) represent this parameter under the following forecast time horizons: (
a) 24 h forecast, 1 h interval; (
b) 96 h forecast, 1 h interval; (
c) 7 d forecast, 1 h interval; (
d) 15 d forecast, 1 h interval; (
e) 30 d forecast, 1 h interval; (
f) 90 d forecast, 1 h interval. The subplot distribution in
Figure 18,
Figure 19,
Figure 20 and
Figure 21 follows the same structure.
Figure 17.
Prediction accuracy and performance of
under different reference frames. The six subplots (
a–
f) represent this parameter under the following forecast time horizons: (
a) 24 h forecast, 1 h interval; (
b) 96 h forecast, 1 h interval; (
c) 7 d forecast, 1 h interval; (
d) 15 d forecast, 1 h interval; (
e) 30 d forecast, 1 h interval; (
f) 90 d forecast, 1 h interval. The subplot distribution in
Figure 18,
Figure 19,
Figure 20 and
Figure 21 follows the same structure.
The performance of varied across different time intervals when predicted using the TE-LSTM, A-LSTM, and A-TE-LSTM models. The RMSE values for TE-LSTM consistently improved over LSTM, ranging from 9.4% to 10.5%. In contrast, A-LSTM exhibited greater fluctuations, decreasing by 4.37% in the 24-hour forecast and achieving a maximum improvement of 20.92% in longer-term predictions. Notably, A-TE-LSTM demonstrated the most significant performance, with enhancements ranging from 12.61% to 25.06%, particularly evident in long-term forecasting scenarios.
In the prediction of the parameter, all four models demonstrated commendable performance. The TE-LSTM model showed an average improvement ranging from 7.18% to 13.42%, indicating a stable enhancement effect. The A-LSTM model also exhibited notable gains, with a maximum increase of 22.46%. However, the A-TE-LSTM model stood out overall, achieving improvements between 21.90% and 27.29%, further validating its superiority in feature extraction and accuracy. Overall, all three models performed admirably in predicting the parameter.
The performance of the models in predicting the parameter varied significantly across different time intervals. The TE-LSTM model demonstrated an average enhancement ranging from 6.90% to 15.81%. The A-LSTM model also showed substantial improvements, achieving a maximum increase of 25.92%. Notably, the A-TE-LSTM model excelled, achieving an overall enhancement between 19.71% and 38.39%, particularly remarkable in short-term forecasts (24 h and 96 h), thereby showcasing its robust capability for feature extraction. The pronounced differences in predictive performance across various time periods can be attributed to the influence of nonlinear errors, such as multipath effects, which significantly impacted the parameter. These error sources can accumulate over time, severely affecting overall prediction accuracy.
The parameters and exhibited minimal correction effects and were less influenced by external disturbances, resulting in relatively low variability. Consequently, the performance of the TE-LSTM, A-LSTM, and A-TE-LSTM models showed significant discrepancies. For , the TE-LSTM model demonstrated suboptimal performance, with average improvements ranging from −6.26% to −0.60%, indicating poor predictive capability. The A-LSTM model exhibited stable performance in short-term predictions (24 h and 96 h), albeit with limited improvements; negative values were observed in some cases, such as 30-day and 90-day forecasts. The A-TE-LSTM model performed slightly better in certain scenarios, particularly in the 7-day and 15-day forecasts, achieving improvements of 7.22% and 6.12%, respectively. However, the overall enhancement remained constrained, likely due to frequent orbital adjustments of in the short term, which complicated the modeling of nonlinear errors.
The prediction results for revealed poor performance in the TE-LSTM model’s 24-h forecast, with a negative improvement of −5.20%. However, in longer-term predictions, particularly for 7-day and 15-day forecasts, improvements reached as high as 6.28%. The A-LSTM model exhibited relatively stable performance, while the A-TE-LSTM model excelled, maintaining high improvement rates across all forecasts, especially in the 96-hour and 7-day predictions, with enhancements of 15.17% and 13.27%, respectively.
The presence of substantial noise in some parameters contributed to an overall increase in prediction errors. This noise manifested as random errors and irregular fluctuations; for instance, the RMSE metric for displayed significant volatility, indicating that noise considerably affected the model’s accuracy in later predictions. Even the IF-TEA-LSTM model struggled to fully capture these noise effects, complicating its ability to mitigate irregular errors and thus impacting overall predictive performance.
Through an in-depth analysis of the RMSE, we observed significant characteristic differences among satellite types. The GEO-type LSTM model exhibited consistently high RMSE values, indicating considerable errors in processing GEO orbital data. In contrast, IGSO and MEO types excelled in error control, showcasing lower RMSE values and tighter distributions. Notably, certain models demonstrated superior performance for specific orbit types, particularly the A-TE-LSTM model, which consistently achieved lower RMSE values, highlighting its potential advantages in particular application scenarios, as illustrated in
Figure 22.
Therefore, it is crucial to distinguish between different orbit types, with particular emphasis on the classification of GEO orbital types, in order to mitigate substantial errors typically associated with these types compared to others. The primary objective of this study is anomaly detection, where RMSE accuracy and the presence of minor abrupt anomalies do not significantly affect the overall detection performance.
Considering that the model processes multi-parameter, multi-variable time-series data with high correlation among parameters, anomalies can still be effectively identified by analyzing trends and fluctuations. For example, while parameters such as and may exhibit short-term volatility, a joint analysis of these parameters can uncover underlying anomaly patterns, thus enhancing the accuracy of detection.
4.2. BDS Broadcast Ephemeris Clock Parameter Prediction
The clock bias is treated separately due to its inherent complexity, which distinguishes it from the orbital parameters. The performance of the LSTM model on raw clock bias data was found to be limited, leading to poor fitting results. To address this issue, a stabilization process was applied to the clock bias data, incorporating outlier detection techniques to identify anomalies. Following this, segmented linear interpolation was utilized to replace the anomalous data points. Predictive experiments conducted on the refined dataset demonstrated the effectiveness of this approach, significantly improving the accuracy of clock bias predictions.
Figure 23 presents a comparison between the actual and predicted values of the broadcast ephemeris clock bias sequence. In theory, if the accuracy of the actual values exceeds 99.5%, the predicted values should closely align with them, forming a nearly smooth diagonal line. This alignment would indicate the model’s capability to effectively capture the fluctuations of non-stationary sequences, such as clock bias. However, traditional time-series models like LSTM often struggle to accurately fit such complex nonlinear sequences, resulting in diminished prediction performance.
To address this challenge, second-order differencing was applied to the clock bias sequence, reducing trend-related interference and allowing the model to extract more meaningful information. As shown in subplot (b) of
Figure 23, the predictions after differencing reveal significant deviations from the ideal diagonal line, highlighting anomalous timestamps.
Additionally, specific experimental satellites were selected for model validation using representative samples. The RMSE values in
Table 5 further demonstrate the LSTM model’s performance, particularly noting the improvements after differencing. Although no significant anomalies were detected in the sample, minor deviations were observed in the model’s predictions. These deviations remained within acceptable limits, reinforcing the conclusion that differencing preprocessing enhances the predictive accuracy of complex non-stationary sequences. These results emphasize the crucial role of data preprocessing in improving the performance of time-series prediction models.
As shown in the table, the IF-TEA-LSTM model consistently outperformed other models in clock bias prediction and anomaly detection. The following points summarize its performance:
For satellites C21 and C22, within a 96-hour short-term forecast window, the RMSE of the IF-TEA-LSTM model improved by 62.56% and 63.52% compared to the standard LSTM, 36.82% and 34.62% compared to A-LSTM, and 60.70% and 63.45% compared to TE-LSTM.
For satellites C26 and C27, the IF-TEA-LSTM model showed RMSE improvements of 69.75% and 77.61% over LSTM, 52.12% and 44.91% over A-LSTM, and 70.38% and 76.75% over TE-LSTM.
It is important to note that the MAE, RMSE, MSE, and range values in the table were derived from normalized and differenced data. Therefore, the units differ from those of the raw data (e.g., clock bias measured in seconds, refined to nanoseconds). These metrics were used for the relative model performance comparison and do not represent the final absolute errors. Upon denormalization to the original scale, the error metrics will revert to units consistent with the raw data.
4.3. Performance of BDS Broadcast Ephemeris Data Anomaly Detection
We analyzed the broadcast ephemeris data of BDS2 and BDS3 satellites, focusing on the performance differences across different orbits (GEO, IGSO, MEO) using an RNN, a GRU, LSTM, and their enhanced versions. The experimental results demonstrate that LSTM outperformed the RNN and GRU in time-series modeling, showing superior precision, recall, and F1-score. Although the GRU offers improved computational efficiency, its recall rate was relatively lower when handling long-term broadcast ephemeris data, indicating its limitations in capturing long-term dependencies. Moreover, the GRU’s extended training time, without corresponding performance gains, suggests structural limitations in learning long sequences. The detailed results for each satellite type are shown in
Table 6.
For the GEO orbital data, the IF-TEA-LSTM model demonstrated the best performance, achieving 86.43% precision and 74.57% recall, with A-LSTM and LSTM closely following. This indicates that more complex models have a clear advantage in capturing the temporal features of GEO data, particularly in identifying anomalies. In contrast, the GRU and RNN models exhibited lower recall, revealing their limitations in handling GEO data.
For the IGSO data, the IF-TEA-LSTM model continued to excel, with precision reaching 89.64% and recall at 79.62%. A-LSTM and LSTM also performed well in this orbit, especially in terms of precision and F1-score, underscoring the effectiveness and adaptability of LSTM-based models in processing dynamic data from non-synchronous orbits.
In the MEO orbital experiments, IF-TEA-LSTM maintained its lead, with precision at 93.42% and an F1-score of 89.21%, demonstrating strong generalization capabilities in processing medium Earth orbit satellite data. A-LSTM and TE-LSTM also showed robustness, while the GRU and RNN models continued to struggle in terms of recall, reflecting their inability to capture the complex temporal patterns of MEO data.
The results indicate that orbital conditions and data characteristics significantly influence model performance. GEO satellites, due to their large static range and limited dynamic properties, underperformed compared to IGSO and MEO satellites. This highlights the need for more specialized methods when analyzing GEO data. In contrast, the higher dynamism in IGSO and MEO data poses challenges for machine learning models, but with extensive training data (e.g., over 1–3 years, exceeding 1.15 million records), models exhibit strong adaptability and predictive power.
Notably, the TE-LSTM model improved average performance by 4.31% for GEO orbits, while the improvements for IGSO and MEO orbits were 0.878% and 2.16%, respectively. This suggests that the sensitivity of TE-LSTM to threshold constraints is more pronounced for GEO data, where stronger threshold enforcement significantly enhances anomaly detection. For other orbits, the effect is relatively smaller. Therefore, optimizing threshold constraints for different orbit types is crucial for improving anomaly detection accuracy.
Table 7 summarizes the average performance of various models across different orbits, further validating these findings.
The overall experimental results indicate that the proposed IF-TEA-LSTM model demonstrates significant advantages across various evaluation metrics. This finding validates the effectiveness of the methods introduced in this study, showcasing the model’s strong generalization capabilities and precise predictive power. By deeply learning temporal features, the model effectively identifies and captures hidden anomaly patterns, providing a broadly applicable and efficient reference framework for future broadcast ephemeris data processing and anomaly detection.
Furthermore, by integrating a clock bias scoring mechanism, the IF-TEA-LSTM model enhances its anomaly detection capabilities in practical applications. Given the critical role of clock bias data in satellite navigation systems, the introduction of this scoring mechanism not only improves the model’s ability to identify anomalous data points in the broadcast ephemeris but also allows for more accurate localization and classification of potential anomalies. This provides robust technical support for the health monitoring and subsequent maintenance of satellite orbital data. Therefore, the series of innovative improvements based on LSTM and its derivatives not only enhances anomaly detection performance across various orbital data but also lays a solid theoretical and practical foundation for future processing and precise prediction of more complex, multidimensional ephemeris data.