Next Article in Journal
Complementary Design of Two Types of Signals for Avionic Phased-MIMO Weather Radar
Previous Article in Journal
Digital Twin Modeling for Landslide Risk Scenarios in Mountainous Regions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction and Performance of BDS Satellite Clock Bias Based on CNN-LSTM-Attention Model

Faculty of Land Resources Engineering, Kunming University of Science and Technology, Kunming 650093, China
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(2), 422; https://doi.org/10.3390/s26020422
Submission received: 15 December 2025 / Revised: 5 January 2026 / Accepted: 6 January 2026 / Published: 8 January 2026
(This article belongs to the Section Navigation and Positioning)

Abstract

Satellite Clock Bias (SCB) is a major source of error in Precise Point Positioning (PPP). The real-time service products from the International GNSS Service (IGS) are susceptible to network interruptions. Such disruptions can compromise product availability and, consequently, degrade positioning accuracy. We introduce the CNN-LSTM-Attention model to address this challenge. The model enhances a Long Short-Term Memory (LSTM) network by integrating Convolutional Neural Networks (CNNs) and an Attention mechanism. The proposed model can efficiently extract data features and balance the weight allocation in the Attention mechanism, thereby improving both the accuracy and stability of predictions. Across various forecasting horizons (1, 2, 4, and 6 h), the CNN-LSTM-Attention model demonstrates prediction accuracy improvements of (76.95%, 66.84%, 65.92%, 84.33%, and 43.87%), (72.59%, 65.61%, 74.60%, 82.98%, and 51.13%), (70.45%, 68.52%, 81.63%, 88.44%, and 60.49%), and (70.26%, 70.51%, 84.28%, 93.66%, and 66.76%), respectively, across the five benchmark models: Linear Polynomial (LP), Quadratic Polynomial (QP), Autoregressive Integrated Moving Average (ARIMA), Backpropagation Neural Network (BP), and LSTM models. Furthermore, in dynamic PPP experiments utilizing IGS tracking stations, the model predictions achieve positioning accuracy comparable to that of post-processed products. This proves that the proposed model demonstrates superior accuracy and stability for predicting SCB, while also satisfying the demands of positioning applications.

1. Introduction

Real-time Precise Point Positioning (PPP) technology is now used in various fields, including surveying, geographic information systems, and unmanned vehicles. The growing demand for high-accuracy localization underscores the necessity of a high-precision time system [1,2]. The PPP technique hinges on continuous and precise satellite orbits and clocks. Since April 2013, the International GNSS Service (IGS) has offered Real-Time Service (RTS), providing global users with satellite orbits, clock differentials, and other products for both real-time and post-processing applications [3,4,5]. The Final Precise Clock Products provided by IGS have a high accuracy of 75 ps but are published as post-processing data with a delay of about two weeks, making them unusable for real-time applications. The Ultra-Rapid Clock Products are for real-time service, but its accuracy is limited to 3 ns, insufficient for high-precision positioning [6]. At the same time, limitations in network connectivity and localization areas can cause interruptions in the real-time RTS data stream, which in turn affects the accuracy of localization [7]. Thus, forecasting Satellite Clock Bias (SCB) with high-precision post-processing data has been becoming a new method to derive the correction with feasible precision for real-time positioning [8,9,10].
The current commonly used models for SCB prediction include the Quadratic Polynomial (QP) model [11], Gray Model (GM) [12,13], Spectral Analysis (SA) [14], Kalman Filter (KF) [15], and Autoregressive Integrated Moving Average (ARIMA) [16]. These models have greatly improved SCB forecasting, while each model has specific advantages and limitations because of the nonlinear and stochastic nature of SCB. The QP model is computationally simple and delivers superior accuracy for short-term predictions, but the stability and accuracy decrease with increasing forecasting period. The GM model assumes a smooth, exponential trend in the underlying function; otherwise, its predictive performance deteriorates significantly. The KF model predicts SCB by incorporating the motion properties of atomic clocks and stochastic a priori knowledge. The ARIMA model predicts most accurately when its p and q parameters are chosen appropriately. The SA model can achieve higher prediction accuracy, but a very long and stable time series are required. Satellite atomic clocks are time-varying complex and influenced by multiple factors, leading to complexity in SCB estimation. This is precisely where machine learning techniques specialize because they can generate predictions by learning patterns and relationships from historical data in combination with mathematical algorithms [17]. Wang et al. [18] used a Radial Basis Function (RBF) model to predict SCB of GPS satellites and achieved notable prediction accuracy. Similarly, Wang et al., [19] proposed a Wavelet Neural Network (WNN)-based SCB prediction model, which shows improved prediction accuracy compared to traditional QP and GM based models. However, the RBF method lacks a robust theoretical basis to select sample parameters, and the determination of the WNN structure remains challenging.
Among the many machine learning models, Recurrent Neural Networks (RNNs) are another widely used machine learning method, which utilizes the characteristics of network structure to search for sequence correlation by directionally connecting nodes into a loop. However, RNNs are prone to extreme nonlinear behavior due to multiple propagation during parameter updating. Subsequently, researchers introduced the “gate” mechanism and proposed the Long Short-Term Memory Neural Network (LSTM), which effectively solved the parameter problem of RNNs and demonstrated improved prediction accuracy. Currently, the forecasting of SCBs using LSTM networks has also achieved good results, and the LSTM forecasting algorithms [20,21] have fitted the satellite clock difference data to make predictions, which have achieved good prediction results. The QP-LSTM model [22] integrates a conventional model to enhance its performance in predicting a single data type. The LSTM-Attention model [23] leverages a self-attention mechanism to effectively balance global attention with local feature capture, thereby enhancing both prediction accuracy and stability.
Building upon the advancements of LSTM and its variants, existing models have significantly improved the prediction of SCB. However, two key aspects warrant further exploration to enhance performance. First, from an architectural perspective, most approaches focus on optimizing a single component, leaving the potential of a coherent, synergistic integration of these complementary strengths under-explored. Second, from a data utilization perspective, the prevailing single-satellite paradigm predicts each clock bias independently, potentially overlooking valuable inter-satellite correlations and constellation-level systematic behaviors that could serve as informative auxiliary features. To address these gaps, this study proposes a novel hybrid CNN-LSTM-Attention architecture that not only integrates the strengths of these components into a unified framework but also innovatively employs a multivariate input strategy to leverage data from other satellites as contextual features. First, the Convolutional Neural Network (CNN) extracts effective features in satellite clock offsets through the operation of convolutional kernels to obtain an accurate representation of the main features. Then LSTM simulates the nonlinear trend of satellite clock deviation and extracts the timing change information. Finally, the attention mechanism applies adaptive weights to the LSTM outputs to generate the final SCB prediction. Among them, we attempted a multivariate input prediction approach. In the CNN-LSTM-Attention model, the SCB of a single satellite is taken as the primary prediction target, while data from other satellites are incorporated as feature inputs into the prediction model. This process is sequentially carried out to complete the prediction tasks for all BDS satellites. Finally, we conducted simulated real-time PPP experiments using the predicted satellite clock error data. A rigorous quality assessment of IGS tracking station data confirms the practical utility and robustness of the proposed model.

2. Data Sources

We evaluated the proposed model and algorithm for clock difference forecasting. IGS was utilized to provide after-the-fact precision satellite clock difference data, which was downloaded from https://cddis.nasa.gov/archive/gnss/products (accessed on 25 June 2025). The data are selected from DOY 11, 2024 with a sampling interval of 30 s. This study focuses on the modeling, forecasting, and analysis of post-processed high-precision satellite clock offsets from the BDS system. The BDS constellation includes satellites in three types of orbits: geostationary orbit (GEO), inclined geosynchronous orbit (IGSO), and medium-Earth orbit (MEO), equipped with either rubidium or hydrogen atomic frequency standards. For this analysis, we selected only satellites with complete SCB data, defined as the uninterrupted 24 h series sampled at 30 s intervals (containing 2880 data points). Table 1 presents their designations, orbit types, and onboard clock types.
The model’s prediction performance was evaluated through real-time kinematic PPP experiments. For this purpose, observational data from 30 globally distributed tracking stations of the International GNSS Service were utilized. The specific coordinates of these stations are provided in Table 2, and their geographical distribution is illustrated in Figure 1. The IGS stations are shown as red triangles in Figure 1. Among them, the IGS station is sourced from the WHU data center, and the station data is sourced from ftp://igs.gnsswhu.cn/pub/gps/data/daily/ (accessed on 15 September 2025).

3. Model and Methodology

3.1. CNN Principle

CNNs are feedforward neural networks that utilize convolutional computations and deep architectures. Initially developed for image classification, object detection, image segmentation, and other visual tasks, CNN have achieved remarkable success in the field of computer vision. Moreover, they demonstrate significant potential across multiple domains, spanning the fields of speech recognition, natural language processing, and time series forecasting [24]. The core of convolutional computation is the operation of sliding kernels across input data to identify locally relevant features. These local features can be hierarchically combined to form higher-level abstract representations, thereby enabling effective analysis and understanding of the input data. In this study, the CNN component is employed to extract locally relevant features from the SCB data. This is achieved through the sliding operation of convolutional kernels across the input sequence, which hierarchically identifies and combines patterns to form effective feature representations for subsequent processing. The convolutional operation in the CNN model is implemented by sliding convolutional kernels over the input sequence. This process effectively captures local dependencies and spatial patterns within the data, and its computation can be formally described as follows:
h k = σ ( W k x + b k )
W k x i , y = m = 0 α 1 n = 0 β 1 w m , n × x i + m , y + n
where x is the input data; h k denotes the output data; σ denotes the sigmoid activation function; W k and b k are the weight coefficients and bias functions, respectively; k denotes the number of convolutional kernels of the network; denotes the discrete convolution operation; α and β denote the size parameters of the convolution kernel; w m , n is a matrix of weight functions and m and n are the rows and columns of the matrix.

3.2. LSTM Principle

The LSTM network is a specialized type of RNN designed to model temporal sequences and capture long-range dependencies. Its core innovation lies in the introduction of a gated mechanism, which overcomes the vanishing and exploding gradient problems inherent in standard RNNs when processing long sequences [25,26]. This mechanism enables LSTM to effectively control and retain information over extended time periods, making it particularly suited for time series analysis and prediction tasks. As illustrated in Figure 2, this structure consists of a cell state and three regulatory gates (input, forget, and output gates), which collectively manage the flow of information. In the LSTM unit, the information flow is regulated by three gates: the forget gate, the input gate, and the output gate. The detailed computational process is as follows:
f t = σ ( W f h t 1 , x t + b f )
i t = σ ( W i h t 1 , x t + b i )
o t = σ ( W o h t 1 , x t + b o )
C t = tanh ( W C h t 1 , x t + b C )
C t = f t C t 1 + i t C t
h t = o t tanh ( C t )
where σ denotes the sigmoid activation function; tanh is the hyperbolic tangent activation function; f t , i t and o t are the forget gate, input gate and output gate, respectively; W and b are the weight coefficients and bias function, respectively; x t , h t , C t and C t are the input information, output information, candidate cell state and updated cell state at time t, respectively.
Each LSTM cell contains three gates: forget, input, and output. The forget gate regulates the retention of prior information; the input gate regulates the integration of new data; and the output gate regulates the exposure of the cell state. Collectively, these gates enable long-term memory retention, allowing the model to handle long sequences more effectively.

3.3. CNN-LSTM-Attention Model

The attention mechanism is inspired by human cognition, which selectively focuses on salient information while filtering out irrelevant stimuli from complex, high-volume inputs. In machine learning, this is implemented as a computational resource allocation strategy that amplifies key features and suppresses less useful ones, effectively learning to assign varying weights to different parts of the input [27].
The structure of the attention mechanism, depicted in Figure 3, is built upon three primary vectors, namely the Query, Key, and Value. The query vector refers to the range of the query and represents the feature vector of subjective awareness. The Query vector serves as the primary reference for attention. It is generated from either the input sequence or its previous hidden state. The key vector is a vector of salient feature information about the object. It represents the feature vector of each input element, against which the Query is matched to compute the attention weights. The value vector is the feature vector representing the object itself, which usually appears in pairs with the key vector. The Value vector represents the actual output for each input element, and the final attention output is computed as a weighted sum of these values based on the attention weights. These three vectors cooperate with each other to realize the calculation process of the attention mechanism. Through the attention convergence of Query and Key to realize the attention weight allocation to Value, to generate the final output results.
We propose a CNN-LSTM-Attention model for SCB prediction, as illustrated in Figure 4. The architecture sequentially integrates an input layer, a CNN module, an LSTM layer, an attention mechanism, and a fully connected output layer. The CNN module, comprising two 1D convolutional layers followed by max-pooling operations, extracts local features and short-term patterns from the input SCB time series while reducing dimensionality and enhancing robustness. The extracted features are then fed into the LSTM layer, which captures long-term trends and complex nonlinear dependencies in the data through its gated cell structure, outputting a sequence of hidden states that encode temporal information.
Subsequently, an attention mechanism assigns adaptive weights to the LSTM hidden states via a fully connected layer and a SoftMax function, quantifying the contribution of each time step to the final prediction. A context vector is generated as the weighted sum of these states, emphasizing the most relevant historical information. This integrated design ensures synergistic cooperation: the CNN focuses on local feature extraction, the LSTM models long-term temporal dependencies, and the attention mechanism dynamically highlights critical time steps. The context vector is finally passed through a fully connected layer to produce the SCB prediction, collectively improving both the accuracy and stability of the forecasting results.
Subsequently, researchers employed the Pearson correlation analysis method to investigate the inter-satellite correlations among different BeiDou satellites [28] and proposed a parallel predictive multi-satellite joint clock error forecast model. By integrating CNN with LSTM networks, this model utilizes an inter-satellite correlation-based multivariate prediction approach to simultaneously forecast clock errors for multiple satellites, thereby enhancing prediction performance. This study explores a multivariate forecasting approach, as illustrated in Figure 5, where the SCB of a single satellite serves as the primary prediction target, while data from other satellites are incorporated as input features into the model. The CNN-LSTM-Attention model is utilized to sequentially complete the SCB prediction tasks for all BDS satellites.
In this context, CX denotes the SCB values corresponding to BDS satellites with different PRN numbers. The SCB data from all satellites within the entire constellation will be utilized as input for the model. The system will then proceed to sequentially generate SCB predictions for each BDS satellite.

3.4. Data Pre-Processing and Parameter Settings

SCB constitutes a univariate time series characterized by a limited number of sampling points and typically exhibits an underlying linear trend for a given satellite. The generation of SCB is influenced by a combination of gravitational, velocity, and relativistic effects, making it difficult to derive accurate features directly from these physical influences. In practice, ground control systems worldwide determine the SCB by comparing the satellite’s clock time with the standard GPS time using historical tracking data. Forecasts are then generated either by modeling the specific clock’s characteristics or by applying extrapolation techniques to the analyzed deviation trends.
When using deep learning network models for SCB prediction, they are affected by raw data series with linear trends, resulting in poor extraction of nonlinear data features. Therefore, we remove the constant term from the SCB data sequence by applying first-order differencing according to Equations (11) and (12). This centers the differenced series around zero, thereby making its underlying variation and trend more pronounced for subsequent analysis.
First, the Median Absolute Deviation (MAD) method [29] is employed to detect and process coarse differences in the raw satellite clock difference sequence, where the bit number equation is:
T M A D = M e d i a n | y i m i d | 0.6745
where mid means the value in the middle of the clock difference sequence. If | y i | > m i d + x T M A D (x ∈ (1, 5), and usually x = 3), it is considered to be an outlier and replaced by the value calculated by segmented linear interpolation. The processed SCB data are arranged into n-dimensional sequences:
X = { x 1 , x 2 , x 3 , , x t }
where x denotes the original SCB data sequence from the first data point (i.e., the first observation in the series) to the t-th data point. The data sequence is then differenced once to obtain a first-order differenced data sequence.
Δ x i = x i + 1 x i
X Δ = { Δ x 1 , Δ x 2 , Δ x 3 , , Δ x t 1 }
However, for each satellite, the first-order differenced SCB series values have a different range of numerical magnitudes. To mitigate scale discrepancies after first-order differencing and prevent any single variable from dominating the analysis, the data were normalized to the [0, 1] interval using min-max scaling. This process enhances model stability and generalization. The normalization is defined as:
Δ x i = Δ x i X Δ max X Δ max X Δ min
X Δ = { Δ x 1 , Δ x 2 , Δ x 3 , , Δ x t 1 }
where X Δ max and X Δ min are the maximum and minimum values in the first-order differenced SCB data sequence, respectively, and Δ x i denotes the data of the i-th calendar element. Finally, we will group the normalized data using a sliding window to reduce the computational complexity. A window size of m is defined, segmenting the normalized data into t-m-1 input-output sequences for supervised training.
{ Δ x 1 , Δ x 2 , Δ x 3 , , Δ x m , Δ x m + 1 } { Δ x 2 , Δ x 3 , Δ x 4 , , Δ x m + 1 , Δ x m + 2 } { Δ x t m 1 , Δ x t m , Δ x t m + 1 , , Δ x t 2 , Δ x t 1 }
For each small sample dataset, we use the first m data as input data to the model, and the last point as the label of the predicted value, and input all the arrays into the model sequentially for SCB prediction. Figure 6 illustrates the complete data processing workflow, from the input of the raw SCB sequence through model training to final prediction.
In the training phase, we predict the m+1th data point using the previously known m data points. And in the prediction phase, the first prediction value will be predicted based on the previous m data points. Subsequently, in the second prediction step, the window is advanced by one time step: the earliest data point is removed, the latest observation is incorporated, and the prediction for the m+2nd point is generated based on the updated sequence of the most recent m points. Finally, a sliding window of m data lengths will be utilized sequentially to complete our prediction task.
Finally, upon completion of the aforementioned prediction tasks, we obtained the preliminary prediction outputs. To further restore their practical significance, the data must undergo post-processing procedures, primarily involving denormalization and linear transformation. Denormalization remaps the standardized data output by the model back to the original value range to recover its actual physical dimensions, while linear transformation ensures that the predicted results reside in the same space as the true values. Through these steps, the final directly applicable SCB prediction results are obtained. See Table 3 for detailed experimental parameters.
This paper presents a hybrid CNN-LSTM-Attention model implemented on the TensorFlow framework; its parameter configurations are detailed in the accompanying table. The proposed architecture processes time series data through sequential layers: A 1D CNN with 64 filters (kernel size = 5) for local feature extraction, followed by an LSTM layer (64 hidden units) to capture long-term dependencies. The model was trained for 200 epochs using mini-batches of 64, with Mean Squared Error (MSE) as the loss function and the Adam optimizer [30] for stable convergence. The network accepts an input dimension of 64 and produces a single output, representing the predicted satellite clock bias. This integrated design effectively synergizes convolutional, recurrent, and attention mechanisms to achieve high-precision forecasting.

4. Experimentation and Analysis

After preprocessing, the short-term forecasting experiments are performed using the prediction model developed in this study as well as other conventional models. The known precision SCB for the forecast period serves as the ground truth to evaluate each model’s prediction accuracy. Error, root mean square error (RMSE) and Mean error are used as the indexes of clock difference accuracy, and their calculation formulas are as follows:
E r r o r = y ^ S C B i y S C B i
R M S E = 1 m i = 1 m ( y ^ S C B i y S C B i ) 2
M e a n = 1 m i = 1 m ( y ^ S C B i y S C B i )
Im provement - Rate = M b a s e M n e w M b a s e × 100 %
where m is the number of data epochs, y ^ S C B i represents the predicted SCB at the i-th epoch, y S C B i denotes the measurements at the i-th epoch. For multiple satellites, the average RMSE served as the clock error accuracy metric. M denotes the evaluation metric. For error-based metrics, the formula is adjusted so that a positive value indicates performance enhancement.

4.1. Performance Assessment of Forecasting Models

The forecasting performance of the CNN-LSTM-Attention model was evaluated by benchmarking it against five established models: LP, QP, ARIMA, BP, and LSTM. In this study, we utilize the post-processed precise SCB data for DOY 11, 2024, available from the IGS Wuhan University Data Center (WHU). The first 18 h of SCB data from that day were used for model training and fitting. Subsequently, the trained model was applied to perform one-step-ahead predictions for four different forecast horizons (1 h, 2 h, 4 h, and 6 h). Specifically, each prediction task was accomplished through a single sliding-window prediction step, rather than employing an iterative rolling or recursive prediction process. We selected the known precise SCB from the forecast period as the ground truth. Against this reference, the overall mean prediction accuracy across all satellites was computed. As summarized in Table 4 and the accompanying figure, the prediction performance across models is presented through statistical data and accuracy metrics, respectively.
The determination of the hyperparameters for each model is described as follows: or the ARIMA model, the order parameters (p, d, q) were automatically determined through a grid search within a pre-defined range, with p ranging from 0 to 2, d from 0 to 1, and q from 0 to 2., and the optimal parameter combination was selected based on the minimized Bayesian Information Criterion (BIC), aiming to achieve a balance between model fit and simplicity. The LSTM model adopts a single-layer LSTM architecture with 500 hidden units, incorporating a Dropout layer (rate = 0.2) and a ReLU activation function to mitigate overfitting. The model was trained using the Adam optimizer, with a maximum of 600 epochs, an initial learning rate of 0.005, and a piecewise learning rate decay strategy to enhance training performance. The BP neural network features a three-layer feedforward structure, consisting of 15 input nodes, 5 hidden nodes, and 1 output node. The Particle Swarm Optimization algorithm was employed to initialize the network’s weights and thresholds, with a population size of 5, 30 iteration steps, and an adaptive mutation mechanism to improve global search capability. The network was trained for a maximum of 2000 epochs with a learning rate of 0.001, and data normalization and denormalization were applied to enhance training stability and prediction accuracy.
The overall superiority of the CNN-LSTM-Attention model is evidenced by the results in Table 4 and Figure 7 Compared with LP, QP, ARIMA, BP and LSTM models, the prediction accuracy shows a percentage improvement rate of 76.95%, 66.84%, 65.92%, 84.33%, and 43.87% in the 1 h prediction task; 72.59%, 65.61%, 74.60%, 82.98%, and 51.13% in the 2 h prediction task; 70.45%, 68.52%, 81.63%, 88.44%, and 60.49% in the 4 h prediction task; 70.26%, 70.51%, 84.28%, 93.66%, and 66.76% in the 6 h prediction task.
Notably, the model’s superiority over traditional forecasts remains significant with extending prediction horizons. In contrast, its improvement compared to the BP neural network and LSTM models continues to expand with longer forecast durations. These results indicate that the model excels in short-term forecasting, while also demonstrating more stable advantages over traditional LSTM methods in long-term prediction tasks. This suggests that the model possesses strong adaptability and generalizability across various time scales in satellite clock bias prediction.
Figure 8 illustrates the trend of forecast errors in the prediction task, with curves in different colors representing different BDS satellites. As shown in Figure 8, forecast errors accumulate across all six models as the prediction time lengthens. The CNN-LSTM-Attention model demonstrates superior error control capability throughout the entire prediction period. Its prediction error sequence is noticeably more convergent, with a significantly smaller fluctuation range compared to the LP, QP, ARIMA, BP, and LSTM models. Especially under longer prediction time horizons, the model still maintains a lower error level without exhibiting obvious divergence, reflecting its excellent stability and generalization performance. In contrast, the other models show gradually increasing errors after the medium to long term, and their error distributions remain more dispersed than that of the CNN-LSTM-Attention model. The results indicate that the CNN-LSTM-Attention model can more effectively capture complex patterns in clock bias sequences and exhibits consistent prediction performance across different time scales.
As shown in Figure 9 and Table 5, the CNN-LSTM-Attention model achieves the best accuracy for the majority of satellites. The overall RMSE values across all satellites exhibit minimal fluctuation, demonstrating its high stability and consistent performance. These results indicate that the model not only improves prediction accuracy but also exhibits strong generalization capability across different satellites. Compared to other benchmark models, the proposed architecture shows a superior ability to capture complex patterns in satellite clock bias data, thereby enhancing its reliability across all forecasting horizons. Given these advantages, the model is ideal for a range of high-precision applications, particularly real-time positioning and timing systems.

4.2. PPP Experiment

The predicted SCB products were employed for PPP calculation to evaluate the model’s performance. Specifically, kinematic PPP data processing was performed for the selected IGS tracking stations mentioned above. The positioning results were then compared with those obtained using the post-processed SCB products available from the WHU to verify the predictive performance. Here, data from post-processed products covering the experimental prediction interval were selected for simultaneous kinematic PPP processing. In the experiment, we utilized the RTPPP software developed by GFZ to conduct real-time PPP tests. The processing mode was set to “kinematic for positioning verification”. Antenna phase center corrections were applied using the “IGS 20.atx” model, with an elevation cut-off angle of 7°. For the antenna phase center corrections in this study, we utilized the product data disseminated by the IGS, available at: https://files.igs.org/pub/station/general (accessed on 15 September 2025). Tropospheric delay was modeled using ZHD with the Saastamoinen model, while ionospheric delay was addressed by eliminating first-order effects via dual-frequency GNSS signals. For satellite orbits, WUH precision satellite orbits were adopted, and clock corrections were based on 6 h predicted SCB data from the aforementioned forecasting experiment. The calculated station coordinates in simulated real-time positioning mode were compared with the known static coordinates of the IGS stations. In this context, dynamic PPP continuously processes GNSS observations in real time, updating receiver coordinates epoch by epoch. This method simulates a dynamic or kinematic scenario in practical positioning applications, where the receiver may be in motion and its position changes over time. By employing predicted satellite clock and orbit products in this kinematic PPP framework, the experiment effectively evaluates the performance of the forecasting model under real-world, time-varying conditions, thereby providing a realistic assessment of the predictive SCB products’ capability in supporting high-precision real-time navigation and positioning.
Subsequently, we compared the average positioning errors between the predicted and product data in the N, E, and U directions. Table 6 summarizes the positioning results. The statistical analysis excluded PPP results from the first two hours. The kinematic PPP with BDS satellites achieves positioning accuracy comparable to that of post-processed SCB products in all directions. This result confirms both the effectiveness and practicality of the prediction model. The horizontal positioning accuracy (N and E) is notably higher, with most station errors within 10 cm, while the vertical (U) direction shows slightly larger discrepancies.
Comparison between the positioning results derived from the predicted data and the post-processed SCB products reveals a strong consistency across the vast majority of stations, with both sets of results exhibiting positioning accuracy of the same order of magnitude. Although the post-processed precise products demonstrate superior accuracy at some stations (USN7, YEL2), the predicted data conversely achieve better performance at others (AGGO). The predicted SCB approach the accuracy of the post-processed products, demonstrating their capability to support real-time positioning requirements. Furthermore, as can be observed from the error sequences of stations AGGO, KARR, PERT, and YEL2 in Figure 10, the PPP results utilizing the predicted clock offsets are stable and reliable. The experimental results confirm that the employed clock offset prediction model can provide a reliable service for BDS real-time precise point positioning, holding significant practical value.
The positioning accuracy was subsequently assessed using the RMSE of the kinematic PPP solutions derived from the predicted data. As shown in Table 7, the results achieve decimeter-level accuracy across all stations, albeit with expectedly lower precision in the vertical component. Among them, some stations (such as KOUG and YKRO) exhibit relatively larger errors, while the majority of stations maintain high accuracy, confirming the effectiveness of the method for real-time positioning applications. The significant errors observed at stations KOUG and YKRO are primarily attributed to their coastal geographic locations near the equator. Such environments are highly susceptible to strong multipath effects, such as those caused by sea surface reflections. PPP solutions are particularly sensitive to this type of multipath interference, especially in coastal areas at low latitudes, where signal reflection and delay issues can further amplify positioning errors.

5. Conclusions

A CNN-LSTM-Attention model is proposed for SCB prediction. This approach enhances the forecasting accuracy of BDS satellite clock differences by leveraging all informative content in the data. Our experimental study reveals that:
(1)
This paper describes the development of a hybrid model based on CNN, LSTM, and Attention mechanisms for predicting BDS SCB. In this architecture, local features extracted by the CNN layer are passed to the LSTM to model temporal dependencies. The Attention mechanism then enhances this process by adaptively weighting critical timesteps, collectively improving prediction accuracy and stability. Meanwhile, we sequentially complete the prediction tasks by leveraging the correlations among BDS satellites.
(2)
Regarding both accuracy and stability, the CNN-LSTM-Attention model significantly outperforms all benchmark models (LP, QP, ARIMA, BP, LSTM). Furthermore, the model excels in short-term predictions and demonstrates remarkable stability in the long term, as the RMSE for most satellites stays below 1 ns.
(3)
Finally, we conducted simulated real-time PPP experiments using predicted SCB data and post-processed SCB products over the same time period, performing kinematic PPP data processing for 30 selected globally distributed IGS tracking stations. Experimental results demonstrate comparable positioning accuracy between the predicted data and post-processed products in the N, E, and U directions, confirming its suitability for positioning applications.
Looking forward, to enhance the applicability of the model in continuous operational scenarios including multi-day or cross-day forecasting, we plan to incorporate a discontinuity correction mechanism. Daily boundary discontinuities, a recognized feature of continuous SCB products, will be addressed by applying a constant offset to align the clock bias series at day transitions. This approach is expected to eliminate prediction jumps and improve the consistency and robustness of long-term SCB forecasts in sustained positioning applications.

Author Contributions

J.T. and J.M. conceived and designed the experiments; J.T. and H.T. performed the experiments; J.M. wrote the paper; J.M. and H.T. analyzed the data; J.T. and X.W. revised the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Yunnan Fundamental Research Projects (202401AS070067), China; and the Yunnan Provincial Young and Middle-aged Academic and Technical Leaders Reserve Talents Project (202405AC350017), China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Both training data and accuracy testing data used in this study are from the post-processing high-precision SCB products provided by IGS WHU. Among them, the experimental data research is publicly available. Available at: https://cddis.nasa.gov/archive/gnss/products (accessed on 25 June 2025).

Acknowledgments

The SCB products were provided by the WHU, and the authors extend their gratitude.

Conflicts of Interest

The authors declare that they have no competing interests.

References

  1. Geng, J.; Wen, Q.; Chen, G.; Dumitraschkewitz, P.; Zhang, Q. All-frequency IGS phase clock/bias product combination to improve PPP ambiguity resolution. J. Geod. 2024, 98, 48. [Google Scholar] [CrossRef]
  2. Han, J.; Zhang, J.; Zhong, S.; Peng, B.; Lu, R.; Zhou, C. The effect of satellite code bias on GPS PPP time transfer. J. Spat. Sci. 2025, 70, 29–44. [Google Scholar] [CrossRef]
  3. Caissy, M.; Agrotis, L.; Weber, G. The IGS real-time service//EGU General Assembly Conference Abstracts. Geophys. Res. Abstr. 2013, 15, EGU2013-11168. [Google Scholar]
  4. Johnston, G.; Riddell, A.; Hausler, G. The international GNSS service. In Springer Handbook of Global Navigation Satellite Systems; Springer: Cham, Switzerland, 2017; pp. 967–982. [Google Scholar] [CrossRef]
  5. Wang, L.; Li, Z.; Ge, M.; Neitzel, F.; Wang, X.; Yuan, H. Investigation of the performance of real-time BDS-only precise point positioning using the IGS real-time service. GPS Solut. 2019, 23, 66. [Google Scholar] [CrossRef]
  6. Huang, G.W.; Zhang, Q.; Xu, G.C. Real-time clock offset prediction with an improved model. GPS Solut. 2014, 18, 95–104. [Google Scholar] [CrossRef]
  7. Nie, Z.; Gao, Y.; Wang, Z.; Ji, S.; Yang, H. An approach to GPS clock prediction for real-time PPP during outages of RTS stream. GPS Solut. 2018, 22, 14. [Google Scholar] [CrossRef]
  8. Duan, B.; Hugentobler, U.; Montenbruck, O. A method to assess the quality of GNSS satellite phase bias products. GPS Solut. 2024, 28, 89. [Google Scholar] [CrossRef]
  9. Li, X.; Zheng, H.; Li, X.; Yuan, Y.; Wu, J.; Han, X. Open-source software for multi-GNSS inter-frequency clock bias estimation. GPS Solut. 2023, 27, 84. [Google Scholar] [CrossRef]
  10. Syam, W.P.; Priyadarshi, S.; Roqué, A.A.G. Fast and reliable forecasting for satellite clock bias correction with transformer deep learning. In Proceedings of the 54th Annual Precise Time and Time Interval Systems and Applications Meeting, Long Beach, CA, USA, 23–26 January 2023; pp. 76–96. [Google Scholar]
  11. Huang, G.; Cui, B.; Zhang, Q.; Fu, W.; Li, P. An improved predicted model for BDS ultra-rapid satellite clock offsets. Remote Sens. 2018, 10, 60. [Google Scholar] [CrossRef]
  12. Liang, Y.J.; Ren, C.; Yang, X.F.; Pang, G.F.; Lan, L. A grey model based on first differences in the application of satellite clock bias prediction. Chin. Astron. Astrophys. 2016, 40, 79–93. [Google Scholar] [CrossRef]
  13. Zheng, Z.-Y.; Chen, Y.Q.; Liu, X.S. An improved grey model and its application research on the prediction of real-time GPS satellite clock errors. Chin. Astron. Astrophys. 2009, 33, 72–89. [Google Scholar] [CrossRef]
  14. Zheng, Z.Y.; Dang, Y.M.; Lu, X.S.; Xu, W.M. Prediction model with periodic item and its application to the prediction of GPS satellite clock bias. Acta Astron. Sin. 2010, 51, 95–102. [Google Scholar] [CrossRef]
  15. Davis, J.; Bhattarai, S.; Ziebart, M. Development of a Kalman filter based GPS satellite clock time-offset prediction algorithm. In Proceedings of the 2012 European Frequency and Time Forum, Gothenburg, Sweden, 23–27 April 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 152–156. [Google Scholar] [CrossRef]
  16. Xi, C.; Cai, C.-N.; Li, S.M.; Li, X.H.; Li, Z.B.; Deng, K.Q. Long-term clock bias prediction based on an ARMA model. Chin. Astron. Astrophys. 2014, 38, 342–354. [Google Scholar] [CrossRef]
  17. Zewdie, G.K.; Valladares, C.; Cohen, M.B.; Lary, D.J.; Ramani, D.; Tsidu, G.M. Data-driven forecasting of low-latitude ionospheric total electron content using the random forest and LSTM machine learning methods. Space Weather 2021, 19, e2020SW002639. [Google Scholar] [CrossRef]
  18. Wang, G.; Liu, L.; Xu, A.; Su, X.Q.; Liang, X.H. The application of radial basis function neural network in the GPS satellite clock bias prediction. Acta Geod. Cartogr. Sin. 2014, 43, 803–807. [Google Scholar] [CrossRef]
  19. Wang, X.; Chai, H.; Wang, C.; Xiao, G.; Chong, Y.; Guan, X. Improved wavelet neural network based on change rate to predict satellite clock bias. Surv. Rev. 2021, 53, 325–334. [Google Scholar] [CrossRef]
  20. He, S.; Liu, J.; Zhu, X.; Dai, Z.; Li, D. Research on modeling and predicting of BDS-3 satellite clock bias using the LSTM neural network model. GPS Solut. 2023, 27, 108. [Google Scholar] [CrossRef]
  21. Huang, B.; Ji, Z.; Zhai, R.; Xiao, C.; Yang, F.; Yang, B.; Wang, Y. Clock bias prediction algorithm for navigation satellites based on a supervised learning long short-term memory neural network. GPS Solut. 2021, 25, 80. [Google Scholar] [CrossRef]
  22. Wen, T.; Ou, G.; Tang, X.; Zhang, P.; Wang, P. A Novel Long Short-Term Memory Predicted Algorithm for BDS Short-Term Satellite Clock Offsets. Int. J. Aerosp. Eng. 2021, 2021, 4066275. [Google Scholar] [CrossRef]
  23. Cai, C.; Liu, M.; Li, P.; Li, Z.; Lv, K. Enhancing satellite clock bias prediction in BDS with LSTM-attention model. GPS Solut. 2024, 28, 92. [Google Scholar] [CrossRef]
  24. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Chen, T. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
  25. Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef]
  26. Yu, Y.; Si, X.; Hu, C.; Zhang, J. A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar] [CrossRef] [PubMed]
  27. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. Comput. Sci. 2016. [Google Scholar] [CrossRef]
  28. Li, N.; Zhao, L.; Li, H. BDS multiple satellite clock offset parallel prediction based on multivariate CNN-LSTM model. GPS Solut. 2024, 28, 189. [Google Scholar] [CrossRef]
  29. Law, J. Robust Statistics—The Approach Based on Influence Functions; John Wiley & Sons: New York, NY, USA, 1986. [Google Scholar] [CrossRef]
  30. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
Figure 1. Location information of 30 IGS observation stations from the WUH Data Center.
Figure 1. Location information of 30 IGS observation stations from the WUH Data Center.
Sensors 26 00422 g001
Figure 2. LSTM neural network structure diagram.
Figure 2. LSTM neural network structure diagram.
Sensors 26 00422 g002
Figure 3. Attention Mechanism Flowchart.
Figure 3. Attention Mechanism Flowchart.
Sensors 26 00422 g003
Figure 4. Architecture of the CNN-LSTM-Attention Model.
Figure 4. Architecture of the CNN-LSTM-Attention Model.
Sensors 26 00422 g004
Figure 5. Predictive Modeling Methodology.
Figure 5. Predictive Modeling Methodology.
Sensors 26 00422 g005
Figure 6. Data Preprocessing Flowchart.
Figure 6. Data Preprocessing Flowchart.
Sensors 26 00422 g006
Figure 7. The RMSE for tasks with different forecast time periods.
Figure 7. The RMSE for tasks with different forecast time periods.
Sensors 26 00422 g007
Figure 8. Prediction residuals from different models: LP (a), QP (b), ARIMA (c), BP (d), LSTM (e), and CNN-LSTM-Attention (f) (Each curve represents the predicted SCB error of a different satellite, with distinct colors used to differentiate between satellites).
Figure 8. Prediction residuals from different models: LP (a), QP (b), ARIMA (c), BP (d), LSTM (e), and CNN-LSTM-Attention (f) (Each curve represents the predicted SCB error of a different satellite, with distinct colors used to differentiate between satellites).
Sensors 26 00422 g008
Figure 9. The average prediction accuracy of each satellite in different models obtained based on different prediction tasks (shown from top to bottom).
Figure 9. The average prediction accuracy of each satellite in different models obtained based on different prediction tasks (shown from top to bottom).
Sensors 26 00422 g009
Figure 10. Comparative diagram of positioning errors.
Figure 10. Comparative diagram of positioning errors.
Sensors 26 00422 g010
Table 1. Information of satellites and atomic clocks in different BDS systems.
Table 1. Information of satellites and atomic clocks in different BDS systems.
SystemModelTrackClockPRN
BDSBDS-2GEORbC01 C02 C03 C04 C05
MEOC11 C12 C14
IGSOC06 C07 C08 C09 C10 C13
BDS-3MEORbC19 C20 C21 C22 C23 C24 C32 C33 C36 C37 C41 C42
PHMC25 C26 C27 C28 C29 C30 C34 C35 C43 C44 C45 C46
IGSOPHMC38 C39 C40
Table 2. Geographic coordinates of IGS tracking stations.
Table 2. Geographic coordinates of IGS tracking stations.
StationsLatitude ( ° )Longitude ( ° )StationsLatitude ( ° )Longitude ( ° )
AGGO−34.874−58.140RGDG−53.786−67.752
AREG−16.465−71.493SFDM34.460−118.755
AREQ−16.466−71.493SOD367.42126.389
CIBG−6.490106.849STHL−15.943−5.667
CKSV22.999120.220STR1−35.316149.010
DAV1−68.57777.973TEJA−39.805−73.253
GAMG35.590127.920ULAB47.865107.052
GRAC43.7546.921USN738.921−77.066
KARR−20.981117.097USUD36.133138.362
KMNM24.464118.389WSRT52.9156.604
KOUG5.098−52.640WTZZ49.14412.879
MIZU39.135141.133WUH230.532114.357
ORID41.12720.794WUTH77.00315.539
P05348.726−107.725YEL262.481−114.481
PERT−31.802115.885YKRO6.871−5.240
Table 3. Parameterization of CNN-LSTM-Attention models.
Table 3. Parameterization of CNN-LSTM-Attention models.
No.ParametersValue
1Loss functionMSE
2OptimizerAdam (learning rate = 0.001)
3Convolutional layer filters64
4Convolutional layer kernel size5
5Convolutional layer activationReLU
6LSTM Hidden Layer Size64
7Attention layer activationSigmoid
8Batch size64
9Training epoch200
10Input dimension size64
11Output dimension size1
12Train/Test split ratio75%/25%
13Data normalizationMin-Max scaling [0, 1]
Table 4. Average prediction accuracy RMSE for different models.
Table 4. Average prediction accuracy RMSE for different models.
Model1 h2 h4 h6 h
LP0.4650.5370.6800.841
QP0.3230.4280.6380.849
ARIMA0.3140.5801.0951.592
BP0.6840.8651.7383.945
LSTM0.1890.3010.5090.753
CNN-LSTM-Attention0.1070.1470.2010.250
Table 5. Comparison of statistical results for the six models across different orbital types.
Table 5. Comparison of statistical results for the six models across different orbital types.
Prediction
Task
ModelBDS-2BDS-3
TrackGEOMEOIGSOMEOIGSO
ClockRbRbPHMPHM
1 hLP1.3570.3811.0640.1900.0960.440
QP0.5770.2130.7230.1680.1590.484
ARIMA0.6600.1940.3850.2210.2340.413
BP2.0260.8191.5210.2680.2430.064
LSTM0.4350.3110.3980.0870.0590.171
CNN-LSTM-Attention0.2210.1060.2380.0630.0550.040
2 hLP1.4650.4041.2940.2310.1150.527
QP0.9010.2060.9990.1940.1970.587
ARIMA1.3220.2720.6410.4250.4200.787
BP2.2251.1451.8180.4420.3710.086
LSTM0.8070.3410.5850.1340.1040.313
CNN-LSTM-Attention0.2750.1400.3460.0890.0810.043
4 hLP1.6600.5821.6800.3000.1890.631
QP1.5260.2931.4850.2480.2920.762
ARIMA2.7960.3981.0520.8390.7371.493
BP3.4862.7423.0991.1980.9640.357
LSTM1.2910.3901.0150.2460.2130.546
CNN-LSTM-Attention0.3010.1890.4370.1280.1530.056
6 hLP1.9220.6952.0950.3990.2610.771
QP2.0200.3492.0850.3120.3810.939
ARIMA4.0480.5281.5131.2611.0492.218
BP8.5426.5646.0612.7532.2560.965
LSTM1.6490.5191.6610.3930.3370.777
CNN-LSTM-Attention0.2860.2740.4710.1930.2230.060
Table 6. Kinematic PPP Accuracy Statistical Results.
Table 6. Kinematic PPP Accuracy Statistical Results.
StationsPrediction Data Mean/cmPost-Processed Products Mean/cm
NEUNEU
AGGO0.87.31.3−3.35.20.0
AREG5.28.6−1.87.36.61.4
AREQ−0.47.9−3.31.13.80.6
CIBG7.4−1.4−4.93.9−0.2−1.3
CKSV−13.9−3.82.7−1.0−2.52.8
DAV113.34.2−3.2−4.40.3−0.5
GAMG−10.97.313.0−0.9−2.50.7
GRAC−0.29.75.4−0.52.6−2.4
KARR2.80.6−3.95.5−1.1−2.4
KMNM−16.9−3.78.4−4.2−1.75.3
KOUG45.810.89.941.05.58.0
MIZU−137.4−15.6−2.0−2.0−5.5
ORID7.910.1−2.51.8−4.25.6
P053−6.7−1.87.8−0.20.5−0.4
PERT10.01.3−4.06.70.80.1
RGDG2.88.51.3−3.21.7−0.9
SFDM−11.45.48.71.70.42.9
SOD310.1−7.4−9.92.3−0.80.1
STHL1.48.34.5−2.51.20.9
STR16.7−1.7−8.40.40.5−1.1
TEJA15.516.65.47.99.74.2
ULAB8.4−7.91.36.2−0.5−2.1
USN7−14.54.56.40.5−1.8−0.6
USUD−16.110.63.7−2.4−0.9−1.0
WSRT16.0−3.610.01.8−2.85.9
WTZZ23.9−20.80.82.4−5.33.6
WUH2−12.1−3.93.9−1.9−3.40.7
WUTH11.28.40.90.1−0.1−0.5
YEL25.92.05.4−0.10.9−0.3
YKRO−29.417.2−25.0−5.44.3−5.7
Table 7. Positioning Accuracy of the Prediction Data.
Table 7. Positioning Accuracy of the Prediction Data.
StationsN (cm)E (cm)U (cm)StationsN (cm)E (cm)U (cm)
AGGO2.22.25.4RGDG3.412.25.8
AREG2.311.421.4SFDM4.210.823.5
AREQ4.310.618.5SOD32.011.34.5
CIBG4.82.315.3STHL9.59.16.8
CKSV7.913.48.0STR110.63.117.0
DAV115.726.532.2TEJA5.314.111.7
GAMG9.813.735.8ULAB2.110.914.1
GRAC5.27.011.7USN73.310.323.0
KARR4.32.54.0USUD7.817.820.2
KMNM8.415.912.7WSRT4.615.317.2
KOUG31.332.8106.7WTZZ18.123.04.0
MIZU5.715.423.1WUH27.89.813.0
ORID10.78.110.6WUTH14.013.611.2
P0536.55.410.2YEL26.83.510.2
PERT3.52.617.2YKRO23.726.234.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, J.; Tang, J.; Teng, H.; Wu, X. Prediction and Performance of BDS Satellite Clock Bias Based on CNN-LSTM-Attention Model. Sensors 2026, 26, 422. https://doi.org/10.3390/s26020422

AMA Style

Ma J, Tang J, Teng H, Wu X. Prediction and Performance of BDS Satellite Clock Bias Based on CNN-LSTM-Attention Model. Sensors. 2026; 26(2):422. https://doi.org/10.3390/s26020422

Chicago/Turabian Style

Ma, Junwei, Jun Tang, Hanyang Teng, and Xuequn Wu. 2026. "Prediction and Performance of BDS Satellite Clock Bias Based on CNN-LSTM-Attention Model" Sensors 26, no. 2: 422. https://doi.org/10.3390/s26020422

APA Style

Ma, J., Tang, J., Teng, H., & Wu, X. (2026). Prediction and Performance of BDS Satellite Clock Bias Based on CNN-LSTM-Attention Model. Sensors, 26(2), 422. https://doi.org/10.3390/s26020422

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop