Preprint
Essay

Clock Bias Prediction of Navigation Satellite Based on BWO-CNN-BiGRU-Attention Model

Altmetrics

Downloads

81

Views

47

Comments

0

Submitted:

21 August 2024

Posted:

22 August 2024

You are already at the latest version

Alerts
Abstract
Satellite clock bias (SCB) prediction is a crucial technology for satellite navigation systems, holding significant importance for Global Navigation Satellite System. This paper proposes a deep learning model for SCB prediction based on the fusion of the Beluga Whale Optimization (BWO), Convolutional Neural Network (CNN), Bidirectional Gated Recurrent Unit (BiGRU), and an attention mechanism. The CNN is utilized to extract the spatiotemporal characteristic information from the clock bias sequence, while the BiGRU fully extracts relevant features through forward and backward propagation. The introduction of an attention mechanism aims to preserve essential features within the clock bias sequence to enhance feature extraction by both CNN and BiGRU networks. Additionally, the BWO is employed to optimize parameter selection in order to improve model accuracy. Experimental verification demonstrates that for the BeiDou Navigation Satellite System's (BDS) hydrogen-maser atomic clocks, the predicted clock bias for 6 hours, 3 days, and 15 days are 0.078 ns, 0.475 ns, and 2.130 ns respectively - representing improvements of 31%, 45%, and 66% over CNN-BiGRU-Attention; 6%, 51%, and 56% over CNN-BiGRU; and 32%, 35%, and 73% over BiGRU, respectively.
Keywords: 
Subject: Engineering  -   Control and Systems Engineering

1. Introduction

In contemporary society, Global Navigation Satellite Systems (such as GPS, GLONASS, Galileo, and BeiDou) have become indispensable tools in people's daily lives. These systems play a crucial role in various fields including navigation, positioning, the internet of vehicles, earthquake early warning, and resource management. The precision of satellite navigation significantly influences the development level of these related fields. Satellite clock bias (SCB) is a highly significant parameter within satellite navigation systems, directly impacting the accuracy of navigation positioning. Therefore, the precise prediction of SCB holds great significance for enhancing the precision and reliability of satellite navigation systems.
Various methods can be employed for SCB prediction, which can generally be categorized as follows: the physical modeling (e.g., Quadratic Polynomial, QP (Wang et al. 2020)), the statistical modeling (e.g., Grey Model - GM (Cui & Jiao, 2005; Liang et al., 2016), Autoregressive Integrated Moving Average Model – ARIMA (Xu & Zeng, 2009; Li, 2013; Xi et al., 2014; Zhang et al., 2022), Kalman Filter Model (Zhu et al., 2008; Davis et al., 2012), Exponential Smoothing (Wang et al., 2017; Yu et al., 2020), etc.), the machine learning modeling (e.g., Support Vector Machine - SVM (He et al., 2019), Back Propagation (BP) Neural Network (Lv et al., 2022), Elman Recurrent Neural Network (Liang et al. 2022), Wavelet Neural Network - WNN (Ai et al., 2016; Wang et al., 2016), Radial Basis Function - RBF (Li, 2013; Wang et al., 2014), Long Short-term Memory Network - LSTM (Huang et al., 2021; He et al., 2023), etc.), the combined prediction models that integrate multiple single models (e.g., GM and WNN (Wang et al., 2014), wavelet transform and SVM (Lei et al., 2014), etc.), as well as some improved models within their own categories (Lu et al., 2008; Zheng et al., 2008; Wang et al., 2017; Huang et al., 2018; Wang et al., 2021; Bai et al., 2023; Cai et al., 2024). Each prediction method has specific characteristics suitable for certain applications but may also exhibit limitations due to the complex time-frequency characteristics of spaceborne atomic clocks and susceptibility to external environmental factors. For instance, as a quadratic polynomial model forecasts clock bias, its predictive errors escalate notably over longer forecasting horizons rendering it unsuitable for long-term predictions. The predictive precision of GM is heavily influenced by its exponential coefficient. ARIMA necessitate stable clock bias data and are thus ill-suited for long-term forecasting. Kalman filters' efficacy hinges upon understanding atomic clock operational traits and stochastic prior knowledge. Exponential Smoothing methods are typically suited only for short- to medium-term projections, and may underperform with data exhibiting long-term trends or complex nonlinear relationships. The selection of kernel function parameters in the SVM prediction model exerts significant impact on predictive performance, yet determining the optimal parameters poses challenges. WNN models' network topology is cumbersome to determine. RBF models prediction suffer from a lack of theoretical foundation regarding determining sample length, sample size, and intervals between samples. Optimizing combinatorial weights in combinatorial forecasting models poses difficulties. The current state of clock bias prediction is characterized by the diversification of SCB models, but they do not have strong universal adaptability. Therefore, exploring a combination model with strong adaptability and high prediction accuracy is a hot direction for future clock bias forecasting models.
The research on satellite clock prediction increasingly focuses on deep learning methods based on neural networks (Lu et al., 2023). This paper introduces a novel SCB prediction model based on BWO-CNN-BiGRU-Attention for the first time. CNN-BiGRU-Attention is a deep learning model comprising CNN and BiGRU with an attention mechanism. The model utilizes a CNN to capture local correlations between different time steps and share local features across various time points. The BiGRU can effectively model the time series data in both directions, thereby enabling the model to capture more comprehensive information. Additionally, the attention mechanism assigns a corresponding weight to each time step, facilitating automatic selection of the most crucial time points and features, thus enhancing the performance and effectiveness of the model. The performance of deep learning models heavily relies on hyperparameter selection and adjustment. To optimize model parameter selection, this paper employs the optimization algorithm BWO for parameter optimization, further enhancing accuracy and robustness.
To validate the effectiveness and applicability of this new predictive model, 300s precision clock bias data from September 6th to November 9th in 2023 provided by IGMAS are used for experimentation. The experimental results demonstrate that for BDS satellite-borne hydrogen atomic clock, the predicted bias at intervals of 6 hours, 3 days and 15 days are 0.078 ns ,0 .475 ns, and 2 .130 ns, respectively. These values were found to be higher improvement than CNN-BiGRU-Attention by 31%, 45%, and 66%; higher than CNN-BiGRU by 6%,51%, and 56%; and higher than BiGRU by32%, 35% and 73% for the respective time intervals.
The structure of this paper is as follows: Section 2 briefly reviews the basic knowledge of CNN, BiGRU, Attention mechanism and BWO algorithm. Section 3 constructs a SCB prediction model based on BWO-CNN-BiGRU-Attention, and Section 4 provides an case analysis.Section 5 summarizes the full text.

2. Basic Knowledge

2.1. Convolutional Neural Network

Convolutional Neural Network (CNN), proposed by Yann Lecun in 1989, are essentially a multi-layer perceptron and one of the representative algorithms of deep learning (LeCun et al., 2015). CNN model is usually composed of input layer, convolution layer, pooling layer, fully connected layer and output layer, the specific structure is shown in Figure 1:
Input layer: The input layer serves as the entry point for the entire network and typically requires preprocessing of the raw data (the data preprocessing will be covered in Section 3 of this article).
Convolution layer: This layer comprises the convolution computation and activation layers. The convolution computation involves matrix operations between the convolution kernel matrix and corresponding position data matrices, with inputs denoted as x R M × N , the convolution kernel as W R U × V , and defined by
, where * represents the convolution operator. The formula is as follows:
z i j = u = 1 U v = 1 V w u v x i u + 1 , j v + 1
The activation layer functions to nonlinearly transform the output from the convolution computation through an activation function:
Y = f ( z )
Here, f ( ) denotes a nonlinear activation function, with commonly used options including Sigmoid, tanh, ReLU and their variations.
Pooling layer: The primary purpose of this layer is to compress data, reduce neural network parameters and computations while preventing overfitting. A common pooling method is Max Pooling which utilizes maximum values within pooled areas to highlight key features in data. Other pooling methods include average pooling, overlapping pooling, spatial pyramid pooling , and random pooling (DataCamp, 2024), etc.
Fully connected layer: Neurons in this fully connected layer establish connections with all neurons in preceding layers; its role is to integrate distributed feature representations learned by the network and map them into sample label space.

2.2. Bidirectional Gated Recurrent Unit

Gated Recurrent Unit (GRU) networks represent an optimization of LSTM networks characterized by simpler model structures that incur lower computational costs while exhibiting improved convergence rates (Cho et al., 2014). The GRU structure is depicted in Figure 2.
The gated update formula for GRU is as follows:
z t = σ ( W z [ h t 1 , x t ] + b z )
r t = σ ( W r [ h t 1 , x t ] + b r )
h ˜ t = tanh ( W h [ r t h t 1 , x t ] + b h )
h t = ( 1 z t ) h t 1 + z t h ˜ t
Where, x t represents the input vector at time t, z t and r t denote the update gate and reset gate. h t represents the hidden layer state, while h ˜ t signifies the candidate hidden layer state. W z , W r , and W h represent the weight matrix, and b z , b r and b h are bias vector.
However, GRU networks are limited to processing data in a single direction, relying solely on past data for making predictions. In contrast, the Bidirectional Gated Recurrent Unit (BiGRU) network consists of both forward and backward neural networks, enabling it to leverage information from both preceding and subsequent time steps. This allows for more accurate predictions closer to the actual values.
The structure of the BiGRU model is illustrated in Figure 3 and comprises an input layer, forward hidden layer, backward hidden layer, and output layer. The input layer simultaneously feeds input data into both the forward and backward hidden layers at each time step, facilitating bidirectional flow of data through the GRU network.
At time t, BiGRU's hidden output is determined by two independent GRUs:
h t = GRU ( h t 1 , x t )
h t = GRU ( h t 1 , x t )
h t = W t h t + V t h t + b t
Whereas h t and h t represent the output state of the forward GRU and backward GRU at time t, respectively, W t and V t represent the weight matrix, and b t represents the bias vector.

2.3. Attention Mechanism

The attention mechanism (Vaswani et al., 2017) simulates human brain resource allocation by focusing attention on specific areas while reducing or ignoring others to obtain relevant information effectively without interference from irrelevant data. The core idea involves assigning different weights based on input data importance for comprehensive focus on crucial parts leading to improved model performance and prediction ability. The Attention structure shown in Figure 4:
Let X = [ x 1 , x 2 , x n ] be the input information vector, given a task-related query vector q, and calculate the attention distribution (weight coefficient) α i of the input information:
α i = soft max ( s ( x i , q ) ) = exp ( s ( x i , q ) ) i = 1 n exp ( s ( x i , q ) )
Where s ( x i , q ) is the attention score function and softmax is the normalized exponential function.Through softmax processing, the original calculated score is converted into a probability distribution with the sum of ownership weight as 1, which highlights the weight of important elements. The Attention score function s employs various models such as addition model, dot product model, scaled dot product model and bilinear model, etc. In this paper, the addition model is selected:
s ( x i , q ) = v T tanh ( W x i + U q )
Where, W , U and v are learnable parameters within neural network. Finally, the Attention layer output y comprises weighted summation of input information using weight coefficients from attention distribution:
y = i = 1 n α i x i

2.4. Beluga Whale Optimization

The Beluga Whale Optimization (BWO) algorithm was introduced by Changting Zhong et al in 2022 (Zhong et al., 2022) as a population-based meta-heuristic algorithm designed for numerical optimization and engineering. The algorithm emulates the swimming (Global Exploration), foraging (Local Development), and "whale fall" behaviors of beluga whales, demonstrating strong capabilities in both global and local search. The algorithm flowchart of BWO is shown in Figure5:
The transition from exploration to development is determined based on balance factors:
B f = B 0 ( 1 t 2 T )
Where t represents the current iteration, T denotes the maximum number of iterations, and B0 varies randomly between (0,1) at each iteration. The equilibrium factor governs the exploration phase  while the development phase occurs within B f 0.5 .
(1) Exploration Phase (Global Exploration)
This stage simulates the swimming behavior of beluga whales through different position updates based on odd-even positions. Its mathematical model is as follows:
{ X i , j t + 1 = X i , p j t + ( X r , p 1 t X i , p j t ) ( 1 + r 1 ) sin ( 2 π r 2 ) , j = e v e n X i , j t + 1 = X i , p j t + ( X r , p 1 t X i , p j t ) ( 1 + r 1 ) cos ( 2 π r 2 ) , j = o d d
Where t signifies the current number of iterations, X i , j t + 1 represents the new position of Beluga whale i on dimension j, P j j = 1 , 2 , , d is randomly selected from dimension d, X i , j t denotes the position of Beluga whale i on dimension j, and X i , P j t and X r , P 1 t are respectively current positions of Beluga whales i and r (where r stands for a randomly selected Beluga whale). r1 and r2 are random numbers in the range (0,1) that are used to enhance the stochastic operators in the exploration phase. sin ( 2 π r 2 ) and cos ( 2 π r 2 ) are used to balance the random numbers between the fins.
(2) Development Phase (Local Development)
This stage simulates predatory behavior where belugas move cooperatively to forage based on their surrounding locations. They share location information with each other during hunting to introduce optimal locations along with others. The mathematical model is as follows:
X i t + 1 = r 3 X b e s t t r r X i t + C 1 L F ( X r t X i t )
Where t indicates the current number of iterations; X i t and X r t represent current positions of Beluga whale i and a random Beluga whale; X i t + 1 signifies new position for Beluga whale i; X b e s t t denotes best position among whales; r3 and r4 are random numbers within range (0,1); and C 1 = 2 r 4 ( 1 t / T ) is used to measure the random jump intensity of Levy's flight intensity. L F is Levy flight function, which is calculated as follows:
L F = 0.05 × u × σ | v | 1 / β
σ = ( Γ ( 1 + β ) × sin ( π β / 2 ) Γ ( ( 1 + β ) / 2 ) × β × 2 ( β 1 ) / 2 ) 1 / β
Where u and v are normally distributed random numbers, and β is default constant equaling 1.5.
(3) Whale Fall Phase
During migration or foraging activities, Belugas face threats from killer whales, polar bears, and humans. Most Belugas can escape these threats by sharing information, but some may not survive, falling to deep ocean floors - a phenomenon known as "whale fall." To maintain population size, the updated location is established using Belugas' location combined with length of falling steps. The mathematical model is expressed as:
X i t + 1 = r 5 X i t r 6 X r t + r 7 X s t e p
Where r 5 , r 6 and r 7 represent random numbers between (0,1), and X s t e p signifies step length of the whale fall determined as:
X s t e p = ( u b l b ) exp ( C 2 t T )
Where C 2 is step factor related to probability of whale descent and population size ( C 2 = 2 W f n ), u b and l b represent upper and lower bounds of variables, respectively. It can be observed that step size is affected by the involved variable, number of iterations, and boundary for maximum number of iterations. In this model, the probability W f of whale falling is calculated as a linear function:
W f = 0.1 0.05 t / T

3. Development of SCB Prediction Model Based on BWO-CNN-BiGRU-Attention

Figure 6 illustrates the construction of a satellite clock bias prediction model based on BWO-CNN-BiGRU-Attention:
(1) Model Input
(a) Data Preprocessing
Addressing outliers in the atomic clock data is crucial. The original phase data may suffer from issues such as missing, jumping, and gross errors, which could significantly impact the accuracy of the clock bias prediction model and its predictive performance. Therefore, preprocessing of the original clock data is essential. Additionally, due to their typically large values on the order of E-4 to E-5, identifying outliers directly from these phase data can be challenging. In this study, we transform the phase data into frequency data and then apply outlier processing techniques (specifically using MAD method) to handle outliers in the frequency data (Huang et al. 2022). Following preprocessing of the frequency data, instead of converting it back into phase data for input into our model, we utilize the frequency data directly for training and predicting purposes. Finally, we convert the predicted frequency data back into phase data.
(b) Data Set Partitioning
The satellite clock bias data set is divided into a training set and a test set. The training set is utilized for model parameter learning, while the test set is employed to assess the predictive accuracy of the model.
(c) Data normalization
The satellite clock bias data undergoes normalization. In the context of machine learning models, data normalization represents a crucial preprocessing step. In this paper, MinMaxScaler is used to carry out data normalization processing, and the input and output features of sample data are mapped to the space of [-1,1], which can remove the impact of data dimension and improve the convergence speed of the model.
Following these operations, the data is fed into the model for training and prediction. CNN is utilized to extract local features, while BiGRU captures long-term dependencies in time series both forward and backward. The attention mechanism preserves essential features to a significant extent, with hyperparameter optimization in the CNN model being performed using the BWO algorithm.
(2) BWO Optimization
The performance of deep learning models heavily relies on hyperparameter selection and adjustment. During model training, the BWO algorithm is employed to optimize and update model hyperparameters such as convolution kernel weight, convolution kernel channel, and fully connected layer.
(3) CNN Layer
Upon obtaining model hyperparameters through the BWO optimization algorithm, these parameters are then returned to train and predict within the CNN model.
(4) BiGUR Layer
The output feature vector from the CNN layer is fed into the BiGRU network, where the BiGRU network, through its forward and backward propagation, stacks the feature information to fully explore the correlations within the clock bias data.
(5) Attention Layer
The feature vectors derived from the BiGRU network's training process are fed into the Attention layer. This layer computes the distribution of attention weights and selectively aggregates the feature vectors based on these weights, thereby accentuating the retention of the most salient features.
(6) Model Output
The Attention layer yields the predicted frequency data of the satellite clock biases. Subsequent frequency-to-phase conversion processes are applied to derive the anticipated phase data of the atomic clocks.
(7) Model Evaluation
The performance of the satellite clock bias prediction model, which integrates BWO-CNN-BiGRU-Attention, is assessed using five key metrics: Root Mean Square Error (RMSE), Mean Absolute Error (MAE), Mean Squared Error (MSE), Mean Absolute Percentage Error (MAPE), and the coefficient of determination R2.
In this context, y ^ 1 , y ^ 2 , , y ^ n represents the predicted value of the clock bias, while y 1 , y 2 , , y n denotes the actual value of the clock bias, and n is the number of samples. The smaller the values of the model's parameters such as RMSE, MAE, MSE, and MAPE, the better the model's predictive performance. A coefficient of determination, denoted as R2, closer to 1 signifies a higher degree of fit of the model to the sample data, thereby indicative of superior predictive capabilities.
Table 1. The Evaluation Indicators of SCB Prediction Model.
Table 1. The Evaluation Indicators of SCB Prediction Model.
Evaluation Indicators Calculation Formula How to Evaluate
RMSE
(Root Mean Square Error)
RMSE = 1 n i = 1 n ( y ^ i y i ) 2 the lower the value,
the better the model
MAE
(Mean Absolute Error)
MAE = 1 n i = 1 n | y ^ i y i |
MSE
(Mean Square Error)
MSE = 1 n i = 1 n ( y ^ i y i ) 2
MAPE
(Mean Absolute Percentage Error)
MAPE = 100 % n i = 1 n | y ^ i y i y i |
R2
(R-Square)
R 2 = 1 i = 1 n ( y ^ i y i ) 2 i = 1 n ( y ¯ i y i ) 2 The higher the value,
The better the model

4. Case Analysis

This paper utilizes the precise 300-second clock bias data of the BeiDou Navigation Satellite System (BDS) provided by the International GNSS Monitoring and Assessment System (IGMAS) (IGS, 2024), spanning from September 6, 2023, to November 9, 2023 - a total of 65 days (denoted as 20230906-20231220) - for experimental validation of the effectiveness and universality of the satellite clock bias forecasting model based on BWO-CNN-BiGRU-Attention.
To fully leverage the aforementioned data, the 65-day dataset is segmented into five intervals of 5 days each, denoted as Segment 1 (20230906-20231020), Segment 2 (20230911-20231025), Segment 3 (20230916-20231030), Segment 4 (20230921-20231104), and Segment 5 (20230926-20231109). For each segment, the first 30 days of clock bias data are employed as the training set to conduct short-term (6 hours), medium-term (3 days), and long-term (15 days) forecasts.
The primary focus of this paper is to evaluate the performance of the BDS satellite clock bias forecasting model using the BWO-CNN-BiGRU-Attention approach. Statistics indicate that the BDS satellites in orbit primarily operate Rb (rubidium) and H (hydrogen) atomic clocks. Table 2 provides the PRN (Pseudo-Random Noise) codes corresponding to different types of BDS satellite-borne atomic clocks.
Upon analysis, it is found that the GEO (Geostationary Earth Orbit) data from the 65-day BDS clock bias data provided by IGMAS is insufficient, while the IGSO (Inclined Geosynchronous Satellite Orbit) and MEO (Medium Earth Orbit) satellite clock bias data are complete without any missing instances. Consequently, one satellite-borne atomic clock is randomly selected from both the IGSO and MEO for experimentation (PRN27 and PRN40).
Below, taking the clock bias data of PRN27 for the Segment 1 period as an example, the process of using the BWO-CNN-BiGRU-Attention model to forecast 15 days ahead based on 30 days of clock bias data is illustrated.

4.1. Data Preprocessing

During the period from September 6, 2023, to October 20, 2023, the raw phase data of the PRN27 satellite-borne atomic clock is presented in Figure 7.
Examination of Figure 7 reveals that anomalies are not readily discernible from the raw phase data. To overcome this, a phase-to-frequency transformation is applied to obtain the clock bias frequency data. As shown in Figure 8 (above), certain outliers within the clock bias data become clearly identifiable. Utilizing the Median Absolute Deviation (MAD) method, outliers in the PRN27 clock bias frequency data are removed, with a threshold factor set to 3. The outliers are then interpolated using a linear interpolation method, as depicted in Figure 8 (below).

4.2. Training and Prediction with the BWO-CNN-BiGRU-Attention Model

This study employs two 2×2 convolutional layers and max pooling layers to extract features from satellite clock bias data, with the ReLU function chosen as the activation function for the convolutional layers and max pooling selected for the pooling layers. After processing by the convolutional and pooling layers, the feature vectors are passed through a fully connected layer structure and activated by the Sigmoid function to obtain the final output vector of the CNN layer. Considering the training time, the initial population size of the BWO algorithm is set to n = 5, with a maximum number of iterations T = 3, and the model prediction step is set to 10, meaning that the frequency value of the 11th moment is predicted using the frequency values of the preceding 10 moments. The main hyperparameters are set as shown in Table 3 below:
For PRN27, the 15-day frequency prediction results are depicted in Figure 9:
For PRN40, the corresponding 15-day frequency prediction outcomes are illustrated in Figure 10:
Following the completion of the frequency predictions, the forecasted frequency information is transformed into predicted phase information, with the accuracy of the clock bias predictions determined by RMSE between the actual and predicted phase data. Figure 11 and Figure 12 present the first-order differences between the actual and predicted phases for PRN27 and PRN40 over a 15-day period, respectively, where the green line represents the actual first-order phase difference, and the red line indicates the predicted first-order phase difference.
Table 4 presents the clock bias prediction results for two BDS satellites, PRN27 (Hydrogen) and PRN40 (Hydrogen), based on the BWO-CNN-BiGRU-Attention model.

4.3. Results Comparison

To validate the effectiveness of the satellite clock bias prediction using the BWO-CNN-BiGRU-Attention model, this paper compares its performance with that of the CNN-BiGRU-Attention, CNN-BiGRU, and BiGRU models on the same dataset. The average values from five different time segments are used to represent the clock bias prediction results of the models, with the RMSE results detailed in the following Table 5:
For ease of comparison, each evaluation metric has been normalized to the same dimension. A detailed comparison of the prediction evaluation metrics for the BWO-CNN-BiGRU-Attention, CNN-BiGRU-Attention, CNN-BiGRU, and BiGRU models can be found in Figure 13 and Figure 14. Due to space constraints, only the results for the 15-day forecast are presented in the comparative figures. It is evident that the satellite clock bias prediction model based on BWO-CNN-BiGRU-Attention outperforms the others across all metrics.
Additionally, the navigation satellite clock bias prediction model based on the BWO-CNN-BiGRU-Attention framework has also demonstrated optimal performance for 6-hour and 3-day forecasts.
This paper utilizes the average prediction results of PRN27(H) and PRN40(H), two satellites, to represent the predictive outcomes for BDS satellite clock bias across various models. According to the aforementioned calculations, the RMSE results for the 6-hour, 3-day, and 15-day predictions of the satellite clock bias based on the BWO-CNN-BiGRU-Attention model are 0.078 ns, 0.476 ns, and 2.13 ns, respectively. In contrast, the RMSE for the 6-hour, 3-day, and 15-day predictions based on the CNN-BiGRU-Attention model are 0.113 ns, 0.856 ns, and 6.18 ns, respectively. For the CNN-BiGRU model, the RMSE for the same prediction periods are 0.083 ns, 0.981 ns, and 4.84 ns, respectively. Lastly, the BiGRU model yields an RMSE of 0.115 ns, 0.734 ns, and 7.94 ns for the 6-hour, 3-day, and 15-day predictions, respectively.
In summary, the computational outcomes indicate that the satellite clock bias predictions yielded by the BWO-CNN-BiGRU-Attention model are markedly superior to those of the CNN-BiGRU-Attention, CNN-BiGRU, and BiGRU models. Specifically, the model's predictions for the 6-hour, 3-day, and 15-day intervals have demonstrated improvements of 30%, 44%, and 66%, respectively, over the CNN-BiGRU-Attention model. Furthermore, the predictions have shown enhancements of 6%, 51%, and 56%, respectively, when compared to the CNN-BiGRU model. Lastly, in comparison to the BiGRU model, the improvements are 32%, 35%, and 73% for the same prediction intervals. These results underscore the efficacy of the BWO-CNN-BiGRU-Attention model in providing more accurate predictions for satellite clock biases over various time horizons.

5. Conclusions

Satellite clock bias is one of the primary sources of error in satellite navigation and positioning, particularly for real-time applications. The significance of predicting satellite clock bias lies in enhancing the accuracy and timeliness of satellite navigation and positioning systems, thereby providing users with more precise and up-to-date location services. By conducting assessments and forecasting research on satellite clock bias, the precision and reliability of satellite navigation and positioning can be significantly improved.
This paper introduces a novel method for predicting satellite clock bias based on the Beluga Whale Optimization algorithm optimized CNN-BiGRU-Attention model. Experimental results demonstrate that this method offers high predictive accuracy and efficiency. The RMSE for short-term (6-hour), medium-term (3-day), and long-term (15-day) predictions are 0.078 ns, 0.476 ns, and 2.13 ns, respectively. These results are notably superior to those of the CNN-BiGRU-Attention, CNN-BiGRU, and BiGRU models. Future research can further explore the application and refinement of this method to meet the practical demands of navigation systems.

Author Contributions

SWL and CJP proposed the method used in this paper and complete the experiment; HCB, LYX, WML, and ZH discussed and analyzed the experimental results and the manuscript, and proposed suggestions for improvements; ZR, FGT, and ZT discussed and analyzed the data preprocessing of the experimental process and the experimental results. All authors read and approved the final manuscript.

Funding

There is no any funding or project support for this manuscript.

Acknowledgments

Thanks Senior Engineer Shi Fengfeng, Lin Yuting, Xu Jinfeng and Guo Hairong from Beijing Satellite Navigation Center for their discussion and guidance. Thanks to International GNSS Monitoring & Assessment System (http://www.igmas.org) for providing data support.
Competing Interests: The authors declare that they have no competing interests.

References

  1. Ai QS, Xu TH, Li JJ, Xiong HW (2016) The short-term forecast of BeiDou satellite clock bias based on wavelet neural network. In: 2016 China Satellite Navigation Conference, pp.145-154. [CrossRef]
  2. Bai HW, Cao QQ, An SB (2023) Mind evolutionary algorithm optimization in the prediction of satellite clock bias using the back propagation neural network. Scientific Reports 13:2095. [CrossRef]
  3. Cai CL, Liu MY, Li PC, Li ZX, Lv KH (2024) Enhancing satellite clock bias prediction in BDS with LSTM-attention model. GPS Solutions 28:92. [CrossRef]
  4. Cho K, Merriënboer BV, Gulcehre C,, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1724–1734, Doha, Qatar.
  5. Cui XQ, Jiao WH (2005) Grey system model for the satellite clock error predicting. Geomatics and Information Science of Wuhan University 30(5):447-450.
  6. DataCamp (n.d.) An Introduction to Convolutional Neural Networks (CNNs). https://www.datacamp.com/tutorial/introduction-to-convolutional-neural-networks-cnns. Accessed 21st, August, 2024.
  7. Davis J, Bhattarai S, Ziebart M (2012) Development of a Kalman filter based GPS satellite clock time-offset prediction algorithm. In: 2012 IEEE European Frequency and Time Forum, pp,152-156. [CrossRef]
  8. He LN, Zhou HR, Liu ZQ, Wen YL, H XF (2019) Improving clock prediction algorithm for BDS-2/3 satellites based on LS-SVM method. Remote Sensing 11(21):2554. [CrossRef]
  9. He SF, Liu JL, Zhu XW, Dai ZQ, Li D (2023) Research on modeling and predictions of BDS-3 satellite clock bias using the LSTM neural network model. GPS Solutions 27(3):108. [CrossRef]
  10. Huang BH, Ji ZX, Zhai RJ, Xiao CF, Yang F, Yang BH, Wang YP (2021) Clock bias prediction algorithm for navigation satellites based on a supervised learning long short-term memory neural network. GPS Solutions 25(2):80. [CrossRef]
  11. Huang BH, Yang BH, Li MG, Guo ZK, Mao JY, Wang H (2022) An improved method for MAD gross error detection of clock error. Geomatics and information science of Wuhan University 47(5):747-752. [CrossRef]
  12. Huang GW, Cui BB, Zhang Q, Fu WJ, Li PL (2018) An improved predicted model for BDS ultra-rapid satellite clock offsets. Remote Sensing 10(1):60. [CrossRef]
  13. IGS (2024). Products. http://www.igs.org/products, accessed 14th, May, 2024.
  14. LeCun Y, Bengio Y, Hinto G (2015) Deep Learning. Nature 521:436-444. [CrossRef]
  15. Lei Y, Zhao DN, Li B, Gao YP (2014) Prediction of satellite clock bias based on wavelet transform and least square support vector machines. Geomatics and Information Science of Wuhan University 39(7):815-819.
  16. Li XY, Dong XR, Zheng K, Liu YT (2013) Research of satellite clock error prediction based on RBF neural network and ARMA model. In: 2013 China Satellite Navigation Conference, pp. 325-334. [CrossRef]
  17. Liang Y, Xu J, Wu M (2022) Elman Neural Network Based on Particle Swarm Optimization for Prediction of GPS Rapid Clock Bias. In: Yang, C., Xie, J. (eds) China Satellite Navigation Conference (CSNC 2022) Proceedings. Lecture Notes in Electrical Engineering, vol 910. Springer, Singapore. [CrossRef]
  18. Liang YJ, Ren C, Yang XF, Pang GF, Lan L (2016) A Grey model based on first difference in the application of the satellite clock bias prediction. Acta Astronom Sinica 56(3):264-277. [CrossRef]
  19. Lu XF, Yang ZQ, Jia XL, Cui XQ (2008) Parameter optimization method of gray system theory for the satellite clock error predicting. Geomatics and Information Science of Wuhan University 33(5):492-495.
  20. Lu YW, Zheng LQ, Hu C (2023) Analysis and comparison of satellite clock error prediction based on various deep learning algorithms. GNSS World of China 48(5):46-55, 91. [CrossRef]
  21. Lv D, Liu GY, Ou JK, Wang SL, Gao M (2022) Prediction of GPS Satellite Clock Offset Based on an Improved Particle Swarm Algorithm Optimized BP Neural Network. Remoet Sensing 14(10), 2407. [CrossRef]
  22. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems:6000-6010.
  23. Wang GC, Liu LT, Xu AG, Su XQ, Liang XH (2014) The application of radial basis function neural network in the GPS satellite clock bias prediction. Acta Geodaetica Et Cartographica Sinica 43(8):803-807.
  24. Wang L, Zhang Q, Huang GW, Tian J (2017) GPS satellite colck bias prediction based on exponential smoothing method. Geomatics and Information Science of Wuhan University 42(7):995-1001. [CrossRef]
  25. Wang X, Chai HZ & Wang C (2020) A high-precision short-term prediction method with stable performance for satellite clock bias. GPS Solutions 24, 105. [CrossRef]
  26. Wang X, Chai HZ, Wang C, Xiao GR, Chong Y, Guan XG (2021) Improved wavelet neural network based on change rate to predict satellite clock bias. Survey Review 53(379):325-334. [CrossRef]
  27. Wang YP, Lv ZP, Chen ZS, Huang LY (2014) A method of satellite clock bias prediction based on grey model and wavelet neural network. Journal of Geodesy and Geodynamics 34(3):155-159.
  28. Wang YP, Lv ZP, Chen ZS, Huang LY, Li LY, Gong XC (2016) A new data preprocessing method of satellite clock bias and its application in WNN to predict medium-term and long-term clock bias. Geomatics and Information Science of Wuhan University 41(3):373-379. [CrossRef]
  29. Wang YP, Lu ZP, Qu YY, Li LY, Wang N (2017) Improving prediction performance of GPS satellite clock bias based on wavelet neural network. GPS Solutions 21(2):523-534.
  30. Xi C, Cai CL, Li SM, Li XH, Li ZB, Deng KQ (2014) Long-term clock bias prediction based on an ARMA model. Chinese Astronomy and Astrophysics 38(3):342-354. [CrossRef]
  31. Xu JY, Zeng AM (2009) Application of ARIMA (0,2,q) model to prediction of satellite clock error. Journal of Geodesy and Geodynamics 29(5):116-120.
  32. Yu Y, Huang M, Wang CY, Hu R, Duan T (2020) A new BDS-2 satellite clock bias prediction algorithm with an improved exponential smoothing method. Applied Sciences 10(21):7456. [CrossRef]
  33. Zhang GC, Han SH, Ye J, Hao RZ, Zhang JC, Li X, Jia K (2022) A method for precisely predicting satellite clock bias based on robust fitting of ARMA models. GPS Solutions 26:3. [CrossRef]
  34. Zheng ZY, Lu XS, Chen YQ (2008) Improved grey model and application in Real-Time GPS satellite clock bias prediction. In: 2008 Fourth International Conference on Natural Computation, pp. 419-423. [CrossRef]
  35. Zhong CT, Li G, Meng Z (2022) Beluga whale optimization: A novel nature-inspired metaheuristic algorithm. Knowledge - Based Systems, 5 (251). [CrossRef]
  36. Zhu XW, Xiao H, Yong SW, Zhuang ZW (2008) The Kalman algorithm used for satellite clock offset prediction and its performance analysis. Journal of Astronautics 29(3):966-970.
Figure 1. Structural Diagram of CNN.
Figure 1. Structural Diagram of CNN.
Preprints 115939 g001
Figure 2. Structural Diagram of GRU.
Figure 2. Structural Diagram of GRU.
Preprints 115939 g002
Figure 3. Structural Diagram of BiGRU.
Figure 3. Structural Diagram of BiGRU.
Preprints 115939 g003
Figure 4. Structural Diagram of Attention.
Figure 4. Structural Diagram of Attention.
Preprints 115939 g004
Figure 5. The Flowchart of BWO Algorithm.
Figure 5. The Flowchart of BWO Algorithm.
Preprints 115939 g005
Figure 6. Structural Diagram of SCB Prediction Model based on BWO-CNN-BiGRU-Attention.
Figure 6. Structural Diagram of SCB Prediction Model based on BWO-CNN-BiGRU-Attention.
Preprints 115939 g006
Figure 7. Raw Clock Bias Phase Data of PRN27.
Figure 7. Raw Clock Bias Phase Data of PRN27.
Preprints 115939 g007
Figure 8. Clock Bias Frequency Data of PRN27.
Figure 8. Clock Bias Frequency Data of PRN27.
Preprints 115939 g008
Figure 9. The 15-day Clock Bias Frequency Prediction Performance for PRN27 during Segment 1.
Figure 9. The 15-day Clock Bias Frequency Prediction Performance for PRN27 during Segment 1.
Preprints 115939 g009
Figure 10. The 15-day Clock Bias Frequency Prediction Performance for PRN40 during Segment 1.
Figure 10. The 15-day Clock Bias Frequency Prediction Performance for PRN40 during Segment 1.
Preprints 115939 g010
Figure 11. Actual and Predicted First-order Phase Differences for PRN27's 15-day Prediction.
Figure 11. Actual and Predicted First-order Phase Differences for PRN27's 15-day Prediction.
Preprints 115939 g011
Figure 12. Actual and Predicted First-order Phase Differences for PRN40's 15-day Prediction.
Figure 12. Actual and Predicted First-order Phase Differences for PRN40's 15-day Prediction.
Preprints 115939 g012
Figure 13. Comparison of Normalized Evaluation Metrics for the 15-day Predictions of PRN27 and PRN40.
Figure 13. Comparison of Normalized Evaluation Metrics for the 15-day Predictions of PRN27 and PRN40.
Preprints 115939 g013
Figure 14. Detailed Comparison of Evaluation Metrics for the 15-day Predictions of PRN27 and PRN40.
Figure 14. Detailed Comparison of Evaluation Metrics for the 15-day Predictions of PRN27 and PRN40.
Preprints 115939 g014
Table 2. Spaceborne Atomic Clock Types of BDS.
Table 2. Spaceborne Atomic Clock Types of BDS.
Orbits PRN/Types
GEO
(7)
01(Rb) 02(Rb) 03(Rb) 04(Rb) 05(Rb) 59(H) 60(H)
IGSO (18) 06(Rb) 07(Rb) 08(Rb) 09(Rb) 10(Rb) 16(Rb) 38(H) 39(H) 40(H)
11(Rb) 12(Rb) 13(Rb) 14(Rb) 19(Rb) 20(Rb) 21(Rb) 22(Rb) 23(Rb)
MEO
(19)
24(Rb) 25(Rb) 26(H) 27(H) 28(H) 29(H) 30(H) 32(Rb) 33(Rb) 34(H)
35(H) 36(Rb) 37(Rb) 41(Rb) 42(Rb) 43(H) 44(H) 45(H) 46(H)
Table 3. Main Hyperparameter Selection for BWO-CNN-BiGRU-Attention Network Model.
Table 3. Main Hyperparameter Selection for BWO-CNN-BiGRU-Attention Network Model.
No Parameter Value
1 Population size 5
2 Max generation 3
3 Convolution kernel size 2×2
4 Activation function Relu
5 Optimizer Adam
6 Gradient threshold 1
7 Initial learn rate 0.01
8 Learn rate drop factor 0.0001
9 Input dimension size 10
10 Output dimension size 1
11 Max epochs 50
12 Attention mechanism Self-attention
Table 4. Prediction RMSE Results for PRN27 and PRN40 Based on BWO-CNN-BiGRU-Attention (ns).
Table 4. Prediction RMSE Results for PRN27 and PRN40 Based on BWO-CNN-BiGRU-Attention (ns).
Time PRN27 (MEO) PRN40 (IGSO)
6 hours 3 days 15 days 6 hours 3 days 15 days
Segment 1 6.46E-11 2.80E-10 3.19E-09 1.10E-10 1.28E-09 3.59E-09
Segment 2 3.12E-11 4.38E-10 1.89E-09 9.91E-11 3.63E-10 1.90E-09
Segment 3 3.78E-11 2.33E-10 2.53E-09 1.13E-10 4.53E-10 1.66E-09
Segment 4 5.55E-11 3.21E-10 1.48E-09 9.81E-11 4.53E-10 2.90E-09
Segment 5 9.40E-11 8.51E-11 8.54E-10 7.51E-11 1.10E-09 1.31E-09
Average(Segments1-5) 5.66E-11 2.72E-10 1.99E-09 9.90E-11 6.79E-10 2.27E-09
Table 5. RMSE Comparison of Clock Bias Prediction Results for Various Models (ns).
Table 5. RMSE Comparison of Clock Bias Prediction Results for Various Models (ns).
Model PRN27 (MEO) PRN40 (IGSO)
6 hours 3 days 15 days 6 hours 3 days 15 days
Average(Segments1-5) CNN-BiGRU-Attention 8.06E-11 8.49E-10 6.34E-09 1.45E-10 8.63E-10 6.01E-09
CNN-BiGRU 7.18E-11 9.12E-10 5.21E-09 9.46E-11 1.05E-09 4.47E-09
BiGRU 7.11E-11 3.37E-10 8.95E-09 1.59E-10 1.13E-09 6.93E-09
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated