You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

29 October 2023

The Short Time Prediction of the Dst Index Based on the Long-Short Time Memory and Empirical Mode Decomposition–Long-Short Time Memory Models

,
,
and
1
Institute of Space Weather, School of Atmospheric Physics, Nanjing University of Information Science & Technology, Nanjing 210044, China
2
State Key Laboratory of Space Weather, Chinese Academy of Sciences, Beijing 100190, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Applications of Machine Learning in Earth Sciences—2nd Edition

Abstract

The Dst index is the geomagnetic storm index used to measure the energy level of geomagnetic storms, and the prediction of this index is of great significance for geomagnetic storm studies and solar activities. In contrast to traditional numerical modeling techniques, machine learning, which emerged decades ago based on rapidly developing computer hardware and software and artificial intelligence methods, has been unprecedentedly developed in geophysics, especially solar–terrestrial space physics. This study uses two machine learning models, the LSTM (Long-Short Time Memory, LSTM) and EMD-LSTM models (Empirical Mode Decomposition, EMD), to model and predict the Dst index. By building the Dst index data series from 2018 to 2023, two models were built to fit and predict the data. Firstly, we evaluated the influences of the learning rate and the amount of training data on the prediction accuracy of the LSTM model, and finally, 10−3 was thought to be the optimal learning rate. Secondly, the two models were used to predict the Dst index in the solar active and quiet periods, respectively, and the RMSE (Root Mean Square Error) of the LSTM model in the active period was 7.34 nT and the CC (correlation coefficient) was 0.96, and those of the quiet period were 2.64 nT and 0.97. The RMSE and CC of the EMD-LSTM model were 8.87 nT and 0.93 in the active period and 3.29 nT and 0.95 in the quiet period. Finally, the prediction accuracy of the LSTM model in the short time period was slightly better than the EMD-LSTM model. However, there will be a problem of prediction lag, which the EMD-LSTM model can solve and better predict the geomagnetic storm.

1. Introduction

As an essential physical field of the Earth, the geomagnetic field is an integral part of space weather, which has a complex spatial structure and temporal evolution characteristics. A deep understanding of the spatial and temporal characteristics, origin, and relationship between the geomagnetic field and other geophysical phenomena provides excellent help for studying the magnetic field of the Earth and other planets. The Dst index, an important index reflecting perturbation of the geomagnetic field, was first proposed by Sugiura in 1964. It is also known as the geomagnetic hourly geomagnetic index, which describes the process of magnetic storms and reflects changes in equatorial westerly ring current [1]. This index is calculated using the horizontal component H of the geomagnetic field at four middle- and low-latitude observatories after removing the secular variation and the ring current latitudinal correction [2]. The Dst index is one of the most widely used indices in geomagnetism and space physics because it can monitor the occurrence and change of magnetic storms succinctly and continuously. The modeling and prediction of the Dst index have significant scientific and applicated meanings for studying changes in geomagnetic external fields, magnetic storms, and even solar activity.
Numerical methods, such as least squares estimation and spline interpolation [3], are generally used to predict geomagnetic indices. These traditional methods mainly carry out numerical processes like iteration from the data itself, error approximation, seeking extremum, and so on, to obtain the best-approximated value that satisfies the accuracy requirement. However, they cannot highlight the characteristics of the data based on in-depth learning and mining of the existing and constantly emerging new data. In recent years, with the continuous improvement in computing power and the rapid development of artificial intelligence (AI), deep learning, a type of multi-layer neural network learning algorithm, has become an essential part of the machine learning field. Deep learning is usually used to process images, sounds, text signals, etc. It is mainly applied to multi-modal classification problems, data processing, and simulation, as well as artificial intelligence [4]. The Long-Short Time Memory (LSTM) network method [5] is the most widely used simulation and forecasting algorithm model. This method has a complex hidden unit, which can selectively increase or decrease information. It, therefore, can accurately model data with short-term or long-term dependencies in the data for accurate modeling [6]. So far, LSTM has shown excellent forecasting ability in long-term time series data processing [7]. We believe that it is very suitable for forecasting the Dst index. Therefore, in this study, we use LSTM and its mixed model—EMD-LSTM—for Dst index prediction and try to optimize the prediction results.
The rest of this study is organized as follows. Section 2 describes related studies. Section 3 gives the data sources and selection conditions for this paper. Section 4 describes the rationale of the methodology used. Section 5 explains the results of the different methods. The conclusions and discussion are given in Section 6 and Section 7, respectively.

3. Data

3.1. Data Sources

The Dst index used in this study was obtained from the website of the World Data Center for Geomagnetism, Kyoto. The Dst index is an hourly average and can be used directly without special processing. The data from 1957 to 2016 are the final Dst index, the data from 2017 to 2021 are the provisional Dst index, and the data after 2022 are the real-time Dst index. Both 2015 and 2023 are solar active years. Dst records show that three large geomagnetic storms occurred in 2023, so this study mainly focuses on the period from 1 January 2015 to 21 May 2023, from which the appropriate data (see Section 3.2 for details) are selected under different prediction conditions.

3.2. Data Selection

One study showed that the Dst index in 2019 was relatively flat [24]. Most of the indices are between −40 and 20 nT. Therefore, the data from 2019 are chosen as the test set for the quiet period, and the data from the previous year are used as the training set. In contrast, three major geomagnetic storms occurred before 21 May 2023, of which the Dst index most drastically varied during the 24 April storm, with an amplitude of about 200 nT. Thus, the data between 1 January 2023 and 21 May 2023 are selected to test the prediction of the active period, and the data from the previous year are used as the training set/ More attention will be paid to the 24 April test.

4. Method

4.1. The LSTM Model

This study selects the LSTM network model, a gated recurrent neural network model, for forecasting. The basic principle is that the model has a recurrent unit in which the output of the previous moment is the input of the current moment, the output of the current moment is the input of the next moment, and a recurrent unit contains multiple components [25]. The specific process is divided into four parts. The first part is shown in Figure 1.
Figure 1. The first part of a single recurrent unit of the LSTM model.
According to the above figure, we have
I t = σ ( X t W x i + H t 1 W h i + b i ) F t = σ ( X t W x f + H t 1 W h f + b f ) O t = σ ( X t W x o + H t 1 W h o + b o )
σ ( x ) = 1 1 + e x
where the input Xt of the current step and the hidden state Ht−1 of the previous step will be sent into the input gate (It), forget gate (Ft), and output gate (Ot). Xt denotes the input data of the current step; the hidden state Ht−1 is the transformed result of the previous step, and the later Ht can be obtained later in the same way; Wx and Wh are the weights corresponding to the inputs and the hidden state, respectively; and each b represents bias, where all of them are parameters. The outputs of the three gates are obtained using the activation function, i.e., the sigmoid function (σ), which is used to transform the outputs nonlinearly. The second part is shown in Figure 2.
Figure 2. The second part of a single recurrent unit of the LSTM model.
According to the above figure, we have
C ˜ t = tanh ( X t W x c + H t 1 W h c + b c )
tanh ( x ) = 1 e 2 x 1 + e 2 x
The LSTM requires calculating the candidate memory cell C ˜ t , which is similar to the previous three gates but uses the tanh function as the activation function. After that, the third part is shown in Figure 3.
Figure 3. The third part of a single recurrent unit of the LSTM model.
According to the above figure, we have
C t = F t C t 1 + I t C ˜ t
The information flow in the hidden state is controlled by input gates, forget gates, and output gates with element domains in [0, 1]. The computation of the current step memory cell Ct combines the information of cell Ct−1 and the current step candidate memory cell and then controls the information flow through the forget gate and the input gate. The fourth part is shown in Figure 4.
Figure 4. The fourth part of a single recurrent unit of the LSTM model.
According to the above figure, we have
H t = O t tanh ( C t )
With the memory cells, the information flow from the memory cells to the hidden state Ht is controlled by an output gate.
After the end of a cycle unit, the hidden state Ht updates parameters and weights with the loss function and optimization algorithm, which means one training is completed. The loss function with Adam’s optimization algorithm represents the mean square error.
The error(gt) between the predicted value and the measurement is
g t = ( y ^ ( i ) y ( i ) ) 2
The optimization algorithm is used to iterate the parameters. The degree of parameter change in each training can be controlled by changing the learning rate. pt (Wx, Wh, b) can be obtained using Adam’s algorithm as follows
v t = γ 1 v t 1 + ( 1 γ 1 ) g t s t = γ 2 s t 1 + ( 1 γ 2 ) g t g t
v ^ t = v t 1 γ 1 t , s ^ t = s t 1 γ 2 t
g ^ t = η v ^ t s ^ t + τ p t = p t 1 g ^ t
where η is the learning rate and also a hyperparameter that needs to be set, and γ is a hyperparameter with domain [0, 1]. After the training set is put into the model and the training is complete, it is then put into the test set for forecasting, and the results are shown graphically against the exact values. The modeling error is expressed using the RMSE and the CC as follows,
RMSE ( X , h ) = 1 m i = 1 m ( h ( x i ) y i ) 2
CC = ( x x ¯ ) ( y y ¯ ) ( x x ¯ ) 2 ( y y ¯ ) 2

4.2. The EMD-LSTM Model

Another forecasting model used in this study is the EMD-LSTM model, which is a model that combines the Empirical Mode Decomposition (EMD) method with LSTM. The EMD was proposed by Huang et al. [26] to decompose signals into eigenmodes. The basic principle of the EMD algorithm is firstly find the maximum and minimum points of the original signal X(t) and then fit these extreme points using curve interpolation to obtain the upper envelope Xmax(t) and lower envelope Xmin(t) and finally determine the average of upper and lower values:
m 1 ( t ) = X max ( t ) + X min ( t ) 2
The first intrinsic mode function (IMF) can be obtained by subtracting the original signal X(t) from the m1(t), and the other IMFs can be obtained in the same way. Finally, the residual component (res) with only one extreme point that cannot be further decomposed can be acquired, which means the algorithm is finished. The core of the EMD algorithm is to decompose the original signal into several IMF components and one res component, which are of decreasing complexity. The sum of all these components is the original signal.
Compared with the LSTM algorithm, the EMD-LSTM is based on the idea that the EMD algorithm decomposes the training data and then puts the decomposed data into the LSTM model for training. Since most of the components derived from EMD decomposition have lower complexity, and LSTM can obtain better fitting results for them. So, the error of the final prediction result will be reduced to some extent. The working flow chart is shown in Figure 5.
Figure 5. The overall structure of the EMD-LSTM model.
In Figure 5, the training data are first decomposed using the EMD algorithm, then each component is trained using the LSTM model, merging all predictions to obtain the final prediction.

5. Results

After determining the modeling Dst data, the exact influences of different parameters on the prediction accuracy of the LSTM model are tested, and the parameters are also fixed for subsequent training. In order to test the validity and accuracy of the model, the Dst index is modeled and predicted for the solar quiet and active periods, respectively, and the consequent predictions are analyzed.

5.1. The Results of the Short-Period Prediction using the LSTM Model

First, we selected 90-day data (90 days since 1 January 2015) and 1096-day data (three years since 1 January 2015) as the training data to test the performance of the LSTM model for short-term (one-hour ahead) prediction, which aims to compare influences on the prediction accuracy at different learning rates and lengths of training data. The learning rate directly affects the degree of parameter changes after training; the more minor the parameter, the lower the amplitude of change during each training, and the more training times are needed to achieve the ideal fitting effect. Usually, the learning rate is chosen to be 10−3, 10−4, 10−5, or even smaller. Since the number of training times directly affects the training time, 10−3, 10−4, and 10−5 are the learning rates used to compare the prediction results. Here, we list the predictions based on 90 days with different learning rates and trainings.
Figure 6, Figure 7 and Figure 8 and Table 2 show the one-hour ahead prediction of a 7-day length (168 h) using the 90-day data. As the learning rate decreases, the training number increases by around 400% to keep the close prediction error; the RMSE only decreases by 0.07 nT, the CC increases by 0.01, and the learning rate decreases from 10−3 to 10−5.
Figure 6. The model prediction through 650 trainings with a learning rate of 10−3.
Figure 7. The model prediction through 3400 trainings with a learning rate of 10−4.
Figure 8. The model prediction through 20,000 trainings with a learning rate of 10−5.
Table 2. The 7-day length prediction using 90-day data.
The predictions based on 1096-day data are as follows.
Figure 9, Figure 10 and Figure 11 and Table 3 show the one-hour ahead prediction results for a 7-day length using three years of data since 2015. The situation is similar to the 90-day results: the RMSE only decreases by 0.03 nT, while the learning rate changes from 10−3 to 10−5. In addition, the prediction error decreases by about 1.3 nT on average relative to the RMSE of the 90-day data, and the CC improves by 0.03, which suggests that increasing the prediction data can effectively improve the accuracy.
Figure 9. The model prediction through 650 trainings with a learning rate of 10−3.
Figure 10. The model prediction through 3300 trainings with a learning rate of 10−4.
Figure 11. The model prediction through 18,000 trainings with a learning rate of 10−5.
Table 3. The 7-day length prediction using 1096-day data.
However, the above figures show that the model-predicted values have a noticeable lag using different training data lengths. This lag phenomenon has also been observed in other studies using LSTM for prediction, such as Cai et al. [27] and Yin et al. [28]. This stems from the fact that when using the LSTM model for time-series prediction, the model algorithm is calculated using the values in the fixed time window as samples. In order to minimize the error, using the value at moment t as the prediction value of moment t + 1 not only requires no additional operations, but the error is very tiny, and the algorithm tends to use the value at moment t or t − 1 as the prediction value, thus generating a lag in the final prediction. Commonly used solutions include increasing the training data length and indirect prediction. To take advantage of the feature of the EMD-LSTM method and obtain segmented average values and then perform LSTM prediction, we use Figure 9 as an example and perform tests as follows.
In Figure 12, under the same conditions as in Figure 9, using the EMD-LSTM model, the original data are decomposed into 14 components using the EMD algorithm, which are individually predicted with LSTM fitting and finally merged to obtain the prediction results. The EMD-LSTM model fits better than the LSTM model because the EMD algorithm can split the data series by detecting the peak and valley values of the two curves, which directly and effectively improves the lag in the prediction.
Figure 12. Results of the EMD-LSTM model when using the same parameters as in Figure 9.
According to the above tests, a lower learning rate has few accuracy improvements. At the same time, the training times and training time increased significantly, so we chose a learning rate of 10−3 for subsequent model training.
By observing the actual Dst data, the data from 2019 are selected as the test set to predict the quiet period, and the data from 1 January 2023 to 21 May are selected as the test set for the active period prediction. The results are as follows.
Figure 13 shows the results of using the 2018 Dst data with one-hour ahead forecasting for 2019. The results show that the fitted values agree well with the actual values. The RMSE is 2.64 nT, the CC is 0.97, and the precision of the results is between those obtained by LSTM using 90 and 1096 days of data.
Figure 13. Results of model prediction through 750 training sessions at a learning rate of 10−3.
In order to further test the model, the dramatic changes in the Dst index in 2023 are now predicted based on the 2022 data, and we expect to model and predict the corresponding geomagnetic storms. The following are the results based on the LSTM model.
Figure 14 shows the one-hour ahead prediction of the data from 1 January to 21 May 2023, with an RMSE of 7.34 nT and a CC of 0.96, using the Dst data of 2022. The storm with the largest amplitude occurred on 24 April, and its prediction results are shown in Figure 15, with an RMSE of 20.47 nT and a CC of 0.91 before and after the storm (300 h). It is found that the fitting before and after the storm valley is good, while the prediction of the main phase is not good, and the difference is more than 100 nT.
Figure 14. The model prediction through 750 trainings with a learning rate of 10−3.
Figure 15. Predictions before and after the 24 April 2023 geomagnetic storm.

5.2. The Results of the Short-Period Prediction Using the EMD-LSTM Model

In order to effectively solve the time lag problem and to improve the accuracy of the prediction, the EMD-LSTM method is used to predict the Dst index in 2019 and 2023. The following is part of the forecast results for the quiet period.
Figure 16 shows that the EMD algorithm mathematically decomposes the original Dst index data into several wavelength components. Each part possesses different characteristics (determined using the original data), and the complexity of these components decreases in order. The error during model training is mainly concentrated in the first few components rather than the entire Dst index, so using EMD-LSTM can make the prediction closer to the actual value.
Figure 16. Data decomposition using EMD based on the 2018 Dst index.
We first predict the data from 2019.
Figure 17 shows the result of using the Dst data from 2018, making a one-hour ahead prediction and predicting the data from 2019. The RMSE is 3.29 nT, and the CC is 0.95, which makes the prediction effect more satisfactory. Compared with the LSTM method, the RMSE increased by 0.65 nT and the CC decreased by 0.02.
Figure 17. Results of model prediction through 600 training sessions with a learning rate of 10−3.
Next, we test the prediction for 2023.
Figure 18 shows different parts of the 2022 Dst index decomposed using EMD.
Figure 18. Results of the Dst index data decomposition using the EMD method for 2022.
Figure 19 shows a one-hour ahead prediction from 1 January to 21 May 2023 using the 2022 Dst data. The results show that the RMSE is 8.87 nT and the CC is 0.93. There is an increase of 1.53 nT in the RMSE and a decrease of 0.03 in the CC compared with the LSTM method, and the fitting of the main phase is obviously better than the LSTM method.
Figure 19. The model prediction through 650 trainings with a learning rate of 10−3.
For further inspection, we list the modeling of the geomagnetic storm that occurred on 24 April 2023.
Figure 20 shows that the modeling is highly consistent in the start timing, amplitude, and recovery value of the magnetic storm. In the period before and after the storm (300 h), the RMSE is 16.91 nT and the CC is 0.96; they reduced by 3.56 nT and improved by 0.05 compared with the LSTM model. The fitted values are slightly higher during the initial and main phases, while they are more consistent during the recovery phase. The Dst index during the main phase can be effectively and precisely predicted. In addition, the lag in the predicted values is greatly improved by the indirect prediction with the decomposition of the original data using the EMD algorithm.
Figure 20. Predictions of the geomagnetic storm that occurred on 24 April 2023.
Comparing the two methods, from the viewpoint of the RMSE and CC, LSTM is slightly better than EMD-LSTM, but EMD-LSTM is better during the magnetic storm.

6. Conclusions

In this study, we trained and predicted the LSTM and EMD-LSTM models by observing and comparing the exact Dst index and selecting appropriate data segments. The factors affecting the prediction accuracy of the LSTM model were first tested, and the parameters used in the LSTM model were determined. Then, the two models were used to predict the Dst index for the quiet and active periods. After analyzing the prediction results, the related conclusions are reached as follows:
(1)
In predicting the Dst index using the LSTM model, a decrease in the learning rate and an increase in the number of training times have no significant influences on improving the prediction accuracy. When using 90-day length data for 7-day prediction, the change in the RMSE is within 0.1 nT and the CC varies around 0.01 under different learning rates. The change in the RMSE is within 0.05 nT and the CC does not change when using 1096-day data, whereas increasing the training data length can improve the prediction accuracy. The prediction accuracy can be improved by increasing the length of the training data: the RMSE decreases from an average of 3.25 nT to 1.95 nT and the CC improves from 0.91 to 0.94 when using 1096 days of data, which suggests that the base dataset is a critical factor in controlling the prediction.
(2)
The LSTM and EMD-LSTM models have better Dst prediction for the quiet period than the active period. The reason is that the fluctuation in the Dst index during the quiet period is slight (±50 nT), so the resulting model based on higher temporal resolution (time-averaged) is robust. The interferences from other factors, especially the solar activity, are less in the quiet period; combining the appropriate training rate and training times can predict the Dst changes better. During geomagnetic storms, the fluctuation amplitude in Dst caused by the solar wind can reach several hundred nT or more. Compared with the use of the LSTM model, the error is significantly reduced when training data are decomposed using the EMD algorithm and then put into the LSTM training, particularly during the big magnetic storm, but the overall prediction accuracy is lower than that of LSTM.
(3)
Although the overall prediction accuracy of the LSTM model is slightly higher than the EMD-LSTM model (the RMSE is reduced by about 1.5 nT and the CC is improved by about 0.03), there is a noticeable lag in the prediction results of the former. EMD-LSTM significantly improves the lag in the prediction results of LSTM. In addition, the prediction accuracy of EMD-LSTM during magnetic storms is better than that of LSTM. In practical applications, it is appropriate to select a suitable model by referencing the intensity of solar activity, such as the sunspot number, and if its value is high, the EMD-LSTM model can be chosen, or the LSTM model can be chosen if the situation is opposite. If the error requirement is not strict, using the LSTM method with a higher learning rate is economical to reduce the computation amount and satisfy the accuracy requirement.

7. Discussion

We tested the LSTM and EMD-LSTM models in terms of their ability for Dst index prediction. Our results prove the effectiveness of deep learning methods, which can be used to improve the prediction accuracy of the model by changing the model parameters and length of the training data, processing the training data, replacing the training data. Thus, this method is worthy of deeper study in the future. The newly launched MSS-1 satellite [29], which provides high accuracy and nearly east–west oriented (inclination angle of 41°) magnetic field data, can lay a solid foundation for studying the removal of external field interference and even for studying short-term changes in the ionospheric field and the magnetospheric ring current.
There are also some shortcomings in this study. The models only use one kind of data for training and prediction and lack the necessary physical constraints. Thus, the next step will be to screen other space weather indices and select indices with higher correlation with the Dst index, such as the magnetospheric ring current index (RC index) or the geomagnetic perturbation index (Ap index), for constraints and co-estimating, using multiple input parameters for training and predicting. In this way, realistic prediction results can be obtained by adding physical constraints.

Author Contributions

Conceptualization, Y.F.; methodology, J.Z. (Jinyuan Zhang) and Y.F.; software, Y.L. and J.Z. (Jiaxuan Zhang); validation, Y.F.; writing—original draft preparation, Y.F.; writing—review and editing, Y.F.; supervision, Y.F.; funding acquisition, Y.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (42250103, 41974073, 41404053), the Macau Foundation and the pre-research project of Civil Aerospace Technologies (Nos. D020308), and the Specialized Research Fund for State Key Laboratories.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

All measurement data will be made available upon request.

Acknowledgments

We acknowledge the High-Performance Computing Center of Nanjing University of Information Science & Technology for their support of this work. We also thank the reviewers for their constructive comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sugiura, M. Hourly values of equatorial Dst for the IGY. Ann. Int. Geophys. Year 1964, 35, 9–45. [Google Scholar]
  2. Wu, Y.Y. Problems and thoughts on the Dst index of geomagnetic activity. Adv. Geophys. 2022, 37, 1512–1519. [Google Scholar]
  3. Tran, Q.A.; Sołowski, W.; Berzins, M.; Guilkey, J. A convected particle least square interpolation material point method. Int. J. Numer. Methods Eng. 2020, 121, 1068–1100. [Google Scholar] [CrossRef]
  4. Sun, Z.J.; Xue, L.; Xu, Y.M.; Wang, Z. A review of deep learning research. Comput. Appl. Res. 2012, 29, 2806–2810. [Google Scholar]
  5. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  6. Yang, L.; Wu, Y.X.; Wang, J.L.; Liu, Y.L. A review of recurrent neural network research. Comput. Appl. 2018, 38, 1–6+26. [Google Scholar]
  7. Liu, J.Y.; Shen, C.L. Forecasting Dst index using artificial intelligence method. In China Earth Science Joint Academic Conference, Proceedings of the 2021 China Joint Academic Conference on Earth Sciences (I)-Topic I Solar Activity and Its Space Weather Effects, Topic II Plasma Physical Processes in the Magnetosphere, Topic III Planetary Physics, Zhuhai, China, 11–14 November 2021; Beijing Berton Electronic Publishing House: Beijing, China, 2021; p. 12. [Google Scholar] [CrossRef]
  8. Kim, R.S.; Cho, K.S.; Moon, Y.J.; Dryer, M.; Lee, J.; Yi, Y.; Kim, K.H.; Wang, H.; Park, Y.D.; Kim, Y.H. An empirical model for prediction of geomagnetic storms using initially observed CME parameters at the Sun. J. Geophys. 2010, 115, A12108. [Google Scholar] [CrossRef]
  9. Kremenetskiy, I.A.; Salnikov, N.N. Minimax Approach to Magnetic Storms Forecasting (Dst-index Forecasting). J. Autom. Inf. Sci. 2011, 3, 67–82. [Google Scholar] [CrossRef]
  10. Tobiska, W.K.; Knipp, D.; Burke, W.J.; Bouwer, D.; Bailey, J.; Odstrcil, D.; Hagan, M.P.; Gannon, J.; Bowman, B.R. The Anemomilos prediction methodology for Dst. Space Weather 2013, 11, 490–508. [Google Scholar] [CrossRef]
  11. Banerjee, A.; Bej, A.; Chatterjee, T.N. A cellular automata-based model of Earth’s magnetosphere in relation with Dst index. Space Weather 2015, 13, 259–270. [Google Scholar] [CrossRef]
  12. Chandorkar, M.; Camporeale, E.; Wing, S. Probabilistic forecasting of the disturbance storm time index: An autoregressive Gaussian process approach. Space Weather 2017, 15, 1004–1019. [Google Scholar] [CrossRef]
  13. Bej, A.; Banerjee, A.; Chatterjee, T.N.; Majumdar, A. One-hour ahead prediction of the Dst index based on the optimum state space reconstruction and pattern recognition. Eur. Phys. J. Plus 2022, 137, 479. [Google Scholar] [CrossRef]
  14. Nilam, B.; Tulasi Ram, S. Forecasting Geomagnetic activity (Dst Index) using the ensemble kalman filter. Mon. Not. R. Astron. Soc. 2022, 511, 723–731. [Google Scholar] [CrossRef]
  15. Chen, C.; Sun, S.J.; Xu, Z.W.; Zhao, Z.W.; Wu, Z.S. Forecasting Dst index one hour in advance using neural network technique. J. Space Sci. 2011, 31, 182–186. [Google Scholar] [CrossRef]
  16. Revallo, M.; Valach, F.; Hejda, P.; Bochníček, J. A neural network Dst index model driven by input time histories of the solar wind–magnetosphere interaction. J. Atmos. Sol.-Terr. Phys. 2014, 110–111, 9–14. [Google Scholar] [CrossRef]
  17. Lu, J.Y.; Peng, Y.X.; Wang, M.; Gu, S.J.; Zhao, M.X. Support Vector Machine combined with Distance Correlation learning for Dst forecasting during intense geomagnetic storms. Planet. Space Sci. 2016, 120, 48–55. [Google Scholar] [CrossRef]
  18. Andriyas, T.; Andriyas, S. Use of Multivariate Relevance Vector Machines in forecasting multiple geomagnetic indices. J. Atmos. Sol.-Terr. Phys. 2017, 154, 21–32. [Google Scholar] [CrossRef]
  19. Lethy, A.; El-Eraki, M.A.; Samy, A.; Deebes, H.A. Prediction of the Dst index and analysis of its dependence on solar wind parameters using neural network. Space Weather 2018, 16, 1277–1290. [Google Scholar] [CrossRef]
  20. Xu, S.; Huang, S.; Yuan, Z.; Deng, X.; Jiang, K. Prediction of the Dst index with bagging ensemble-learning algorithm. Astrophys. J. Suppl. Ser. 2020, 248, 14. [Google Scholar] [CrossRef]
  21. Park, W.; Lee, J.; Kim, K.C.; Lee, J.; Park, K.; Miyashita, Y.; Sohn, J.; Park, J.; Kwak, Y.S.; Hwang, J.; et al. Operational Dst index prediction model based on combination of artificial neural network and empirical model. J. Space Weather Space Clim. 2021, 11, 38. [Google Scholar] [CrossRef]
  22. Hu, A.; Shneider, C.; Tiwari, A.; Camporeale, E. Probabilistic prediction of Dst storms one-day-ahead using full-disk SoHO images. Space Weather 2022, 20, e2022SW003064. [Google Scholar] [CrossRef]
  23. Finlay, C.C.; Kloss, C.; Olsen, N.; Hammer, M.D.; Tøffner-Clausen, L.; Grayver, A.; Kuvshinov, A. The CHAOS-7 geomagnetic field model and observed changes in the South Atlantic Anomaly. Earth Planets Space 2020, 156, 72. [Google Scholar] [CrossRef] [PubMed]
  24. Zhang, H.; Xu, H.R.; Peng, G.S.; Qian, Y.D.; Zhang, X.X.; Yang, G.L.; Shen, C.; Li, Z.; Yang, J.W.; Wang, Z.Q.; et al. A prediction model of relativistic electrons at geostationary orbit using the EMD-LSTM network and geomagnetic indices. Space Weather 2022, 20, e2022SW003126. [Google Scholar] [CrossRef]
  25. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005, 18, 5–6. [Google Scholar] [CrossRef] [PubMed]
  26. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. Lond. A Mat. Proc. Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  27. Cai, J.; Dai, X.; Hong, L.; Gao, Z.; Qiu, Z. An air quality prediction model based on a noise reduction self-coding deep network. Math. Probl. Eng. 2020, 2020, 3507197. [Google Scholar] [CrossRef]
  28. Yin, H.; Jin, D.; Gu, Y.H.; Park, C.J.; Han, S.K.; Yoo, S.J. STL-ATTLSTM: Vegetable Price Forecasting Using STL and Attention Mechanism-Based LSTM. Agriculture 2020, 10, 612. [Google Scholar] [CrossRef]
  29. Zhang, K. A novel geomagnetic satellite constellation: Science and applications. Earth Planet. Phys. 2023, 7, 4–21. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.