You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

12 July 2023

Earthquake Detection Using Stacked Normalized Recurrent Neural Network (SNRNN)

,
,
,
and
1
College of Instrumentation & Electrical Engineering, Jilin University, Changchun 130061, China
2
College of Geoexploration Science & Technology, Jilin University, Changchun 130026, China
3
Institute of Integrated Information for Mineral Resources Prediction, Jilin University, Changchun 130026, China
4
Department of Artificial Intelligence, University of Management and Technology, C-II Block C 2 Phase 1 Johar Town, Lahore 54770, Pakistan

Abstract

Earthquakes threaten people, homes, and infrastructure. Earthquake detection is a complex task because it does not show any specific pattern, unlike object detection from images. Convolutional neural networks have been widely used for earthquake detection but have problems like vanishing gradients, exploding, and parameter optimization. The ensemble learning approach combines multiple models, each of which attempts to compensate for the shortcomings of the others to enhance performance. This article proposes an ensemble learning model based on a stacked normalized recurrent neural network (SNRNN) for earthquake detection. The proposed model uses three recurrent neural network models (RNN, GRU, and LSTM) with batch normalization and layer normalization. After preprocessing the waveform data, the RNN, GRU, and LSTM extract the feature map sequentially. Batch normalization and layer normalization methods take place in mini-batches and input layers for stable and faster training of the model and improving its performance. We trained and tested the proposed model on 6574 events from 2000 to 2018 (18 years) in Turkey, a highly targeted region. The SNRNN achieves RMSE values of 3.16 and 3.24 for magnitude and depth detection. The SNRNN model outperforms the three baseline models, as seen by their low RMSE values.

1. Introduction

Natural disasters like earthquakes can be very damaging. They also trigger tsunamis, earthquakes, and fires, all of which can inflict significant damage and risk human and property safety. Earthquakes can have different effects on humans depending on the severity of the quake and the surrounding environment. Large earthquakes of six or above on the Richter scale are the scariest. Earthquakes have killed nearly 750,000 people worldwide since 1998. Many of these fatalities happened in nations hit hard by natural disasters, like Japan, China, and Indonesia. Automatic early earthquake detection using the data from seismic station sensors has emerged as an important area of research in recent years for emergency response []. The primary task in earthquake detection is to estimate an incoming event’s magnitude, depth, and location information. Emergency response and early earthquake warning systems (EEWSs) should be able to issue a warning or spread event information on the targeted area within seconds of the detection of seismic waves without the need for human intervention []. The number of seismic networks and monitoring sensors has steadily increased in recent years, and the continuous growth of seismic records calls for new processing algorithms that assist in solving problems in seismology. Several seismic networks with multiple stations, like the Southern California Seismic Network and the Turkey Network, have provided catalogs of earthquake events in the last decade. Seismology benefits significantly from a practical and comprehensive analysis of these catalogs. Using computational approaches like machine learning and deep learning is more promising for automatically predicting earthquakes [].
Many different machine learning models, including decision trees (DTs), support vector machines (SVMs), and k-nearest neighbors (k-NNs), have been applied to the problem of earthquake detection. Ref. [] used the SVM model for an on-site early warning system. Ref. [] used k-means clustering techniques for earthquake magnitude detection from global earthquake catalogs. Ref. [] used k-NN, SVM, DT, and random forest (RF) for earthquake detection and found that RF outperformed the other models. Several factors like feature selection method, dataset size, and class imbalance distribution affect the performance of machine learning models. To overcome the machine learning issues, deep learning has been introduced.
Deep learning is a subfield of machine learning. Like artificial neural networks, deep learning-based models use multiple layers to input and process data and output the results. Deep learning models are considered good when the dataset is complex and large because they employ several layers to deal with such data []. A convolutional neural network (CNN) was used to classify earthquake events into macro, micro, and artificial earthquakes. Ref. [] presented a CNN and a graph convolutional network (GCN) based on deep learning models to detect earthquakes from multiple stations. Ref. [] proposed a method that uses a CNN and graph partitioning algorithms to detect events with extremely low signal-to-noise ratios. Recurrent neural networks (RNNs) have a different architecture than CNNs. RNNs use various gates for different operations instead of convolutional and pooling operations. Ref. [] used an RNN-based DeepShake model, and Ref. [] used a CNN model to predict earthquake shaking intensity from ground motion observation. Ref. [] used LSTM and BLSTM for magnitude detection. Ref. [] proposed a transformer-based TEAM method that issues accurate and timely warnings. Model interpretation, hyperparameter tuning, and GPU-based special devices are significant issues with deep learning models. Past research studies propose three approaches to overcome the problems of deep learning and machine learning methods: layer or batch normalization, the attention mechanism in deep learning architectures, and ensemble learning.
Multiple or ensemble learning models have grown in computational intelligence in the last couple of decades. Different models (i.e., base classifiers) are organized sequentially or parallel to the design of a more powerful ensemble model. An ensemble model may consist of only machine learning models, deep learning models, or hybrid models. Ensemble-based studies report that ensemble models outperform the individual models of either machine learning or deep learning [,]. In machine learning-based ensemble studies, Ref. [] proposed a machine learning-based ensemble method that combines SVM, k-NN, DT, and RF to design an ensemble model that can effectively detect earthquakes. Ref. [] ensembled four machine learning models, AdaBoost, XGBoost, DT, and LightXGBoost, in a stack using multiple settings. Another stacked-based ensembled model was proposed in [] to ensemble three models, bagging, AdaBoost, and stacking, for earthquake causality prediction. Ref. [] ensembled a CNN and LSTM and proposed MagNet for earthquake magnitude estimation. Ref. [] proposed an LSTM-GRU-based ensemble method that outperformed the LSTM and GRU on two datasets. Ref. [] proposed a hybrid model and used SVM and three ANN models for earthquake prediction.
At different stations, the same earthquake event may be recorded differently. Applying normalization techniques like batch normalization and layer normalization can improve classification performance due to their normalization effects on raw seismic data. Within each layer, the feature maps are first standardized using the mean and standard deviation, and subsequently, these maps are transformed into standardized values using either a shift factor or a scale factor. Several recent studies showed that using layer normalization and batch normalization layers in deep learning models improves their computational and detection performance. Ref. [] used batch normalization techniques in a GCNN model for earthquake source characterization. Ref. [] showed that the training of a deep learning model can be enhanced using batch normalization. When a model uses layer normalization, all the cells receive the same feature distribution independently for each batch input. Ref. [] used LSTM with an attention layer that effectively detected large-magnitude earthquake events and outperformed the LSTM and ANN models. Ref. [] proposed an attention-based fully connected CNN model, and Ref. [] used the attention layer in the LSTM model for earthquake detection. Ref. [] applied batch normalization and attention layers in a GCNN model for magnitude detection.
Taking advantage of both normalization and ensemble technique methods, in this study, we propose an RNN-based stacked normalized recurrent neural network (SNRNN) model that ensembles three recurrent neural network models in a stack: GRU, LSTM, and SimpleRNN. A GRU is an RNN model that uses less memory and is faster when the data have longer sequences. The LSTM and GRU used their gates to handle the gradient descent problem. The experimental results and their analysis show that the proposed model SNRNN is effective in earthquake detection tasks. The contribution of this study is as follows:
  • We propose a deep learning model: a stacked normalized recurrent neural network (SNRNN), which is an ensemble model of three SimpleRNN, GRU, and LSTM models with normalization layers;
  • We evaluate the performance of three individual RNN models and stacked RNN models using layer normalization and batch normalization methods;
  • We compare the proposed SNRNN model to classic recurrent models and find that it outperformed all other models, achieving the lowest RMSE values.

3. Stacked Normalized Recurrent Neural Network (SNRNN) Architecture

This section discusses, in detail, layer by layer, the proposed stacked normalized recurrent neural network model. The architecture of the SNRNN is shown in Figure 2. The first layer is the data layer, which receives the dataset and input to the proposed model. The second layer performs data preprocessing operations like removing irrelevant event information and splitting the dataset into training and testing subsets. The following three layers ensemble the SimpleRNN, GRU, and LSTM in a stack: The SimpleRNN model receives the preprocessed data and is trained on it to extract valuable features in the third layer. The fourth layer uses the GRU model to refine the features set output using the SimpleRNN. In the fifth layer, LSTM performs the final feature selection using its memory cells. The last layer is the output layer, which finally detects the earthquake event’s magnitude and depth.
Figure 2. The five-layered architecture of the proposed stacked normalized recurrent neural network (SNRNN).

3.1. Layer 1: Input Data

The dataset consists of fourteen attributes: event number, date and time, latitude, longitude, station number and name, magnitude, type, depth, source number, source description, reference, etc. We removed the irrelevant information and kept only the relevant information. As a result, we obtained five input attributes to feed into the proposed model: time, latitude, longitude, magnitude, magnitude type, and depth. To provide greater insight, the architecture of our model’s inputs and outputs is depicted in Figure 3. Following feature selection, the dataset was normalized using a min–max scaler. This scaler transforms the features by transforming their values to the interval [0, 1]. In this way, the original distribution shape was maintained. We split the dataset into training (75%) and testing (25%) subsets to train and test the model.
Figure 3. The framework of our proposed model’s inputs and outputs.

3.2. Layer 2: Recurrent Neural Network Model (RNN)

After preprocessing the dataset, it is given as input to the first SimpleRNN model. The normalization layer first normalizes feedback, and the normalized output is provided as input to the RNN layer. A dense layer is applied to the outcome of the RNN layer to produce the outcome. Layer normalization gives all neurons in a layer the same feature distribution for a given input. The need for batches is removed when normalization is performed across all features and for each piece of input to a given layer []. That is why layer normalization works so well with sequence models like RNNs. Finally, the dense layer works on the RNN layer to produce the output.

3.3. Layer 3: Gated Recurrent Unit Model (GRU)

The output of the RNN model is given to the next layer of the proposed architecture. At this layer, the GRU model is implemented. RNN faces exploding and vanishing gradient problems during backpropagation. GRU overcomes these problems using its gates. GRU is an RNN model with less memory and is faster than LSTM when the data have longer sequences. GRU is slower in training than ANN because GRU cell has more parameters than ANN cell. To address the gradient descent problem, GRUs employ the update gate and reset gate. Reset gates can recall previous knowledge learned and control what information is sent to the output. The model uses the update gate to determine how much information from the past should be transmitted to the future []. As the model can decide to replicate all the knowledge from the past, it eliminates the risk of the vanishing gradient problem, making this a very powerful model. As a first step, we used the following equation to get the updated gate state zt for the time step t:
z t = σ ( W z x t + U z h t 1 )
The weight W(z) of the network and the unit is applied to the input xt when it is entered into the network. The same can be said for ht−1, which, after being multiplied by its weight U(z), stores information about the time before it (t − 1). The two sets of findings are combined and summed up, and then a sigmoid activation function is used to squeeze the total value into the range of 0 to 1. The purpose of the forget gate in the model is to determine how much of the information from the past should be ignored. The following equation was used to calculate the required information. This same formula describes the update gate, but the gate is used in a different way and with varying weights:
r t = σ ( W r x t + U r h t 1 )

3.4. Layer 4: Long Short-Term Memory Model (LSTM)

The long short-term memory (LSTM) model is the next layer of the proposed model. The LSTM is built up of many individual cells or units []. The hidden layer has many cells. The three gates of an LSTM cell are forgotten, input, and output. The forget gate keeps the cell state in memory, while the input and output gates handle incoming and outgoing data []. By utilizing these gates, LSTM can fix the issue of vanishing gradients. The following equations show the list of operations on the input sequence for the LSMT model:
i t = σ ( x t U i + h t 1   W i + b i   )
f t = σ ( x t U f + h t 1 W f + b f )
o t = σ ( x t U o + h t 1 W o + b o )
q t = t a n h x t U q + h t 1 W q + b q  
p t = f t p t 1 + i t q t
h t = o t t a n h p t
r t = σ W r x t + U r h t 1
i t is the output of the input gate, o t is the result of the output gate, and f t is the output of the forget gate. i t , o t , and f t are activated by the sigmoid function σ . These three results are the ensemble of input x t   and the preceding hidden state h t 1 and the biased value b . A t a n h nonlinearity function is applied over the input x t and the previous hidden state h t 1 to generate a temporary result q t . The current step t is then used to calculate the hidden step. Then, to obtain an updated version of the history p t , q t is combined with history p t 1 using the input gate i t and forgot gate f t in that order. In the end, the output gate o t will make use of the updated history p t to determine the final hidden state, which will serve as the output to the softmax layer, producing the outcome.

3.5. Layer 5: Output Layer

The last layer in the proposed architecture is the output layer which is a fully connected softmax layer. Here, the output of the LSTM model is the input of this layer. The outcome is the final detection made by the SNRNN model about the magnitude and depth of the earthquake event.

3.6. Layer Normalization

Layer normalization is a normalization technique that normalizes the inputs to a layer across the feature dimension, which can help improve the stability and performance of deep learning models. In our proposed model, we used layer normalization as a critical component to enhance the stacked normalized recurrent neural network (SNRNN) performance for earthquake detection. While batch normalization is another commonly used normalization technique, we chose layer normalization in our model due to its ability to normalize the inputs across different samples in a batch. This can be particularly useful when working with earthquake data.

4. Turkish Seismic Earthquake Dataset

In this study, we applied our methods to a seismic dataset gathered from Turkey. Turkey is well recognized for its active seismicity in a seismically active region. Regarding the number of people killed by earthquakes, Turkey is rated third in the globe, while in terms of the number of people affected, it is placed ninth. On average, it has at least one earthquake with a magnitude of 5 to 6 every year. Therefore, it is vital to analyze the earthquake data in this region and design an EEWS to detect an incoming earthquake accurately. The dataset was collected from the Disaster and Emergency Management Authority (AFAD) (https://deprem.afad.gov.tr/depremkatalogu?lang=en, accessed on 10 August 2022) catalog with the geographic parameters of latitude [35.67° to 42.38°] and longitude [25.85° to 45.14°]. The dataset consists of 6574 seismic events from 2000 to 2018 (18 years). The events were collected from different locations and stations in Turkey.
There are several features in the USGS dataset. We selected some of the most important and connected features and removed the others. The location information of all the events in the dataset is given in Figure 4. The locations of the high-magnitude (i.e., >7) seismic events are indicated on the map by red dots. The locations of the low-magnitude (i.e., <3) events are represented by white spots on the map. Green- and aqua-colored dots dominate the data, with a magnitude range of 4 to 6. Turkey’s border regions are the most likely to experience earthquakes due to their proximity to the fault lines. Figure 5a shows a histogram showing the magnitude distribution in the dataset. The magnitude value ranges from 4.0 to 7.9, and the mean magnitude is 4.46. The dataset is dominated by events of magnitudes between 4.0 and 5.0. Figure 5b shows a histogram of the depth distribution of earthquake events. The depth ranges from 0.0 to 212 km, and the mean depth is 25 km. The majority of the events have a depth of less than 15 km.
Figure 4. Dataset seismological event locations. High-magnitude events are red, while low-magnitude events are aqua-colored.
Figure 5. (a) Magnitude distribution of events; (b) depth distribution of events. (a) shows a histogram of magnitude distribution. (b) shows a histogram of the depth distribution of earthquake events.
In summary, Table 2 provides important statistical information on the seismic dataset related to earthquake events in Turkey. This information can improve our understanding of the region’s seismic activity and help us to develop more effective earthquake awareness and response strategies.
Table 2. Summary of the Turkey dataset.
Figure 6 presents a risk map of the earthquake catalog of the Turkey region used in this study. A risk map is essential for predicting and mitigating the impacts of seismic events in the region. The map is constructed of peak accelerations, which are measures of the maximum shaking expected at a given location during an earthquake. The peak accelerations were calculated based on the characteristics of the earthquake catalog data, such as the magnitude, location, and depth of each earthquake event. By analyzing the earthquake catalog data and constructing the risk map, we can better understand the potential impact of future earthquakes in the region and help inform earthquake mitigation strategies. The risk map visually represents the distribution of seismic events in different regions of Turkey, with areas of higher hazard indicated in red and orange colors and areas of lower hazard in green and yellow.
Figure 6. Risk map of earthquakes in Turkey.

5. Experiments and Discussion

In this section about the experiment and discussion, we first discuss the experimental settings and optimized parameters of the SimpleRNN, GRU, and LSTM. We discuss and analyze the individual models’ results with and without layer normalization and batch normalization methods. In the end, we discuss the results of our proposed SNRNN model.
All of the experiments were carried out, as a rule, on an Intel Core i7-7700 central processing unit operating at 3.60 gigahertz with 16 gigabytes of memory, an NVIDIA GeForce GTX 1080 graphics card, Windows 10, and Keras using the CUDA toolkit. The experiments were designed to compare the performance of three RNN models (SimpleRNN, GRU, and LSTM) with the proposed SNRNN model. The hyperparameters of these models, like batch size, dropout, and epochs, were tuned before the final experiments. These last parameters are shown in Table 3. For all the models, we used the RMSE performance measure for network parameter optimization. RMSE is the improved form of mean square error (MSE) and can be calculated using the following equation:
RMSE = i = 1 n y i y ^ i 2 n
Table 3. The hyperparameters of RNN, GRU, and LSTM models.
It is important to note that the unit scale of the observed data is preserved by the square root mistake. The y ^ i in the equation represents the predicted values based on the model fit, and y i is the expected value. n is the size of the values in the dataset. In addition to RMSE’s use in assessing quality, a random 75/25 split created a training and testing set from the whole dataset. From the training set, 25% of the data was used for validation during the training.
The first type of experiment compared the performance of the three RNN models: SimpleRNN, GRU, and LSTM. The RMSE values obtained by these models are shown in Figure 7. LSTM is usually considered suitable for time-series data. LSTM outperforms the GRU and SimpleRNN models in predicting magnitude and depth. LSTM achieves 4.35 and 4.84 RMSE values for magnitude and depth. The SimpleRNN does not perform well because of its vanishing gradient problems. The GRU and LSTM have gates to deal with vanishing gradient problems. Therefore, the GRU performed better than the SimpleRNN.
Figure 7. Magnitude and depth estimation error using SimpleRNN, GRU, and LSTM.
For the second type of experiment, we used the normalized models of the SimpleRNN, GRU, and LSTM to compare their performance. We added a batch normalization layer or the normalization layer to normalize the input and the mini-batches to train the model faster and increase performance. Again, LSTM+LN and LSTM+BN outperform the others in predicting magnitude and depth. LSTM+LN performs the best among all the models. It achieves 4.14 and 4.65 RMSE values for magnitude and depth, as shown in Figure 8. The models with batch normalization perform a little better than the standard models. When adding layer normalization to these models, all the models perform significantly better. Normalized GRU decreases its performance more than simple GRU. For the third type of experiment, we used the normalized models of the SimpleRNN, GRU, and LSTM to compare their performance. After analyzing the effects of the normalization techniques on the RNN models, we ensembled the SimpleRNN, GRU, and LSTM models in different settings to make a powerful and efficient stacked model.
Figure 8. Magnitude and depth estimation error using SimpleRNN+BN, SimpleRNN+LN, GRU+BN, GRU+LN, LSTM+BN, and LSTM+LN.
The output of the first model was given as input to the second model, and so on. The results of the stacked models and individual models are shown in Figure 9. We conclude a significant decrease in the RMSE values of the stacked models compared to the individual models. Although the performance of stacking two models is better, stacking three models obtains the lowest RMSE values. From the results, we can easily conclude that ensemble methods are suitable for the task of magnitude and depth detection by overcoming the weaknesses of the base models.
Figure 9. Magnitude and depth estimation error using SimpleRNN, GRU, LSTM, SimpleRNN+GRU, SimpleRNN+LSTM, GRU+LSTM, and SimpleRNN+GRU+LSTM.
The stacked RNN achieves the lowest RMSE values, 3.27 and 3.65, for magnitude and depth detection. Again, the GRU with either LSTM or the SimpleRNN does not show a high performance. In Figure 8 and Figure 9, we see that the stacked-based ensemble model significantly outperforms the individual models, and the models with the RNN models with normalization layers outperform the models with batch normalization layers. Finally, our proposed model, the stacked normalized RNN (SNRNN), is an ensemble model of the SimpleRNN, GRU, and LSTM models where each model has a normalization layer. Normalization is applied to the model input in the stack. The normalization layer normalizes each input in the batch independently across all the features. Batch normalization depends on the batch size, while normalization is independent of the batch size. The RMSE values show that the proposed model SNRNN (on top) outperforms all the models, where normalization helps to achieve 3.16 and 3.24 RMSE values for magnitude and depth detection. Our proposed SNRNN model takes advantage of both the ensemble learning and normalization techniques. All the results are summarized in Table 4.
Table 4. Results of the baseline models and the proposed SNRNN model.

6. Conclusions and Future Work

Natural disasters like earthquakes can be very damaging. Automatic early earthquake detection using the data from seismic station sensors has emerged as an important area of research in recent years for emergency response. This article proposes a stacked-base ensemble method SNRNN for earthquake detection. This model, an ensemble of RNN, GRU, and LSTM base models, incorporating layer normalization techniques, can successfully detect the depth and magnitude of an earthquake event. After the data were preprocessed, the RNN, GRU, and LSTM sequentially extracted the feature map to make the final detection. Batch and layer normalization were utilized to achieve more consistent and faster training. Layer normalization was used to normalize features independently of the batch size. Therefore, layer normalization is more effective than batch normalization when using RNN-based models. We applied RNN, GRU, and LSTM models independently with both normalization methods. But the performance was lower than the proposed SNRNN model. After that, these models were ensembled to design a powerful model. We tested the proposed model on 6574 earthquake events from 2000 to 2018 in Turkey. The proposed model achieves 3.16 and 3.24 RMSE values for magnitude and depth detection. The RNN models outperform layer normalization over batch normalization. We also conclude that the ensemble model outperforms the individual model.
Researchers, seismologists, and the metrological department will all benefit from this innovative SNRNN by learning more about the potential of the ensemble method and how to apply many data mining techniques at once. In the future, we plan to apply similar ensemble techniques for earthquake detection with homogenous deep learning models. Further, the proposed model can be used for ground motion intensity detection.

Author Contributions

Conceptualization, Y.J. and Y.W.; Formal analysis, M.A.B., M.P.A. and H.L.; Funding acquisition, Y.J. and Y.W.; Investigation, H.L.; Methodology, M.P.A.; Software, M.A.B.; Supervision, H.L.; Visualization, M.A.B.; Writing—original draft, M.A.B.; Writing—review and editing, M.P.A. All authors have read and agreed to the published version of the manuscript.

Funding

National Key R&D Plan: 2021YFC2901801.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Datset was collected from AFAD. The data is Publically available on AFAD website.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kim, G.; Ku, B.; Ahn, J.-K.; Ko, H. Graph Convolution Networks for Seismic Events Classification Using Raw Waveform Data from Multiple Stations. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  2. Münchmeyer, J.; Bindi, D.; Leser, U.; Tilmann, F. The transformer earthquake alerting model: A new versatile approach to earthquake early warning. Geophys. J. Int. 2020, 225, 646–656. [Google Scholar] [CrossRef]
  3. Bilal, M.A.; Ji, Y.; Wang, Y.; Akhter, M.P.; Yaqub, M. Early Earthquake Detection Using Batch Normalization Graph Convolutional Neural Network (BNGCNN). Appl. Sci. 2022, 12, 7548. [Google Scholar] [CrossRef]
  4. Song, J.; Zhu, J.; Wang, Y.; Li, S. On-site alert-level earthquake early warning using machine-learning-based prediction equations. Geophys. J. Int. 2022, 231, 786–800. [Google Scholar] [CrossRef]
  5. Yuan, R. An improved K-means clustering algorithm for global earthquake catalogs and earthquake magnitude prediction. J. Seism. 2021, 25, 1005–1020. [Google Scholar] [CrossRef]
  6. Ku, B.; Kim, G.; Ahn, J.-K.; Lee, J.; Ko, H. Attention-Based Convolutional Neural Network for Earthquake Event Classification. IEEE Geosci. Remote Sens. Lett. 2020, 18, 2057–2061. [Google Scholar] [CrossRef]
  7. Yano, K.; Shiina, T.; Kurata, S.; Kato, A.; Komaki, F.; Sakai, S.; Hirata, N. Graph-Partitioning Based Convolutional Neural Network for Earthquake Detection Using a Seismic Array. J. Geophys. Res. Solid Earth 2021, 126, e2020JB020269. [Google Scholar] [CrossRef]
  8. Datta, A.; Wu, D.J.; Zhu, W.; Cai, M.; Ellsworth, W.L. DeepShake: Shaking Intensity Prediction Using Deep Spatiotemporal RNNs for Earthquake Early Warning. Seism. Res. Lett. 2022, 93, 1636–1649. [Google Scholar] [CrossRef]
  9. Joshi, A.; Vishnu, C.; Mohan, C.K. Early detection of earthquake magnitude based on stacked ensemble model. J. Asian Earth Sci. X 2022, 8, 100122. [Google Scholar] [CrossRef]
  10. Cui, S.; Yin, Y.; Wang, D.; Li, Z.; Wang, Y. A stacking-based ensemble learning method for earthquake casualty prediction. Appl. Soft Comput. 2020, 101, 107038. [Google Scholar] [CrossRef]
  11. Mukherjee, S.; Gupta, P.; Sagar, P.; Varshney, N.; Chhetri, M. A Novel Ensemble Earthquake Prediction Method (EEPM) by Combining Parameters and Precursors. J. Sensors 2022, 2022, 5321530. [Google Scholar] [CrossRef]
  12. Berhich, A.; Belouadha, F.-Z.; Kabbaj, M.I. A location-dependent earthquake prediction using recurrent neural network algorithms. Soil Dyn. Earthq. Eng. 2022, 161, 107389. [Google Scholar] [CrossRef]
  13. Asim, K.M.; Idris, A.; Iqbal, T.; Martínez-Álvarez, F. Earthquake prediction model using support vector regressor and hybrid neural networks. PLoS ONE 2018, 13, e0199004. [Google Scholar] [CrossRef]
  14. Bilal, M.A.; Ji, Y.; Wang, Y.; Akhter, M.P.; Yaqub, M. An Early Warning System for Earthquake Prediction from Seismic Data Using Batch Normalized Graph Convolutional Neural Network with Attention Mechanism (BNGCNNATT). Sensors 2022, 22, 6482. [Google Scholar] [CrossRef]
  15. Kalayeh, M.M.; Shah, M. Training Faster by Separating Modes of Variation in Batch-Normalized Models. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1483–1500. [Google Scholar] [CrossRef] [PubMed]
  16. Berhich, A.; Belouadha, F.-Z.; Kabbaj, M.I. An attention-based LSTM network for large earthquake prediction. Soil Dyn. Earthq. Eng. 2023, 165, 107663. [Google Scholar] [CrossRef]
  17. Akhter, M.P.; Zheng, J.; Afzal, F.; Lin, H.; Riaz, S.; Mehmood, A. Supervised ensemble learning methods towards automatically filtering Urdu fake news within social media. PeerJ Comput. Sci. 2021, 7, e425. [Google Scholar] [CrossRef]
  18. Pham, K.; Kim, D.; Park, S.; Choi, H. Ensemble learning-based classification models for slope stability analysis. Catena 2020, 196, 104886. [Google Scholar] [CrossRef]
  19. Jozinović, D.; Lomax, A.; Štajduhar, I.; Michelini, A. Rapid prediction of earthquake ground shaking intensity using raw waveform data and a convolutional neural network. Geophys. J. Int. 2020, 222, 1379–1389. [Google Scholar] [CrossRef]
  20. Abebe, E.; Kebede, H.; Kevin, M.; Demissie, Z. Earthquakes magnitude prediction using deep learning for the Horn of Africa. Soil Dyn. Earthq. Eng. 2023, 170, 107913. [Google Scholar] [CrossRef]
  21. Mousavi, S.M.; Ellsworth, W.L.; Zhu, W.; Chuang, L.Y.; Beroza, G.C. Earthquake transformer—An attentive deep-learning model for simultaneous earthquake detection and phase picking. Nat. Commun. 2020, 11, 3952. [Google Scholar] [CrossRef] [PubMed]
  22. Santurkar, S.; Tsipras, D.; Ilyas, A.; Madry, A. How Does Batch Normalization Help Optimization? Adv. Neural Inf. Process. Syst. 2018, 31, 2483–2493. [Google Scholar]
  23. Huang, L.; Qin, J.; Zhou, Y.; Zhu, F.; Liu, L.; Shao, L. Normalization Techniques in Training DNNs: Methodology, Analysis and Application. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10173–10196. [Google Scholar] [CrossRef] [PubMed]
  24. Singh, S.; Krishnan, S. Filter Response Normalization Layer: Eliminating Batch Dependence in the Training of Deep Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11234–11243. [Google Scholar] [CrossRef]
  25. Xu, J.; Sun, X.; Zhang, Z.; Zhao, G.; Lin, J. Understanding and improving layer normalization. Adv. Neural Inf. Process. Syst. 2019, 32, 1–19. [Google Scholar]
  26. Elsayed, H.S.; Saad, O.M.; Soliman, M.S.; Chen, Y.; Youness, H.A. Attention-Based Fully Convolutional DenseNet for Earthquake Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–10. [Google Scholar] [CrossRef]
  27. Xiao, J. SVM and KNN ensemble learning for traffic incident detection. Phys. A Stat. Mech. Appl. 2018, 517, 29–35. [Google Scholar] [CrossRef]
  28. Ridzwan, N.S.M.; Yusoff, S.H.M. Machine learning for earthquake prediction: A review (2017–2021). Earth Sci. Inform. 2023, 16, 1133–1149. [Google Scholar] [CrossRef]
  29. Mousavi, S.M.; Beroza, G.C. A Machine-Learning Approach for Earthquake Magnitude Estimation. Geophys. Res. Lett. 2020, 47, e2019GL085976. [Google Scholar] [CrossRef]
  30. Al Banna, H.; Ghosh, T.; Al Nahian, J.; Abu Taher, K.; Kaiser, M.S.; Mahmud, M.; Hossain, M.S.; Andersson, K. Attention-Based Bi-Directional Long-Short Term Memory Network for Earthquake Prediction. IEEE Access 2021, 9, 56589–56603. [Google Scholar] [CrossRef]
  31. Ruiz, L.; Gama, F.; Ribeiro, A.R. Gated Graph Recurrent Neural Networks. IEEE Trans. Signal Process. 2020, 68, 6303–6318. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.