1. Introduction
One of the worst natural disasters on the planet, earthquakes frequently result in significant loss for the human race. As a result, it is imperative to foresee when and where they will occur, albeit doing so is difficult because of their inherent randomness [
1].
The frequently used earthquake prediction methods nowadays can be categorized into four groups, according to [
2]. The first two strategies, namely 1) mathematical tools like FDL [
3] and 2) precursor-based techniques that retrieve geographic features like seismic anomalies [
4], cloud images [
5], and animal behavior [
6], were well-liked when there was a dearth of earthquake data. Then, as more and more earthquakes produced larger datasets, machine learning techniques entered the picture. Maria Moustra et al. [
7] used an artificial neural network, one of them, to predict an earthquake in 2011. The model’s accuracy was just about 70%, which was not very good then. It was determined from the research that fewer attributes in the dataset, as well as the class imbalance, led to unsatisfying results.
Finally, deep learning techniques have recently been used to forecast earthquakes. A deep neural network (DNN) model that assessed earthquake magnitudes due to occur within the next seven days using a newly introduced parameter Fault Density (based on the concept of spatial effect) was tested by Mohsen Yousefzadeh et al. [
8]. DNN performed better than other machine learning models, with a test accuracy of about 79% on magnitudes greater than 8. To detect the epicenters of earthquakes that occurred in the past two months, Luana Ruiz et al. [
9] developed a network called the Graph Convolutional Recurrent Neural Network (GCRNN) by combining the properties of convolutional neural networks (CNN) and recurrent neural networks. GCRNN used convolutional filters to maintain the independence of the trained parameter count from the input time sequences. For both the 30-second and 60-second waves, GCRNN achieved an accuracy of about 33% with fewer characteristics being taken into account. In addition, Q. Wang et al. [
2] used a long short-term memory (LSTM) variation of the recurrent neural network to learn the temporal-spatial dependencies of seismic signals and, consequently, forecast earthquakes. Even though they only used seismic data from the previous year as input, the accuracy was about 75%. A CNN-LSTM hybrid network was also suggested by Parisa Kavianpour et al. [
10] to learn the spatial-temporal dependencies of earthquake signals. They used the total number of earthquakes that occurred in a single month between 1966 and 2021 as their input to evaluate the model’s effectiveness in nine regions of the Chinese mainland. They also included other earthquake-related factors, such as latitude and longitude. They compared the results of several models, including MLP and SVM, and found that CNN-LSTM scored the highest. In [
11], the Author(s) focuses on predicting an earthquake in the next 30 seconds. Even though a long-term prediction’s outcome is likely erroneous (not accurate to a single day), 30 seconds allows individuals enough time to react and prevent an unthinkable tragedy. They, finally, deploy different time series models (vanilla RNN, LSTM, and Bi-LSTM) compared with one another in the same study question.
More recent research studies outline alternative neural network technologies. In recent years, neural networks, including LSTM and convolution neural networks, have been widely used in the research of time series and magnitude prediction [
7,
12]. As mentioned earlier, using LSTM networks to record spatiotemporal correlations among earthquakes in various places, the authors of [
3] developed a novel method for earthquake prediction. Their simulation findings show that their strategy performs better than conventional methods. It has been demonstrated that earthquake prediction accuracy can be increased using seismicity indicators as inputs in machine learning classifiers [
13]. Results from using this methodology in four Chilean cities show that reliable forecasts may be made by thoroughly examining how specific parameters should be configured. [
14] describes a proposed methodology trained and tested for the Hindu Kush, Chile, and Southern California regions. It is based on the computation of seismic indicators and GP-AdaBoost classification. Compared to earlier studies, the derived prediction results for these regions show improvement. [
15] explores how artificial neural networks (ANNs) can be used to predict earthquakes. The results from applying ANNs to Chile and the Iberian peninsula are presented, and the findings are compared with those of other well-known classifiers. Adding a new set of inputs improved all classifiers, but according to the conclusion, the ANN produced the best results of any classifier.
Seven different datasets from three areas have been subjected to a methodology for identifying earthquake precursors using clustering, grouping, building a precursor tree, pattern extraction, and pattern selection [
16]. Results compare favorably to the previous edition regarding all measured quality metrics. The authors propose that this method could be improved and used to predict earthquakes in other places with various geophysical characteristics.
The authors of [
17] employ machine learning techniques to identify signals in a correlation time series that predict future significant earthquakes. Decision thresholds, receiver operating characteristic (ROC) techniques, and Shannon information entropy are used to assess overall quality. They anticipate that the deep learning methodology would be more all-encompassing than earlier techniques and won’t require guesswork upfront about whether patterns are significant. Finally, their findings in [
18] demonstrate that the LSTM technique only provides a rough estimate of earthquake magnitude, whereas the random forest method performs best in categorizing major earthquake occurrences. The Author(s) conclude that information from small earthquakes can be used to predict larger earthquakes in the future. Machine learning offers a potential way to enhance earthquake prediction.
1.1. Research Contribution
To technically elaborate and materialize this claim, a study was conducted by utilizing (four) seismic features ((°N), (°E), (km), (Mag.)) based on seismic catalogs from the National Observatory of Athens (NOA), Greece containing the full list (i.e., small and large magnitude occurrences) of past earthquake events and for a specific year frame. This input was granted to construct a solicit time-scaled "sliding-window" (SW) mathematical technique that, if co-deployed with advanced Machine-learning and Game-theoretic algorithms, would significantly impact increasing the future precision accuracy of the earthquake predictability. Additionally, the capability of having architectured a synthesized seismic predictability framework that is capable of endorsing Machine Learning (both (semi)supervised and unsupervised), Game-theory solving models (i.e., OpenAI Gym, which is a toolkit for developing and comparing reinforcement learning (RL) algorithms), as well as the adaptation of the SW model in seismic short and long term casting was explored. The research investigation emphasized two crucial questions:
Will there be a strong event (M ≥ 6.0, 7.0, or 8.0) forecasted in the next year among the specific studied geographical region?
Can we obtain the scientific ability to predict the nearly exact 4-tuple ((°N), (°E), (km), (Mag.)) output of such future major event, as well as the (implicit) almost exact time frame of its occurrence?
2. Proof of Methods
Greece is a natural seismology laboratory because it has the highest seismicity in Europe and statistically produces an earthquake of at least M6.0 almost every year [
19]–
21]. The short repeat time in the area also makes it possible to study changes in regional seismicity rates over "earthquake cycles." Here, we provide the outcomes of the novel earthquake nowcasting approach (EN) developed by Rundle, Turcotte, and colleagues [
22]. With this approach, we counted the number of tiny EQs since the most recent major EQ to determine the region’s current hazard level. The term "natural time," which was coined by Varotsos et al. [
23,
24,
25,
2627], refers to event counting as a unit of "time" as opposed to clock time. According to Rundle et al. [
22], applying natural time to EQ seismicity offers the following two benefits: In addition, when computing nowcasts, the concept of natural time—counts of small EQs—is used as a measure of the accumulation of stress and strain between large EQs in a defined geographic area. First, it is not necessary to declutter the aftershocks because natural time is uniformly valid when aftershocks dominate, when background seismicity dominates, and when both contribute.
In other words, the use of natural time is the foundation of nowcast. As previously mentioned, there are two benefits to using natural time: first, there is no need to separate the aftershocks from the background seismicity; second, only the natural interevent count statistics are used, as opposed to the seismicity rate, which also takes into account conventional (clock) time. Instead of focusing on recurrent events on particular faults, the nowcasting method defines an "EQ cycle" as the recurring large EQs in a vast seismically active region composed of numerous active faults. Following Pasari [
28] (see also Pasari et al. [
29]), we may say that although the concept of "EQ cycle" has been used in numerous earlier seismological investigations [
30],
31]
32], the idea of "natural time" is unique in its properties.
The estimation of seismic risk in large cities around the world [
33], the study of induced seismicity [
34], the study of temporal clustering of global EQs [
35], the clarification of the role of small EQ bursts in the dynamics associated with large EQs [
36], the understanding of the complex dynamics of EQ faults [
37], the identification of the current state of the "EQ cycle" [
38,
39,
40], the nowcasting of avalanches [
41]. The Olami-Fe Here, using the earthquake potential, we examined the greatest events in Greece between January 1 and February 6, 2023, with MW(USGS) 6 (see
Figure 1 and Table 1), utilizing the earthquake potential score (EPS) (see below).
Natural time analysis (NTA), which was discussed in [
27] and more recently in [
42], in general, displays the dynamical evolution of a complex system and pinpoints when it hits a critical stage. As a result, NTA can play a significant role in foreseeing approaching catastrophic occurrences, such as the emergence of massive EQs. In this regard, it was used in situations of EQ in Greece [
23,
25,
26,
43,
44,
45,
46,
47,
48,
49], the USA [
50,
51], Mexico [
70,
71], the Eastern Mediterranean [
55,
56], and globally [
57,
58,
59]. We observe that NTA permits the insertion of an entropy,
S, which is a dynamic entropy [
24] that demonstrates positivity, concavity, and experimental stability for Lesche [
64,
65]. Recently, the research of EQs in Japan [
67,
68,
69,
70] and Mexico [
71,
72] has used complexity metrics [
66] based on natural temporal entropy and
S itself, with encouraging results [
73]. In particular, two quantities—which are discussed below—have lately been stated in the Preface of [
43] and have emerged through natural time analysis to be crucial in determining whether and when the critical moment (mainshock, the new phase) is approaching.
First, let’s talk about seismicity’s order parameter
. Its value
indicates when the system reaches the critical stage, and its variations’ minimum indicates when the Seismic Electric Signals (SES) [
74,
75,
76,
77,
78,
79] activities start [
80]. The SES amplitude is crucial because it allows for estimating the impending mainshock’s magnitude. The epicentral area is determined using the station’s SES selectivity map, which records the pertinent SES (using this methodology, a successful prediction for a
that occurred on June 8, 2008, in the Andravida area of Greece was made [
44,
46,
81]).
Second, the entropy change,
, under time reversal: its value, when minimized a few months in advance, denotes the beginning of precursory phenomena; as for its fluctuations (when the minimum of
appears), they show a clear increase, denoting the time when the
preparation begins, as explained by the physical model that served as the basis for the SES research [
79].
The Tohoku on March 11, 2011, which was the largest event ever recorded in Japan, and the Chiapas on September 7, 2017, which was the largest EQ in Mexico in more than a century, were used to study precursory phenomena before the two subsequent major EQs.
3. Datasets and Feature Engineering
The seismic input catalog used in this research study was provided by the Greece National Observatory of Athens (NOA), (
https://www.gein.noa.gr/services/cat.html, last accessed on 1 January 2024) and includes earthquake events with a magnitude greater than 2.0 in the Greek geographic region from 1950 to 2024. Using feature engineering, several statistical principles were used to create the seismic activity parameters for a specific experimentation trial set; rather than using the original seismic catalog, these parameters obtained from it were utilized as the input features for earthquake prediction.
Numerous seismic properties obtained from earthquake catalogs have been shown in previous studies to be useful in earthquake prediction. The number and maximum/mean magnitude of previous earthquakes, seismic energy release, magnitude deficit, seismic rate fluctuations, and the amount of time since the last significant earthquake are some of these characteristics. Other elements include an ’a’ and a ’b’ value in the Gutenberg-Richter (GR) law. Typical related research work seismic features also include the probability of an earthquake occurring, the divergence from the Gutenberg-Richter rule, and the standard deviation of the estimated b value. The equation of the GR property, which describes the fractal and/or power law magnitude distribution of earthquakes in the defined region and in the defined time interval, is given by the following formula: , is a very nice illustration scenario where we applied, as part of the scope of our ANN framework, a quite accurate Machine-learning logistic regression library to predict the next mathematical future value (of the GR rule), with strict preconditional property of holding at least 3 decimal digits of numerical accuracy in the fractional part of the predicted value.
6. Discussion and Future Work
The scientific luxury of obtaining a network of seismographic instruments and next-generation Internet-of-Things microseismic sensors that can aggregate in real time a massive amount of earthquake data from beneath the earth’s surface would be an add-on for any research attempts like the one above that aims to leverage state-of-the-art Artificial Intelligence (AI) to predict shock events in the short and long term. At first glance, we can increase the number of layers in each network to enable each network to learn more consecutive properties from the data. Then, improving the hyperparameters is another choice.
As Future Work for our research project, we will deploy explicit next-generation (hybrid) Transformer networks, focusing on exploiting eXplainable AI (XAI) techniques and building Large Language Models (LLMs). The latter are known to possess extreme potency in language detection and language generation. The research field of predicting earthquakes via generative AI (GPT-4) would be quite interesting to construct and experiment on.
Ultimately, the investigation showed that earthquake prediction remains a difficult issue. Various deep learning techniques can be applied separately or in combination to determine the best approach for these time series forecasting problems.