1. Introduction
Our research is focused on assessing the predictive capabilities of the equation of Dynamics of Self-Developing Natural Processes (DSDNP equations) and developing algorithms for its practical use. The origin of the first versions of this equation occurred when A. Malyshev (one of the authors of this work) studied plastic deformations that preceded and accompanied the eruptions of Bezymyannyi Volcano in 1981–1984 [
1]. These placative deformations were not accompanied by volcanic earthquakes. However, the analysis of changes in the volume of erupted material showed the presence of a direct (the more, the faster) avalanche-like development before the culmination of eruptions and a reverse (the less, the slower) avalanche-like development in the post-climactic eruptive process. Moreover, the absence of signs of an avalanche-like development indicated an upcoming calm (without climactic) eruption. These observations allowed A. Malyshev to successfully predict the directed blast of Bezymyannyi Volcano on 30 June 1985 [
1,
2,
3], as well as a number of eruptions (with or without paroxysm) in 1986 to 1987 [
1,
3].
The expressions ‘the more, the faster’ and ‘the less, the slower’ are first-order differential equations expressed in words. In these equations, the rate of change in the state of the system depends on its current state. At the same time, the fact of self-development of the system is essential. As a result of the analysis of the patterns of self-development of a wide range of natural processes, it was concluded that, in the case of Self-Developing Systems (SDS), the forces that change their state arise due to their own energy of movement of these systems [
1]:
Fx =
C |
Ex −
E0|
γ =
C |
mx(
x′)
2/2 −
mx(
x′
0)
2/2|
γ. Here,
mx,
Fx and
Ex are, respectively, the “measure of inertness,” the “force” and the “energy of motion” of the system with respect to parameter
x; C is a constant of proportionality; γ is an exponent of nonlinearity and
x′ and
x′
0 are the rates of change of the parameters in the current and stationary states, respectively. This conclusion was transformed into the DSDNP equation [
4,
5]:
where
x is the quantitative parameter of the process,
x′ and
x″ are the rate and acceleration of changes in this parameter at time
t (the first and second derivatives), respectively,
k is the proportionality coefficient and exponents λ and α describe the nonlinearity of the process near the stationary state (
x′ ≈
x′
0) and at a significant distance from it (
x′ >>
x′
0). The above conclusion about the self-development of natural systems corresponds to DSDNP Equation (1) at λ = 2, α = 2γ and
k =
Cmxγ−1.
Equation (1) is difficult to use in practice. Therefore, in our research, we used a simplified version of the equation as an approximation model:
Equation (2) corresponds to potentially catastrophic processes with a large range of changes in parameter x.
Equation (2) formally corresponds to the equation proposed by B. Voight [
6] for describing the dynamics of brittle deformations (disjunctive dislocations) on the eve of the culmination of volcanic eruptions. The Voight equation is used in the Forecasting Failure Method (FFM). However, in earthquake forecasting, Equation (2) was used for the first time in the pioneering work by G. Papadopoulos [
7]. The same equation is used in the Accelerated Moment Release (ARM) method using accumulated moment or Benioff strain [
8,
9,
10]. Attempts to use these methods for predictive purposes have not had success. Moreover, both methods have been criticized. In particular, a number of papers [
11,
12,
13] have claimed that the FFM method is biased and inaccurate even for a retrospective analysis. In turn, the statistical insignificance of the ARM method was justified in Reference [
14], where it was shown that the ARM practice carries the hazard of identifying patterns that are not real but are created by choosing the free parameters for demonstration of the hypothesized pattern. This hazard is particularly high when the results are unstable.
Serious criticisms have led to a reduction in the number of attempts to use Equation (2) for predictive purposes. Currently, only a few researchers using the ARM method (primarily References [
15,
16]) are trying to cope with the criticism (‘gravestone’ on this issue) of J. Hardebeck and coauthors [
14]. In addition, the Self-Developing Process (SDP) method [
17,
18] is actively used to study the seismicity of Sakhalin Island and adjacent territories (Russia). The SDP method was developed by I. Tikhonov based on the DSDNP equation for analyzing the flux of seismic events. Nevertheless, all the criticisms expressed in Reference [
14] apply to this method as well.
In our opinion, both points of view (supporters of both the FFM and ARM methods and their opponents) have a right to exist. Each researcher can use intuition in his scientific research, but he must (1) consolidate the results obtained in objective and reproducible criteria and (2) confirm the receipt of similar results using these criteria on the maximum possible number of examples (wide test control). An analysis of critical comments on the FFM and ARM methods showed that one of their main problems is the instability of the results obtained. We can confirm the seriousness of this problem due to the experience of our own research. The instability of the results, in our opinion, is largely due to the choice of the optimization criterion: the standard least squares method, commonly used by researchers (in particular, in References [
11,
12,
13]), does not provide the stability of approximation modeling and is therefore ineffective. The Revised-ARM method [
15,
16] does not solve this problem either. Therefore, the passage of the ARM method through a wide test control seems unlikely. As for the Tikhonov method [
17,
18], in our opinion, it needs objective criteria for the selection of representative earthquakes and the subsequent receipt of the results in automatic mode.
Our approximation algorithms were configured and successfully tested on synthetic catalogs during the second half of the 1990s. However, subsequent extensive testing on real seismic catalogs showed the instability of the results until the generally accepted optimization criterion for the smallest standard deviations was replaced by optimization for the minimum area deviations in the mid-2000s [
5]. Further, the problem with the missing assessment system for the predictability of seismic trends got in the way of our research. This evaluation system was developed by us by the mid-2010s [
19]. Optimization by area deviations proved inconvenient for predictive estimates. Therefore, it was replaced by optimization based on bicoordinate (mean geometric) deviations, which also provided the necessary stability to the results obtained. As a result, an algorithm was developed to identify seismic trends and assess their predictability (extrapolability). This algorithm has been extensively and successfully tested on real seismic catalogs when they are fully scanned in automatic mode. The initial version of the algorithm was applied to localized volcanic seismicity [
19,
20]. Then, this variant was adapted to a 3D space and used in the study of the predictability of seismic trends according to the Kamchatka Regional Catalog (KAGSR) [
21], seismic catalogs of the US Geological Survey (USGS) [
22,
23] and the Japan Meteorological Agency (JMA).
The results of the above works show a good predictability of the seismicity trends, including both trends of increasing activity before strong earthquakes and trends of the attenuation of activity after these earthquakes. However, it should be emphasized here that ‘the predictability of the seismic trend before a strong earthquake’ and ‘the forecast of a strong earthquake’ are not equivalent concepts, even if a strong earthquake fully corresponds to the extrapolation (forecast) part of the trend. Any earthquake that is part of the seismicity trend forms a deviation from the main trend pattern. The prediction of the time and magnitude of these random fluctuations (located in the band of permissible deviations) is difficult even with good predictability of the trend itself.
Thus, the use of Equation (2) in the prediction of strong earthquakes has natural limitations. Nevertheless, the approximation–extrapolation technique developed on the basis of the DSDNP equation, in our opinion, is a good tool for retrospectively studying the patterns of preparation of strong earthquakes and identifying their hazards in real time. When generalizing the previously obtained results on the flux of seismic energy, it was found [
24] that the range of values of the parameters α and
k in Equation (2), which determined the predictability of strong earthquakes, in the diagram α–lg
k (
Figure A1) is shifted to the left and below the entire area (including conditionally safe) predictability, i.e., to where the figurative points of the most slowly developing and long-term activation processes are located. This makes it possible to differentiate the trends in the activation of the seismic energy flux by the α and
k parameters, on the one hand, into conditionally safe (without precedents of termination by strong earthquakes), and, on the other hand, into potentially hazardous ones that deserve close attention due to the precedents (often repeated) of these trend completions with strong earthquakes. This paper describes an algorithm for estimating the hazard of strong earthquakes and illustrates its use for the flow of seismic energy based on the analysis of data from the USGS catalog as of 1 June 2021.
About terminology. Earthquake sequences are divided by the sign of the coefficient
k in the approximation–extrapolation trend:
k > 0—activation sequences,
k = 0—stationary sequences and
k < 0—attenuation sequences. Foreshock sequences are activation sequences that end with strong earthquakes. Foreshocks—all earthquakes of the foreshock sequence preceding a strong earthquake. Aftershock sequences are attenuation sequences that begin after a strong earthquake. Aftershocks—all earthquakes of the aftershock sequence after the main shock. In this paper, only activation sequences and (among them) foreshock sequences are considered. In relation to the time of the mainshock, there are [
7,
25] close (from several hours to several days), short-term (up to 5 to 6 months) and long-term (several years) foreshocks among the foreshocks. We examine all the listed types of foreshocks without exception.
2. Materials and Methods
2.1. The Model Equation
The solutions of Equation (2) can be expressed in the explicit form:
k = 0 | x = x1 + x’(t − t1), x′ = const |
k ≠ 0, α ≠ 1, α ≠ 2 | x = Xa + [k(α − 1)(Ta − t)](α−2)/(α−1)/[k(2 − α)], Ta = t1 + (x′11−α)/[k(α − 1)], Xa = x1 + (x′12−α)/[k(α − 2)] |
k ≠ 0, α = 1 | x = Xa + (x1 − Xa) exp[k(t − t1)], Xa = x1 − x′1/k |
k ≠ 0, α = 2 | x = x1 + ln|(Ta − t1)/(Ta − t)|/k, Ta = t1 + 1/(kx′1) |
Thus, the solutions of Equation (2) are either directly a linear dependence (
k = 0) or are reduced to linear dependences by taking the logarithm of the differences between the parameter values and/or time and the respective asymptotes:
k ≠ 0, α ≠ 1, α ≠ 2 | ln|x − Xa| = c1ln|t − Ta| + c0, c1 = (α − 2)/(α − 1), c0 = ln|k(α − 1)|(α − 2)/(α − 1) − ln|k(α − 2)| |
k ≠ 0, α = 1 | ln|x − Xa| = c1t + c0, c1 = k, c0 = ln|x1 − Xa| − k × t1 |
k ≠ 0, α = 2 | x = c1ln|Ta − t| + c0, c1 = −1/k, c0 = x1 + ln|Ta − t1|/k |
2.2. The Optimization and Its Criteria
The linearity of solutions of Equation (2) in ordinary or logarithmic coordinates simplifies optimization (finding the best match to the factual data). Direct optimization by five values (α, k, x1, x′1 and t1) is complex and requires large computing resources. However, when using solutions from Equation (2) in linear form (conventional or logarithmic), optimization is reduced to the analysis and comparison of several variants of linear regression. In some cases, optimization may also be required with respect to one or two additional parameters: Ta and/or Xa. The optimal values (α, k, x1, x′1 and t1) for each variant are easily determined analytically from linearity constants (c0, c1) and asymptotes (Ta and/or Xa). This greatly simplifies the optimization procedure and reduces the requirements for computing resources.
For the stability of the results, it is of great importance to choose an optimization criterion—a quantitative characteristic of the correspondence between the factual data and their approximation model. Therefore, to solve the problems that arise (see their detailed description in Reference [
22]), all variants of linear regression are compared in ordinary (nonlogarithmic) coordinates. The bicoordinate rms deviation Δ
xt = {Σ(Δ
xiΔ
ti)/[
n(
xc −
xs)(
tc −
ts)]}
0.5 is used as an optimization criterion to ensure the stability of the results. Here, (
xc −
xs) and (
tc −
ts) are the ranges of variations of the factual data for optimization (these ranges normalize the coordinates to a range from 0 to 1), and Δ
xi and Δ
ti are the deviations of each point of the factual data from the calculated curve along the abscissa and ordinate axes, respectively. In a geometric sense, the bicoordinate deviation corresponds to the side of a square equal in area to a rectangle with sides Δ
xi and Δ
ti, i.e., the bicoordinate deviation is the geometric mean of these deviations. The bicoordinate root mean square deviation is not the only possible criterion for the stability of the optimization results (optimization by area deviations has been successfully applied before), and moreover, it cannot be argued that this criterion is the best. However, it allows you to get stable results and is adequate for the available computing resources. For greater sensitivity, optimization is performed according to the maximum of the regularity coefficient (the inverse value for bicoordinate deviation):
Kreg = 1/Δ
xt.
2.3. The Processing of Seismic Catalogs
A spatial analysis of seismic data was carried out by spherical hypocentral samples with radii of 7.5, 15, 30, 60, 150 and 300 km. For each radius, the sample centers formed a fixed grid over the entire surface of the Earth and in subsurface spaces up to depths of 1000 km. Within this grid, the sample centers are distributed by latitude, longitude and depth, with an offset step that is 1.5 times smaller than the sample radii (i.e., 5, 10, 20, 40, 100 and 200 km, respectively), which provides spatial overlap of the samples. Catalog processing is executed automatically for each radius. Each event in the analyzed catalog is consistently treated as a ‘current’ event (earthquake). The moment of time of this event is taken as the ‘present’. The time preceding this event is considered the ‘past’, and the subsequent time is considered the ‘future’. To analyze the seismicity preceding and following the ‘current’ event, a spherical sample with the center closest to the hypocenter of the ‘current’ earthquake is used. Within this sample, an array of data of the studied flow parameter is formed (in our case, the flux of seismic energy, i.e., the total energy of earthquakes).
The main trends of the ‘current’ seismicity are revealed by test approximations in order to search the factual data for those intervals that demonstrate the best characteristics during optimization. The first test approximation is performed based on the factual data presented by the ‘current’ earthquake and the 6 preceding ones. In subsequent test approximations, the nearest event from the ‘past’ is added to the approximated factual data until the first event in the sample is included in their number. All approximations with Kreg < 10 are ignored. From all the test approximations, the three best variants are selected: the first one is based on the maximum Kreg, and the rest are based on the nearest and main maxima of the Kreg/Klin ratio. The first variant is always determined, the rest depending on the presence and combination of current nonlinear trends. Nonlinearity variants allow you to track new development trends that begin (and, therefore, are still poorly expressed) against the background of the main trends.
2.4. The Estimation of Extrapolation Predictability
The term ‘trend predictability’ is defined here as finding the factual data of the ‘future’ in the band of acceptable errors relative to the calculated curve in its extrapolation part. To estimate the trend predictability, the rms deviation σ of the factual points (
tf,xf) from the calculated curve along the one normal to it is used. It is calculated on the approximation section of the trend in coordinates normalized to a range from 0 to 1 from the first (
ts,
xs) to the last point (
tc,
xc):
Equation (3) is obtained on the basis of elementary geometric constructions (
Figure 1), in which the shortest distance from the factual point (
tf,
xf) to the calculated curve is estimated in the first approximation as the height
h of a right triangle (
tr,
xf)–(
tf,
xf)–(
tf,
xr) lowered by the hypotenuse (
tr,
xf)–(
tf,
xr) from the opposite vertex, which is the factual point (
tf,
xf):
h =
ab/
c. As a result, Equation (3) is an expression for the root mean square value of these distances.
Further, the approximation is extrapolated into the ‘future’ as long as the distance of each subsequent (predicted) factual point (
tp,
xp) to the calculated point is in the band of permissible errors ± 3σ, i.e., the ratio is fulfilled:
The width of the error band here is determined from the statistical rule ‘3 sigma’, according to which, 99.73% of the results fall into such an error band in the case of a normal distribution. Since the ratio (4) uses normalization for the range from 0 to 1 from the first (ts,xs) point to the point being tested (tp,xp), then the average deviation is pre-recalculated to the same normalization range; that is, σ is pre-recalculated on the approximation section of the trend according to Equation (3) with the replacement of tc and xc by tp and xp.
2.5. Quantitative Estimates in Prediction Precedents for Strong Earthquakes
The relative accuracy of the precedent predictions is estimated by the formula:
Further, the following classification of the accuracy of the retrospective predictions is used in the work: quantitative estimations at Δ ≥ 5 (relative error <20%), semi-quantitative estimations at 2 ≤ Δ < 5 (relative error 20–50%) and qualitative estimations at Δ < 2 (relative error > 50%). Here, the relative error is the inverse of the relative accuracy.
The predicted nonlinearity of
Lpn is calculated by the formula:
Factual values (xsh_f, tsh_f) are used in retrospective estimates. When predicting by precedent, instead of them, the calculated values (xsh_i, tsh_i) for a strong earthquake are used in the Equation (6).
In essence, the predictive nonlinearity of Lpn corresponds to the sign of the coefficient k in the DSDNP equation, but not in the integer, except in real terms. For extremely nonlinear activation sequences, parameter predictability dominates over time predictability, so the Lpn value is close to 1. As the ratio in the predictability of the trend in terms of parameter and time is leveled, the Lpn value decreases, reaching 0 for sequences close to stationary development. A further shift of the ratios in trend predictability leads to an increasing increase in time predictability compared to parameter predictability, which corresponds to attenuation sequences. Lpn values tend to be −1 for extremely nonlinear attenuation sequences, in which the time predictability significantly exceeds the parameter predictability. Thus, for the activation sequences considered by us (foreshock sequences in case of the completion of activation by a strong earthquake), the Lpn value varies from 0 to 1. As will be shown below on the example of retro-forecasts, the predictive nonlinearity of the Lpn determines the asymmetry of the band of permissible deviations and thereby reflects the stochasticity/determinism of the position (fluctuations) of the main thrust of the predicted trend.
The approximation–extrapolation coefficient
A shows how many times the general trend of activation exceeds the approximation part included in it. This coefficient is calculated in the coordinates of the full trend, normalized to a range from 0 to 1 and is used to estimate the limit of possible extrapolations. In general, the value of
A can be determined by the ratio of the lengths of the corresponding sections of the calculated curve; however, in the case of step cumulative characteristics of the seismic flow (energy, Benioff strain or the number of events), the formula is simpler and quite effective:
When predicting by a precedent, instead of the factual values (xsh_f,tsh_f) in Equation (7), calculated values (xsh_i,tsh_i) for a strong earthquake are also used.
2.6. The Real-Time Predictive Estimation Algorithm
The use of the precedents from retro-forecast data is based on the possibility of linking the time
tsh_f of the main earthquake to the rate
x′
sh of change of the parameter at the point of the extrapolation curve closest to the main earthquake. In retrospective studies, this point is determined in parameter–time coordinates, normalized to a range from 0 to 1 according to the factual values from point (
ts,
xs) to point (
tsh_f,
xsh_f). If the factual point of a strong earthquake is located within the region of existence of the extrapolation curve (
tsh_f <
Ta at α > 1 and
xsh_f <
Xa at α > 2), then the distances to the extrapolation curve from the factual point of a strong earthquake are determined by the abscissa (
a) and ordinate (
b):
a = |
tsh_f −
t(
xsh_f)|,
b = |
xsh_f −
x(
tsh_f)|. Geometrically, these distances correspond to the cathetus of a right triangle with a vertex at the point (
tsh_f,
xsh_f) (see
Figure 1, assuming that point (
tf,
xf) is point (
tsh_f,
xsh_f) and point (
ti,
xi) is point (
tsh_i,
xsh_i)). Then, the position of point (
tsh_i,
xsh_i) is approximately defined as the intersection of the hypotenuse perpendicular to it from the vertex of the right angle. Based on the proportions existing in a right triangle, we determined the coordinate values for this point:
tsh_i =
tsh_f + (
t(
xsh_f) −
tsh_f)/(
a +
b) and
xsh_i =
xsh_f + (
x(
tsh_f) −
xsh_f)/(
a +
b). Using these coordinates, the rate
x′
sh is calculated:
x′
sh = [
k(α − 1)(
Ta −
tsh_i)]
1/(1−α) at α ≥ 1.5 or
x′
sh = [
k(2 − α)(
xsh_i − Xa)]
1/(2−α) at α < 1.5.
If one of the coordinates of the factual point of a strong earthquake goes beyond the area of existence of the extrapolation curve, then the nearest point of the calculated curve is determined by the second coordinate (abscissa or ordinate, respectively): xsh_i = x(tsh_f) and tsh_i = tsh_f or tsh_i = t(xsh_f) and xsh_i = xsh_f. After that, the rate x′sh is calculated according to the above formulas. If both coordinates go beyond the limits of the existence of the extrapolation curve (tsh_f ≥ Ta and xsh_f ≥ Xa at α > 2), the asymptotic point (Ta,Xa) turns out to be the closest to the earthquake; therefore, an extremely large value is conditionally assumed as the rate x′sh.
The regularities of the precedent foreshock preparation of strong (M7+) earthquakes allow identifying similar trends in the seismic activity increase. The prediction estimates of these trend hazards assume the use of the data of precedent retrospective predictions and are based on the possibility of binding the time of the main earthquake to the rate of change in the x′sh parameter at the point of the extrapolation curve closest to the main earthquake. For this purpose, a database of precedent retrospective predictions is created. This database includes information about the hypocentric radius of the sample, α, k and x′sh, as well as information about this strong precedent earthquake (magnitude, time and place).
The essence of the precedent–extrapolation assessment of a seismic hazard is to identify potentially hazardous spatial zones where there is such an increase in seismic activity that has historical precedents of ending with a strong earthquake. The quantitative aspect of the forecast corresponds to the calculation of the possible time of a similar earthquake based on the database of its previous forecasts. For the activity in each spatial zone, analogs are possible in the preparation of several strong earthquake precedents. In this case, calculations of the possible time of a similar earthquake in the considered spatial region are performed for each of the precedents.
The prediction extrapolations algorithm provides for the following operations:
Search in the catalog for unfinished (not come out of the band of admissible errors at the time of the catalog end) prediction definitions, in which a tendency towards an increase in seismic activity is found.
Comparison of the type of an activity increase with the database of precedent retrospective predictions alongside the sample radius, exponent α (with an accuracy of 0.01) and coefficient k (when comparing lg k with an accuracy of 0.1). All cases of an activity increase that have no analogs in the database of precedent retrospective predictions are ignored.
For each precedent retrospective prediction based on the rate of change in the
x′
sh parameter, the time
tsh and the value of the
xsh parameter are calculated, at which, for a given type of an activity increase, its level will correspond to the level of the precedent shock:
Then, the values of Lpn, A and σt are estimated. The definitions, for which the approximation and extrapolation ratio A exceed the maximum value of Amax for retro-forecast precedents, are considered as having no precedents.
The revealed precedent retrospective predictions are grouped by the main shock. For each group, the calculated average time of the strong earthquake and its standard deviation σsh, as well as the average values of Lpn, A and σt, are determined.
2.7. Initial Data
As the initial data for a precedent-based extrapolation estimate of the M8+ earthquake hazard, this work uses the worldwide United States Geological Survey (USGS) earthquake catalog [
26], which includes data from 1900 to the end of May 2021. By this time, the catalog contains data on 3,889,120 earthquakes with a magnitude
M = −1.0 … +9.5 at its modal value 1.2. The results of processing the Japan Meteorological Agency (JMA, 1919–present) earthquake catalog [
27] and the Earthquakes Catalogue for Kamchatka and the Commander Islands (ECKCI, 1962–present) [
28] were also used to analyze the foreshock predictability and form a database of precedent retro-forecasts. By the end of August 2018, the JMA catalog contained data on 3,498,071 earthquakes with a magnitude
M = −1.6 … + 9.0 at its modal value 0.6. By the end of March 2021, the ECKCI catalog contained data on 428,225 earthquakes with a magnitude
M = –2.7 … + 8.1 at its modal value 0.5. The seismic energy flux
E is considered as the
x parameter, i.e., a cumulative amount of earthquakes energy. In this case, the energy of a single earthquake is estimated according to the existing relationship between its magnitude
M and the energy class
K [
29]:
K = lg
E = 1.5
M + 4.8. The relationship between
M and energy
E is valid for
E in Joules. When processing catalogs, all earthquakes are used, for which there are energy characteristics (
M or
K), regardless of their magnitude. Our research does not require the completeness of seismic catalogs and does not depend on it. We study the flow of seismic energy as it is (where it is recorded and how it is recorded). This is of great importance for this work, since the USGS world catalog is made up of many regional catalogs and, therefore, is extremely heterogeneous in completeness both in space and in time.
4. Discussion
The results obtained confirm the conclusions previously made [
21,
22,
23] about the good predictability of the seismic energy flow when using the DSDNP equation as a model and the prospects of using this equation for the prediction of strong earthquakes. However, here, in order to avoid misunderstandings, it is necessary to specify the terminology used, as well as to return to the topic of natural limitations of the DSDNP equation (and its analogs) in the prediction of strong earthquakes touched upon in the introduction. A forecast is usually understood as determining the location, strength (magnitude), time and probability of occurrence of future earthquakes. An attempt to detail this definition is contained in the overview report by the International Commission of Earthquake Forecasting for Civil Protection [
30] (p. 319): “A prediction is defined as a deterministic statement that a future earthquake will or will not occur in a particular geographic region, time window, and magnitude range, whereas a forecast gives a probability (greater than zero but less than one) that such an event will occur.”
The DSDNP equation, like its analogs, has natural limitations that take it beyond the above definition of forecasts, both unambiguous and probabilistic. Firstly, using this equation, it is possible to successfully identify and predict trends in seismic activity but not fluctuation deviations from these trends. The trend forecast only allows you to determine the band of permissible deviations for it. Secondly, there is a ‘magnitude of uncertainty’ in the forecasts of seismic trends in the energy flux: the same increment of seismic energy can be realized through a singular M8 earthquake, 32 M7 earthquakes or one thousand M6 earthquakes. All these alternatives are equal from the point of view of trend predictability, provided they are in the band of permissible deviations. It is impossible to determine in advance exactly how the predicted potential of the seismic energy flow is realized (in the form of a single strong earthquake or a swarm of moderate ones). Thirdly, the use of unambiguous and probabilistic methods is not applicable to the predictability of the trend itself. From the point of view of the dynamics of self-developing natural processes [
4], the formation of activation trends in accordance with Equations (1) and (2) is due to the system’s exit from the state of equilibrium (of stationary development), which has become unstable, and the subsequent transition of the system to a new state of equilibrium (of stationary development). The change of mode from activation to attenuation in each specific case is determined by the internal state of the system, which cannot be estimated either by unambiguous or probabilistic methods. It is only possible to track potentially hazardous trends, assessing the increase in possible threats or fixing the termination, in fact.
The calculation of the precedent time by Equation (8) is not a real forecast. It is only the formal determination of a reference point in the increasing flow of seismic energy, in the vicinity (tolerance band) of which, in the case of a similar development of the process in accordance with Equation (2), a strong earthquake was once already registered. This allowed us to conclude that this process of increasing activity is potentially hazardous and to assess the possible time frame of this hazard. However, the probability of repeating the characteristics of a precedent earthquake in this process is negligible. If a strong earthquake occurs in a potentially hazardous cycle of increasing activity, it will create its own precedent with its magnitude, with its deviations (although permissible) from the calculated trends in parameter and time and, therefore, with its calculated level of activity at the time of the earthquake.
The described method is limited by the ‘non-repeatability’ of precedent earthquakes, the inability to determine the magnitude range for the future mainshock and the unpredictable completion of hazardous trends (often long before hazardous levels of activity). All this takes the described method beyond both unambiguous and probabilistic earthquake forecasts. Therefore, the only purpose of this method is to identify areas of potentially hazardous increases in the flux of seismic energy and their subsequent monitoring. In fact, precedent-based extrapolation estimates can be considered as a preparatory stage for future real earthquake forecasts. Nevertheless, it may seem that precedent-based extrapolation estimates fall in the category of seismic pattern methods. However, as we have already indicated in the introduction, the forecast of seismic trends cannot be correctly used to predict the time and magnitude of random deviations from these trends (actually the mainshocks). Equating our method with earthquake prediction methods from the category of seismic patterns in order to compare their effectiveness, in our opinion, is doubly wrong. That is why we do not consider it possible to compare our precedent-based extrapolation estimates of the danger of seismic trends with existing earthquake forecasting methods [
30].
Maps of the distribution of precedent clusters are convenient for monitoring spatial hazards (see
Figure A17). The ‘reddening’ of the cluster ellipsoid shows an increase in the number of potentially hazardous trends in its composition. In particular, on the maps of
Figure A17, it can be seen that, after two M7+ earthquakes, the hazard of stronger earthquakes (M8+) in the area of the Aleutian Islands has sharply increased. A decrease in ‘redness’ indicates a decrease in hazards, i.e., a decrease in the number of potentially hazardous trends due to exceeding the limits of permissible deviations or the limit of possible extrapolations (
A >
Amax).
Monitoring the hazard over time is complicated by the qualitative level of accuracy of forecast extrapolations. Updating the database with additional information on corrections for differences between the actual and calculated values of the parameter and time at the time of the earthquake would significantly increase the accuracy of determining the earthquake time when testing on its own predictive precedents (see
Table A5,
Table A6,
Table A7,
Table A8,
Table A9,
Table A10 and
Table A11). However, in general, the situation with the accuracy of time estimates would most likely not have changed, since the low accuracy of extrapolating the flow of seismic energy is ultimately controlled by a large spread of actual data for this parameter. Benioff strains show a more ordered dependence on time. Therefore, their studying looks more promising in terms of the accuracy of the forecast in time.
In the estimates of the precedent time, there is another problem that remains within the framework of this article not only unresolved but not even disclosed. This is the problem of the multivariance of precedent trends in the computational cluster. As an example, it is enough to pay attention to large σsh values, especially against the background of low σt values. In particular, for cluster 17, the standard deviation of the estimated earthquake time (σsh) is 4.5 years, with only a daily expected average deviation in time of the actual data from the calculated curve (σt). This indicates the presence of several alternatives (or their ‘fans’) for the further development of the hazard with significant discrepancies between them. The number of clusters with similar large σsh values is a significant part of their total number, which requires additional research to solve this problem.
Concluding the discussion, it should be noted that, in the current state, the methodology of the precedent–extrapolation estimates is a primary (research) variant that needs further adaptation, debugging, identification and the elimination of minor shortcomings and (possibly) errors. Above, we showed the possibility of its practical use in predictive research both in retrospect and in real time. However, a long process of testing, improvements and debugging awaits this technique ahead.