 Full paper
 Open Access
 Published:
Testing the seismic quiescence hypothesis through retrospective trials of alarmbased earthquake prediction in the Kurile–Japan subduction zone
Earth, Planets and Space volume 73, Article number: 100 (2021)
Abstract
We make trial binary forecasts for the Kurile–Japan subduction zone for the period 1988–2014 by hypothesizing that seismic quiescence (i.e., the absence of earthquakes of M ≥ 5 for a minimum period of T_{q}) is a precursor of a large (7.5 ≤ M_{w} < 8.5) earthquake in the coming period T_{a} within a radius R of the quiescence. We evaluate the receiveroperatingcharacteristic diagram constructed using a range of forecast models specified by (T_{q}, R, T_{a}). A forecast experiment targeting eight large earthquakes in the studied spacetime suggests that the risk of a large earthquake is modestly (probability gain G ~ 2) but significantly (pvalue less than 5%) heightened for several years following a long quiescent period of T_{q} ≥ 9 years, within several tens of kilometers of the quiescence. We then attempt crossvalidation, where we use half the data for training [i.e., optimization of (T_{q}, R, T_{a})] and the remaining half for evaluation. With only four target earthquakes available for evaluation of the forecasts in each of the learning and evaluation periods, our forecast scheme did not pass the crossvalidation test (with a criterion that the pvalue is less than 5%). Hence, we cannot formally deny the possibility that our positive results for the overall period are a ghost arising from overfitting. However, through detailed comparison of optimal models in the overall test with those in the crossvalidation tests, we argue that severe overfitting is unlikely involved for the modest G of ~ 2 obtained in the overall test. There is thus a reasonable chance that the presently tested type of quiescence will pass the crossvalidation test when more target earthquakes become available in the near future. In the meantime, we find that G improves to ~ 5 when target earthquakes are limited to 8 ≤ M_{w} < 8.5, though we cannot say anything about the possible involvement of overfitting because we have only three such very large target earthquakes.
Introduction
Large earthquakes are often preceded by a period of seismic quiescence; that is, a remarkable drop in regional background seismicity lasting for several years. Drawing attention as a possible intermediateterm earthquake precursor (e.g., Scholz 2019), quiescence has a long history of research. Whereas earlier studies (e.g., Inouye 1965; Utsu 1968; Mogi 1969; Ohtake et al. 1977; Kanamori 1981) lacked an objective definition of quiescence, later studies have developed a variety of measures quantifying quiescence, including the ZMAP method (Wiemer and Wyss 1994; Katsumata 2011, 2017a), ETAS modeling (Ogata 1992), and RTL/RTM method (e.g., Sobolev and Tyupkin 1997; Nagao et al. 2011). Thus, seismic quiescence has been established as an objectively demonstrable anomalous incident, which often precedes large (typically M_{w} ≥ 7.5) earthquakes by several years to decades.
Finding more and more earthquakes preceded by quiescence, however, does not answer the real question of interest; does a statistically significant tendency exist for quiescence to precede large earthquakes? Specifically, we must ask if the observed frequency of the coincidence of the two occurrences (quiescence and a subsequent large earthquake) is beyond the level explainable as a mere matter of chance. To our knowledge, this point has not been tested, except for an attempt by Katsumata (2017a), who, unfortunately, failed to set an objective criterion for the coincidence as detailed later. Thus, while long having received research attention, quiescence has remained a wellknown precursor candidate for half a century.
In the meantime, an earthquakepreceding tendency (e.g., Nakatani 2020) has been proven to exist for a special class of quiescence called ‘relative quiescence’ (Ogata 2001), a phenomenon that the aftershock seismicity following a large earthquake starts depleting at one point of time during the aftershock period, compared with extrapolation following Omori's law. However, there is no reason to presume that the affirmative conclusion about the precursory relative quiescence applies to the (general) quiescence seen in the nonaftershock period.
To judge if a certain phenomenon has an earthquakepreceding tendency, we need to scan the entire spacetime of study, not only the times preceding large earthquakes, for the same type of phenomenon. We refer to the phenomena as ‘anomalies’ following the convention adopted in earthquake precursor research, but ‘anomalies’ in this context can be any incidents that are objectively definable; they do not have to be rare or anomalous in a statistical sense.
On the basis of the above concept, Katsumata (2017a) exhaustively scanned background seismicity in subduction zones around Japan in the period 1975–2012, for longterm quiescence anomalies (lasting longer than 9 years). He compared the detected anomalies with four earthquakes having M_{w} ≥ 8.25 that occurred in the studied spacetime, using a contingency table. A Fisher’s exact test found a pvalue (i.e., the probability that the observed or higher extent of correlation emerges by chance under the null hypothesis of no correlation between the anomalies and earthquakes) of 0.021, suggesting that the quiescence has an earthquakepreceding tendency. Conceptually, this is a correct method of evaluation. However, he did not set a consistent, objective criterion with which to judge whether a large earthquake followed a quiescence anomaly, leaving room for doubt. More importantly, it seems that Katsumata (2017a) hesitated to set an objective criterion for the anomaly–earthquake association to get around a technical difficulty inherent to a contingency table as explained below.
In evaluating a contingency table for an earthquakepreceding tendency, the number of anomalies or alerts incurred by them needs to be counted. When multiple incidents of anomalies occur close to each other in space and time, incurred alerts largely overlap. In such cases, one physical incident of a correct (or incorrect) anomaly or alert is evaluated as many successful (or unsuccessful) incidents, while it should be ideally counted as just one incident of success (or failure) in the contingency table. Thus, in practice, number counting is not a good way to quantify anomalies or alerts. Probably as a subconscious workaround for this problem, Katsumata (2017a) subjectively combined clustered anomalies into one incident of quiescence for the contingency table. The evaluation conducted by Katsumata (2017a) thus lacked objectivity, even though his subjective grouping is probably reasonable from a physical point of view.
The above difficulties with the contingency table can be avoided by instead using an alarm map, with each spatiotemporal cell being given a binary value of either alarmon or alarmoff according to any objective rule. A simple example may be alerting a region within a distance R from an anomaly for a time duration T_{a} following the anomaly. The total alerted volume, instead of the number count of alarms or anomalies, can be used for alarm quantification. In this way, largely overlapped alerts from clustered anomalies do not cause disproportionate weighting of their success or failure.
Under the null hypothesis of no correlation between the anomalies and earthquakes, expectancy for the prediction rate (i.e., the ratio of the number of alerted earthquakes to the total number of target earthquakes) is equal to the alarm fraction (i.e., the ratio of the alerted volume to the total spacetime of the study). On this basis, the statistical significance of the earthquakepreceding tendency and its strength can be evaluated using the pvalue and the probability gain, respectively (e.g., Zechar and Jordan 2008; Nakatani 2020).
Note that the above statistical testing cannot tell if individual anomalies that preceded the target class of earthquakes were indeed related to the earthquake. Instead, the 'existence of an earthquakepreceding tendency’ means that one or more of the anomalies followed by earthquakes were indeed related to the subsequent earthquake occurrence.
In the present paper, we evaluate the earthquakepreceding tendency of longterm (~ 10year) quiescence in the Kurile–Japan subduction zone (Katsumata 2017a) adopting the alarmmapbased testing described above.
As the primary goal of the present paper is to demonstrate rigorous yet generic statistical procedures of alarmmapbased evaluation, we only consider quiescence to avoid distraction. However, we recognize that various, sometimes even opposite, senses of seismicity changes [e.g., quiescence and activation (Reasenberg and Matthews 1988) and acceleration and deceleration (Hardebeck et al. 2008)] might precede earthquakes. We emphasize that there is no logical or practical difficulty in producing and evaluating alarm maps based on such mixed behaviors as long as the alerting procedure is stated objectively; e.g., the M8 algorithm (KeilisBorok and Kossobokov 1990).
The statistical power depends strongly on the number of targetclass earthquakes available, and we thus make trial forecasts targeting earthquakes of M_{w} ≥ 7.5 instead of M_{w} ≥ 8.25 for which the precursory longterm quiescence has already been suggested (Katsumata 2017a). However, there are still only nine earthquakes of M_{w} ≥ 7.5, and this small number of target earthquakes is certainly a weak point of the present study. Although more target earthquakes would be available if we add more study regions, we decided not to add other tectonic regions as our priority is to investigate the effects of retrospective optimization (e.g., Mulargia 1997). For that purpose, we examine a suite of forecast models, where we vary the model’s three main adjustable parameters (i.e., the anomaly detection threshold and the temporal and spatial limits within which anomalies are associated with subsequent earthquakes) over a wide range. We also examine the likelihood of overfitting by conducting additional crossvalidation experiments. We therefore take a minimalistic approach in other regards.
Data
The present study area comprises subduction zones along the western margin of the Pacific Plate, spanning 25°–55° N and 138°–163° E and including the east coast of Kamchatka, the Kurile Islands, and the IzuBonin Islands. Forecast targets are shallow (having a centroid depth less than 70 km) earthquakes of M_{w} ≥ 7.5 listed in the Global Centroid Moment Tensor catalog. The term ‘centroid’ means the gravity center of the seismic moment released by an earthquake, while the term ‘hypocenter’ means the location where the main shock rupture started. For large earthquakes, the centroid and hypocenter usually do not coincide. Nine earthquakes (Fig. 1 and Table 1) satisfy these conditions in the period from January 1, 1988 to December 31, 2014. These nine earthquakes are clearly main shocks, sufficiently isolated on a space–time plot of seismicity. Evaluation of a forecast is not straightforward if the forecast targets include aftershocks (Michael 1997).
We exclude EQ#9, the 2011 Tohoku Earthquake (M_{w} 9.1), from our forecast targets because it was preceded by superlong (i.e., superstrong) quiescence that we could not detect owing to a mere practical limitation. As detailed in “Methods” section, we detect changes in seismicity by examining 15year subcatalogs of the background seismicity, such that there is no way of detecting quiescence lasting 15 years or more. If we use sufficiently long subcatalogs, we can correctly recognize the exceptionally long quiescence exceeding 20 years that preceded EQ#9 (Katsumata 2017b). Therefore, counting EQ#9 as a false negative would not be scientifically adequate. We therefore decided to exclude EQ#9. If one still finds it unfair to ignore EQ#9, whose precursor is missed by our presently adopted algorithm, we can modify our forecast target to 7.5 ≤ M_{w} < 8.5, instead of M_{w} ≥ 7.5.
In addition, the 1993 Hokkaido Nanseioki earthquake (M_{w}7.7, 42.71° N, 139.28° E, 16.5 km deep) is excluded from our forecast targets because the earthquake occurred in the "alarmundecidable" spacetime defined in the step (2) of “Alarm map” section. This event is not included in Table 1 or Fig. 1.
To detect quiescence, we analyze seismicity (i.e., bodywave magnitude m_{b} ≥ 5.0, depth ≤ 60 km, 5792 events in total) in the study area for January 1, 1964, through December 31, 2014. We downloaded data from Reviewed ISC Bulletin (ftp://iscmirror.iris.washington.edu/pub/prerebuild/ffb/catalogue/). Although the International Seismological Centre (ISC) has finished rebuilding the bulletin from 1964 to 1979 (Storchak et al. 2017), we use old data stored in the "prerebuild" directory to ensure temporal homogeneity of source parameters, especially the magnitude.
Catalog completeness may vary with space and time. As an example, Michael (2014) found that the magnitude of completeness, Mc, of the global ISCGEM catalog is 6.0 for shallow (depth ≤ 60 km) earthquakes that occurred from 1964 to 1975 around the world. However, in our study area, we find that Mc of the ISC catalog (depth ≤ 60 km) is much better, ranging from 3.5 to 4.7 throughout the study period (Additional file 1: Figure S1). Hence, our analysis using only events with m_{b} ≥ 5.0 is prudent.
Methods
Subcatalogs
We first make 16year subsets of the 5792 earthquakes of m_{b} ≥ 5.0. We make 351 subcatalogs, with the starting date varying from the year 1964.0 to 1999.0, in 0.1year steps, so that the first subcatalog covers 1964.0–1980.0, the second subcatalog covers 1964.1–1980.1, …, and the 351st subcatalog covers 1999.0–2015.0. We then remove aftershocks in each subcatalog, using a declustering algorithm based on the eightparameter ETAS model (Zhuang et al. 2002, 2005), where the observed seismicity is decomposed into the quasisteady background seismicity and the temporary surge due to aftershocktype triggering following the Omori–Utsu law (Utsu 1957).
We decluster each subcatalog independently. We first determine ETAS parameters by fitting only the seismicity in each subcatalog. All eight ETAS parameters were reasonably stable among different subcatalogs (Additional file 1: Figure S2). It is thus unlikely that uncertainties in the ETAS parameters affect the conclusions of the present study, though we do not attempt a quantitative assessment of the impact (e.g., Wang et al. 2010).
We then produce a declustered subcatalog that only retains the events likely to belong to the background seismicity. Specifically, we judge that an earthquake belongs to the background seismicity if the ETASmodeled probability P_{back} that an earthquake belongs to the background seismicity exceeds P_{rand}, a random number drawn from a uniform distribution between 0 and 1 (Zhuang et al. 2002).
We then discard the first 1year portion of each declustered subcatalog because the ETAS model starts working properly only after seeing a sufficient length of prior seismicity. Although the clustering parameters of Page et al. (2016) for subduction zones suggest the possibility of aftershocks having m_{b} ≥ 5.0 in the second and later years, we find that aftershocks having m_{b} ≥ 5 do not exceed the background rate in our study region for more than a year even after an earthquake as large as M8 class, except for the prolonged aftershock period following the 2011 M9 earthquake (Additional file 1: Figure S3). Therefore, discarding the initial year is sufficient.
We thus obtain 351 declustered subcatalogs, each of which is 15 years long. The first subcatalog covers 1965.0–1980.0, the second subcatalog covers 1965.1–1980.1, …., and the last (351st) subcatalog covers 2000.0–2015.0. Hereafter, the term 'subcatalog' refers to these 15year declustered catalogs, not the 16year catalogs before declustering. Each subcatalog contains 671–986 events, or ~ 850 on average.
ETAS parameters are also spatially variable. However, for simplicity and the stability of analysis, we assume ETAS parameters to be spatially uniform. Though not shown, we attempted declustering with spatially variable ETAS parameters, only to find that the alarm maps based on this version of declustered subcatalogs are almost identical to the presently shown and evaluated maps made with spatially uniform ETAS parameters.
Trial forecast—anomaly detection and construction of the alarm map
In the present algorithm of trial forecasts, we first map the occurrence of quiescence anomalies throughout the studied spacetime (“Anomaly detection” section). Using the thus made anomaly map, we issue alarms according to the spatiotemporal distance from anomalies (“Alarm map” section). We update this alarm map every 0.1 years. Both anomaly criteria and alert criteria involve adjustable parameters. We will review the performance of the trial forecasts made with a wide range of parameter values in order to assess if the presently defined type of longterm quiescence has an earthquakepreceding tendency.
Anomaly detection
A variety of quiescence measures have been proposed as described in “Introduction”. In the present study, we adopt the duration of the streak of noearthquake days in the region, motivated by Katsumata (2017b), who investigated the seismicity preceding 23 earthquakes having M_{w} ≥ 8 for the period 1990–2014. He found that a streak of noearthquake days lasting longer than ~ 10 years preceded all but four earthquakes that occurred in regions where the background seismicity was already too low to detect any further drop.
Judgment of quiescence anomalies is made every 0.1 years, for each of the spatial grid points laid at intervals of 0.1° N × 0.1° E. The algorithm is as follows:

(1)
Select a subcatalog, which covers T_{1} through T_{2} (= T_{1} + 15 years). 1965.0 ≤ T_{1} ≤ 2000.0 and 1980.0 ≤ T_{2} ≤ 2015.0. Each grid point shall be diagnosed for the occurrence of a quiescence anomaly as of T_{2}.

(2)
Select a grid point for which the anomaly judgment is made. From the selected subcatalog, find the six nearest earthquakes around the grid point.

(3)
If the epicentral distance to the sixth nearest earthquake exceeds 100 km, conclude the grid point as 'anomalyundetectable' as of T_{2} because an anomalous drop in seismicity is difficult to recognize for the spacetime that is already very quiet.

(4)
Let dt be the time interval from the most recent of the six earthquakes to T_{2}. This dt represents the duration of quiescence (i.e., streak of noearthquake (m_{b} ≥ 5) days) around the grid point as of T_{2}. If dt ≥ T_{q}, conclude the grid point as 'yesanomaly' as of T_{2}. The value of T_{q}, the threshold for dt to be regarded as anomalous, is one of the three adjustable parameters of the present forecast algorithm. It may represent the forecaster's idea about the duration of precursory quiescence. A higher T_{q} means a more stringent selection of quiescence anomalies.

(5)
If neither (3) nor (4) applies to the grid point, conclude it as 'noanomaly'.
Alarm map
On the basis of the anomaly map obtained in “Anomaly detection” section, we now construct an alarm map, any spatiotemporal point of which is given one of the three forecast values: 'alarmon', 'alarmoff', or 'alarmundecidable'. The algorithm is as follows.

(1)
Label the spacetime as 'alarmon' if at least one 'yesanomaly' grid point exists in the preceding period of length T_{a} within a distance R. R and T_{a} are two adjustable parameters of the present forecast scheme. They may represent the spatial and temporal ranges up to which the forecaster expects that quiescence anomalies are likely related to the subsequent earthquake.

(2)
Label the spacetime as 'alarmundecidable' if the spacetime does not satisfy (1) and if at least one 'anomalyundetectable' grid point exists in the preceding T_{a} within the distance R.

(3)
Label all the remaining spacetime as 'alarmoff'. Note that in the alarmoff spacetime, it is guaranteed that no quiescence anomaly occurred within R during the preceding T_{a}.

(4)
Finally, we relabel any spacetime belonging to the aftershock region of earthquakes having M_{w} ≥ 7.5 as 'alarmundecidable'. This is for prudence; application to an aftershock period is beyond the scope of the traditional precursory quiescence hypothesis. Using the ETAS parameters obtained in the declustering procedure, we calculate P_{after} = 1  P_{back} as a function of the time elapsed since the occurrence of the main shock. For the time period when P_{after} ≥ 0.01, the region within R of the grid point closest to the centroid of the main shock is designated as ‘alarmundecidable’. This step overrides decisions made in earlier steps. Spacetime labeled 'alarmundecidable' shall not count in the evaluation (see “Evaluation of the trial forecasts” section).
Figure 2 shows, in the form of yearly snapshots, an example (T_{q} = 11 years, R = 60 km, T_{a} = 7 years) of the spatiotemporal alarm map created adopting the above procedure. The time attached to each snapshot is that when the alarm status was decided and issued; that is, T_{2} of the newest subcatalog available then. For the anomaly map obtained in “Anomaly detection” section, T_{2} ranges 1980.0 ≤ T_{2} ≤ 2015.0, but, for the alarm map, T_{2} ranges 1988.0 ≤ T_{2} ≤ 2015.0 because T_{a} up to 8 years is considered in the present study.
Evaluation of the trial forecasts
For each forecast model, specified by a combination of (T_{q}, R, T_{a}), we thus obtain one spatiotemporal alarm map stating the model's forecasts for 1988.0 through 2015.0. We evaluate the performance of each model by looking at its prediction rate r, compared with the alarm fraction f invested by the model. Considering the existence of 'alarmundecidable' spacetime, the exact formulae for f and r are
and
where V_{on} is the total spatiotemporal volume labeled 'alarmon’, V_{off} is that labeled 'alarmoff’, N_{on} is the number of target (i.e., 7.5 ≤ M_{w} < 8.5) earthquakes that occurred in V_{on}, and N_{off} is the number of target earthquakes that occurred in V_{off}.
To obtain r, one needs to compare the forecast (alarm map) with the catalog of targetclass earthquakes. By contrast, f is independent of target earthquakes; it depends solely on the forecast model (T_{q}, R, T_{a}) and the background seismicity up to the date of alarm issue T_{2}. Figure 3 shows f as a function of T_{2}, for representative models discussed in the present paper: T_{q} = 9, 10, 11, and 12 years, with T_{a} and R, respectively, fixed at 7 years and 60 km. We see f is smaller for higher T_{q}, where models become more selective in recognizing anomalies. Additionally, note that f is quite large (> 10%), meaning that the present paper deals with quite vague forecasts.
Following Zechar and Jordan (2008), we define the probability gain G as
As mentioned earlier (“Introduction” section), f is the expectancy of r under the null hypothesis that the presently defined quiescence anomalies are not relevant to the subsequent occurrence of earthquakes having M_{w} ≥ 7.5. Therefore, the righthand side of Eq. (3) represents the improvement ratio of the prediction rate, whereas the standard definition of probability gain is the improvement ratio of the success rate (i.e., the enhancement of the probability density of earthquake occurrence in the alerted spacetime) (e.g., Aki 1981). However, it can be shown mathematically that the two improvement ratios necessarily coincide (Nakatani 2020). Incidentally, note that the theoretical upper limit for G is 1/f. As seen in Fig. 3, f was > 10% in the present study, such that G cannot exceed 10, even if none of the target earthquakes are missed.
The null hypothesis of no correlation is equivalent to the proposition that the true value of G is unity. As shown in “Results and discussion” section, many of our forecast models exhibited G > 1, suggesting an earthquakepreceding tendency. To check if the tendency is statistically significant, we will calculate the pvalue, which is the probability that G (or r) equal to or higher than the observed occurs in the random forecasts having f equal to that of the forecast (i.e., alarm maps) being scrutinized. The formula of binomial probability giving the pvalue (e.g., Zechar and Jordan 2008) is
In the present study, where we can use a maximum of eight target earthquakes for statistical testing, we tentatively adopt p < 5% as the criterion for statistical significance, though the choice of threshold is a subjective matter after all.
Results and discussion
Experiment 1 (main experiment)
As Mulargia (1997) pointed out, retrospective optimization of a forecasting method is prone to overfitting, which would lead to overrating of the method. We therefore illustrate the method's robustness by showing the results of all 210 forecast models produced in retrospective optimization instead of judging the significance solely according to the lowest pvalue among them. In our optimization, we search with respect to all three main aspects that generally constitute a precursorbased forecast; T_{q} (7 ≤ T_{q} ≤ 13 years, at 1year intervals) regulates the recognition of the suspected precursory phenomenon, while R (50 ≤ R ≤ 100 km, at 10km intervals) and T_{a} (4 ≤ T_{a} ≤ 8 years, at 1year intervals), respectively, regulate the spatial and temporal association between suspected precursors and subsequent earthquakes. Figure 4 is the receiveroperatingcharacteristic (ROC) diagram, constructed by plotting (f, r) for all 210 models. All eight target (7.5 ≤ M_{w} < 8.5) earthquakes (EQ#1–8 in Table 1) are considered in calculating r for each model.
Figure 4 shows that r is higher than f for almost all models, suggesting that the presently defined (“Anomaly detection” section) longterm quiescence provides information that helps discriminate spacetime with heightened risk. At the same time, G (= r/f) is only ~ 2; the information provided by the quiescence is so weak that the risk in the alerted spacetime is only twice the secular level.
In Fig. 4, we use different marker shapes according to the model's pvalue. Furthermore, we use large red markers for the optimal models yielding p < 5% and r > 0.5. Table 2 lists these optimal models. (In Experiment 1, all models that yielded p < 5% also yielded r > 0.5; however, the condition r > 0.5 matters for consistency with corresponding plots in other experiments shown later.) At face value, p < 5% for these models implies the statistical significance of the earthquakepreceding tendency. Below, we look into the characteristics of optimal models for Experiment 1 (Table 2), in an attempt to elucidate the characteristics of precursory longterm quiescence. Note that Experiment 1 is the main and base experiment of the present paper.
Ten of the 15 optimal models use T_{q} = 9 years (Table 2, models 1–1 through 1–10), while R ranges 50–80 km and T_{a} ranges 4–8 years. These 10 models yield high r; eight achieve r = 8/8 and two achieve r = 7/8. For the 10 models, f ranges 50%–70% and G ranges 1.5–1.8. The remaining five models (Table 2, models 1–11 through 1–15) use stricter criteria for quiescence anomalies, T_{q} = 10–12 years. As a result, EQ#1 and #6 become surprise events, lowering r to 6/8. For model 1–15, which adopts T_{q} = 12 years, EQ#5 also becomes a surprise event, further lowering r to 5/8. Nevertheless, these five models with higher T_{q} (i.e., 10–12 years) achieve higher G of 1.9–2.3 because f is much improved (i.e., slashed) to 27%–39% thanks to the stringent selection of quiescence anomalies. The size of the alerted spacetime (R, T_{a}) around an anomaly is generally larger than that for the 10 models with T_{q} = 9 years, but not by much.
As seen above, there is a tradeoff between the two performance demands; that is, higher r and lower f. The pvalue strongly depends on r and f in the relevant range (Fig. 5). The balance between the two seems to be regulated by (T_{q}, R, T_{a}) in a selfevident sense. It is thus hardly meaningful to ask which model is the best. Instead, we emphasize that these wellperforming models (Table 2) lie coherently in the model space (T_{q}, R, T_{a}). This implies that the favorable performance of the optimized models (Table 2) originates from decent optimization reflecting universal properties of mechanisms underlying the earthquakepreceding tendency of the longterm quiescence, rather than originating from the use of deliberate, complex algorithms that do whatever to score well. Such gerrymandering is unlikely because our forecast algorithm, including adjustment through (T_{q}, R, T_{a}), is straightforward. However, the present trial forecast is calibrated with only eight earthquakes. Hence, overfitting, if not gerrymandering, can readily occur. We will check this point by reporting on crossvalidation experiments in “Experiments 2 and 3 (crossvalidation)” section.
Experiments 2 and 3 (crossvalidation)
Although they look good, our results so far could benefit from overfitting as cautioned already; good performance exhibited by the retrospectively optimized forecast models does not assure good performance for future data (e.g., Mulargia 1997). We thus conduct crossvalidation experiments by dividing the study period of Experiment 1 into two, using one period to train models and the other period to evaluate elite models in the learning period. Experiment 2 uses the later period for training and the earlier period for evaluation, whereas Experiment 3 uses the earlier period for training and the later period for evaluation. In both experiments, we explore the same range of the parameter space (T_{q}, R, T_{a}) as in Experiment 1, amounting to 210 models. The target class of forecast remains 7.5 ≤ M_{w} < 8.5, the same as in Experiment 1. Unfortunately, our low number of target earthquakes does not allow us to conduct crossvalidation experiments separately for earthquakes having M_{w} < 8.0 and M_{w} ≥ 8.0, though this would be desirable given the plausible dependence of precursory quiescence on the main shock magnitude (Additional file 2).
Experiment 2
In Experiment 2, models are first optimized through trial forecasts based on subcatalogs in the second half (1995.0 ≤ T_{2} < 2015.0), for which four earthquakes having 7.5 ≤ M_{w} < 8.5 (EQ#5–8) are available as forecast targets. Figure 6 shows the performance in this learning period for all 210 models. We draw common reference contours (blue curves) to directly compare the strength of the apparent correlations seen in Figs. 4 and 6. The three contours correspond to contours of (f, r) that would yield p = 5%, 10%, and 15% provided that eight target earthquakes were available as in Experiment 1. The marker shape convention is the same as that in Fig. 4. This ROC diagram, mostly implying G > 1, is generally similar to that of Experiment 1, but no model achieves p < 5% owing to the limited number of target earthquakes (i.e., four target earthquakes). Thus, formally, our crossvalidation attempt has failed already in the learning period; the number of available earthquakes (four in the learning period, four in the evaluation period) is too few to attempt crossvalidation of the weak (G < 10) earthquakepreceding tendency.
Nonetheless, we proceed by choosing eight models that achieve p < 15% and r > 0.5 in the learning period (indicated by large red symbols in Fig. 6) as optimal models (Table 3). Performances of these learningperiod elites are then evaluated by making forecasts based on the subcatalogs in the first half (1988.0 ≤ T_{2} < 1995.0), the evaluation period of Experiment 2. All four target earthquakes (EQ#1–4) in the first half are used for evaluation. Figure 7 shows the ROC diagram for this evaluation period. Note that marker shapes in Fig. 7 reflect the pvalues of the respective models in the learning period, and large red symbols represent the learningperiod elites listed in Table 3. Table 4 shows how these learningperiod elites perform in the evaluation period.
In both the learning (Fig. 6) and evaluation (Fig. 7) periods, r is higher than f for most models, implying skill better than random forecasts. We thus examine the performance in the two periods in detail to seek signatures of overfitting.
A comparison of Tables 3 and 4 shows that N_{on} is lower by 1 in the evaluation period for most models. In closer examination, we see that almost all the models miss EQ#1 in the evaluation period, worsening the pvalue. We now explain what happened concretely here. All four target earthquakes in the learning period occurred within 4 years of the quiescence of T_{q} = 9 years. Hence, most of the optimized models adopt relatively short T_{a} of 4 or 5 years to avoid excessive f. In contrast, EQ#1, in the evaluation period, occurred 6 years after a 9year quiescence and is missed by all the learningperiod elites except model 2–2, which adopts T_{a} = 6 years. This can be said to be overfitting, if mild.
However, when we look at G, learningperiod elites perform equally well (G ~ 2) in the evaluation period (red markers in Fig. 8a). This is largely explained by the fact that f was generally lower in the evaluation period, being 70–80% of the value in the learning period (Tables 3 and 4). This cancels the adverse effect of the ~ 25% drop in r. Overall, we may say that there is no severe overfitting in Experiment 2.
Furthermore, the (T_{q}, R, T_{a}) range of the optimal models in Experiment 1, the main experiment of the present paper, resembles that in Experiment 2 (Table 3). As described above, the optimal models in Experiment 2 do not seem to involve severe overfitting like in Experiment 3 (“Experiment 3” section). Hence, the favorable results (G ~ 2) of Experiment 1, even though the same data are used for learning and evaluation, are probably not due to overfitting. We thus surmise (but not prove, unfortunately) that the favorable performance (G ~ 2) of Experiment 1 is likely achieved by decent optimization and it may be taken as an encouraging result.
Experiment 3
We conduct another crossvalidation experiment, Experiment 3, where the learning period and evaluation period are flipped relative to those in Experiment 2. Figure 9 is the ROC diagram for the learning period (i.e., the first half) of Experiment 3. For consistency with Experiment 2, we regard models that achieve p < 15% and r > 0.5 (Table 5) as optimal and highlight them using large red symbols, though quite a few (nine) of them achieve p < 5%. Moreover, the ROC for the learning period of Experiment 3 (Fig. 9) looks far better than those for Experiments 1 and 2, yielding higher G often exceeding 3. Table 5 shows that models with higher T_{q} perform particularly well in Experiment 3. In other words, this superb learningperiod performance of Experiment 3 is due to many of the target earthquakes being preceded by particularly prolonged quiescence, allowing the slashing of f through stringent anomaly selection. This strategy does not sacrifice r in the learning period because as many as three earthquakes (EQ#2, 3, 4) were preceded by quiescence anomalies as strong as T_{q} ≥ 12 years (see models 3–30 through 3–39).
Figure 10 is the ROC diagram of Experiment 3 for the evaluation period. Large red symbols highlight the learningperiod elites (Table 5). Table 6 lists their performances in the evaluation period, which are considerably worse than those in the leaning period. Figure 8b (red markers) compares G of respective learningperiod elites between learning and evaluation periods; all these models show a drop, often severe, in G during the evaluation period, implying serious overfitting in Experiment 3, in contrast with the case of Experiment 2 (Fig. 8a).
Let us take a closer look. Results (red markers) in Fig. 8b may be divided into three clusters (groups A, B, and C in Fig. 8b). GroupA models have the most severe performance drop; G in the evaluation period is often below unity. In terms of the groupaveraged G, Group A is the best in the learning period but is the worst in the evaluation period. In contrast, Group C is the worst in the learning period but is the best in the evaluation period. We examine Tables 5 and 6 in detail below to find specific mechanisms of overfitting in Experiment 3.
All GroupA models adopt a high value of T_{q}, as high as 12 years, meaning stringent anomaly selection, which helps slash f. Additionally, values of T_{a} are mostly 7 years for GroupA models and shorter than T_{a} of 8 years adopted by the other models with T_{q} = 12 years, which belong to Group B. Shorter T_{a} also helps suppress f. In the learning period, three (out of four) earthquakes occurred within 7 years of strong quiescence anomalies lasting at least 12 years. Hence, GroupA models could achieve good r (3/4 = 75%) despite their restrictive alerting policy. However, in the evaluation period, only one earthquake (out of four) was preceded by such strong quiescence, resulting in r = 1/4 and G ~ 1. Similar overfitting occurred in GroupB models, which mostly adopt the same stringent T_{q} of 12 years as do GroupA models. The performance drop is less severe because, thanks to the adoption of T_{a} = 8 years instead of 7 years, GroupB models did not miss EQ#7 that occurred 7.5 years after the recognition of the quiescence of T_{q} = 12 years, so that r in the evaluation period is 2/4 instead of 1/4 for Group A.
Note the above difference in quiescence behavior between the learning and evaluation periods cannot be ascribed to the plausible dependence of the quiescence duration on the main shock magnitude (Additional file 2), because only one (EQ#3) of the three earthquakes having M_{w} ≥ 8.0 occurred in the learning period. Instead, the superb performance in the learning period of Experiment 3 arises from pure luck that the quiescence anomalies preceding two (EQ#2 and #4) of the three earthquakes having M_{w} < 8 in the learning period happened to have particularly prolonged durations of at least 12 years. We thus conclude that most (i.e., groups A and B) learningperiod elites commit a classical overfitting. They became too picky in anomaly detection because they mistook mere apparent features introduced by natural fluctuation as a universal property of the precursory quiescence.
Although G drops in the evaluation period also for GroupC models, we do not recognize obvious overfitting. GroupC models, being the least selective among the learningperiod elites of Experiment 3, have r = 3/4 or 4/4 in both learning and evaluation periods. The observed mild drop in G is mostly attributable to higher f in the evaluation period (Fig. 3).
In conclusion, Experiment 3 involves severe overfitting. This is pretty much expected, given the limited number of learning data (i.e., only four target earthquakes in the learning period). Specifically, overfitting occurred because our optimization tried to slash f too aggressively, using apparent features that arose by pure luck. Of course, there is no way to tell that these features are not a real property until cases lacking these features are learned; optimization using a sufficient number of target earthquakes would ease the problem.
We, however, emphasize that the overly stringent policy in anomaly detection adopted by Experiment 3 was not preferred in Experiment 1, the main experiment of the present paper. Its optimal models (Table 2) mostly use T_{q} = 9 years, and only one of the 15 adopts T_{q} = 12 years. Given the insights from Experiment 3, we may say high T_{q} was not much preferred in Experiment 1 because securing a high r is more important than suppressing f to achieve a sufficiently small p, which is our primary criterion in choosing optimal models. We therefore believe that the modest (G ~ 2) but still favorable performance of Experiment 1 is unlikely to be a ghost arising from overfitting.
Although the present crossvalidation tests failed to confirm it formally, we would say, as an optimistic conjecture, that there is a good chance that the earthquakepreceding tendency of longterm quiescence will be proven one day by having a higher number of target earthquakes for trial forecasts, maybe several times more. As an example, if the number of target earthquakes doubles in the future (i.e., 16 in total), the learning and evaluation periods can use eight earthquakes each. With eight earthquakes, a pvalue less than 5% can be shown for the modest (G ~ 2) performance we saw in Experiments 1 and 2 and for group C in Experiment 3. To give some numbers, p < 5% is attained by G ≥ 1.75 (r ≥ 7/8) if f is 50%, and by G ≥ 2.5 (r ≥ 5/8) if f is 25%.
Summary and conclusions
We made simple trial forecast experiments where we issue alarms valid for time T_{a} to regions within a distance R of the location where longterm quiescence is detected. We defined a quiescence anomaly as the absence of nearby earthquakes having m_{b} ≥ 5 for a time duration of at least T_{q}. On the basis of these anomalies, alarm maps were made throughout the study area (i.e., the northwestern margin of the Pacific Plate) by exhaustively scanning the studied spacetime. In each of our forecast experiments (Experiments 1–3), alarm maps, made with a range of (T_{q}, R, T_{a}) values, were evaluated using a ROC diagram. Our main experiment (Experiment 1), targeting all eight earthquakes having 7.5 ≤ M_{w} < 8.5, found a range of forecast models exhibiting pvalues less than 5%, and G of ~ 2, supporting the existence of an earthquakepreceding tendency of longterm quiescence.
These favorable results of Experiment 1, however, could be a ghost owing to overfitting because we used the same data for training and evaluation. We therefore attempted crossvalidation using four of the eight earthquakes to train forecast models and the other four earthquakes to evaluate the optimal models obtained from the learning period. We conducted two crossvalidation experiments (Experiments 2 and 3) by flipping the datasets for training and evaluation. In Experiment 2, the performance was similar for learning and evaluation periods. Additionally, optimal models and their G values were similar to those in Experiment 1. The above implies that the apparent success (G ~ 2) in Experiment 1 is unlikely attributable to overfitting. However, we failed in formal crossvalidation as none of the models achieved p < 5% in the learning period of Experiment 2. This is not surprising, given that statistical power with only four target earthquakes available is likely insufficient to demonstrate the significance of the modest earthquakepreceding tendency of G ~ 2, if any, exhibited in Experiment 1.
In the other crossvalidation experiment (Experiment 3), many models, in the learning period, yielded p < 5% and G much higher than that in our base experiment (Experiment 1). However, this was a typical overfitting effect; the models fared miserably in the evaluation period. Close examination of Experiment 3 has revealed that the overfitting occurred because the models became too picky in anomaly detection because they mistook fortuitous features only seen in the learning period as real properties of the precursory quiescence. Fortunately, that excessive selectiveness was not preferred in the optimal models (Table 2) in our main experiment (Experiment 1), again supporting the likelihood that the favorable results of Experiment 1 are real, though we admit that we failed in formal crossvalidation.
In the course of the present study, we have also found that G is much higher if the forecasts target only earthquakes having M_{w} ≥ 8 (Experiment 5 in Additional file 2); G > 5 and p < 5% were easily achieved. However, this was a result obtained with only three target earthquakes, and we have nothing to say against the possibility of overfitting. Meanwhile, forecasts (Experiment 4 in Additional file 2) targeting only five smaller earthquakes (7.5 ≤ M_{w} < 8) found a correlation weaker than that in Experiment 1, which targeted all earthquakes of 7.5 ≤ M_{w} < 8.5, and a pvalue less than 5% was not obtained. Hence, we have no direct basis for claiming that the quiescence's earthquakepreceding tendency holds for earthquakes of 7.5 ≤ M_{w} < 8 as well. However, the (T_{q}, R, T_{a}) range for optimal models in Experiment 4 was similar to that in Experiment 1. Hence, there remains a reasonable chance that longterm quiescence has a tendency, if weaker, to precede even these smaller earthquakes.
Availability of data and materials
The datasets used and/or analyzed in the current study are available from the corresponding author on reasonable request.
References
Aki K (1981) A probabilistic synthesis of precursory phenomena. In: Simpson DW, Richards PG (eds) Earthquake Prediction (Maurice Ewing Series 4). American Geophysical Union, Washington, D.C., pp 566–574
Bird P (2003) An updated digital model of plate boundaries. Geochem Geophys Geosyst 4(3):1027. https://doi.org/10.1029/2001GC000252
Hardebeck JL, Felzer KR, Michael AJ (2008) Improved tests reveal that the accelerating moment release hypothesis is statistically insignificant. J Geophys Res 113:B08310. https://doi.org/10.1029/2007JB005410
Inouye W (1965) On the seismicity in the epicentral region and its neighborhood before the Niigata earthquake. Kenshinjiho (Quarterly Journal of Seismology) 29:139–144 (in Japanese)
Kanamori H (1981) The nature of seismicity patterns before large earthquakes. In: Simpson DW, Richards PG (eds) Earthquake Prediction (Maurice Ewing Series 4). American Geophysical Union, Washington D. C., pp 1–19
Katsumata K (2011) Precursory seismic quiescence before the M_{w} = 8.3 Tokachioki, Japan earthquake on 26 September 2003 revealed by a reexamined earthquake catalog. J Geophys Res 116:B10307. https://doi.org/10.1029/2010JB007964
Katsumata K (2017a) Longterm seismic quiescences and great earthquakes in and around the Japan subduction zone between 1975 and 2012. Pure Appl Geophys 174:2427–2442. https://doi.org/10.1007/s0002401614158
Katsumata (2017b) Longterm seismic quiescence before shallow great earthquakes with M_{w}8.0 or larger between 1990 and 2014. In: Abstracts of JpGUAGU Joint Meeting 2017, Makuhari Messe, Japan, 20–25 May 2017. https://confit.atlas.jp/guide/event/jpguagu2017/subject/SSS14P06/date?cryptoId=
KeilisBorok VI, Kossobokov VG (1990) Premonitory activation of earthquake flow: algorithm M8. Phys Earth Planet Inter 61:73–83. https://doi.org/10.1016/00319201(90)90096G
Michael AJ (1997) Testing prediction methods: Earthquake clustering versus the Poisson model. Geophys Res Lett 24:1891–1894. https://doi.org/10.1029/97GL01928
Michael AJ (2014) How complete is the ISCGEM global earthquake catalog? Bull Seismol Soc Am 104:1829–1837. https://doi.org/10.1785/0120130227
Mogi K (1969) Some features of recent seismic activity in and near Japan (2), Activity before and after great earthquakes. Bull Earthquake Res Inst Tokyo Univ 47:395–417
Mulargia F (1997) Retrospective validation of the time association of precursors. Geophys J Int 131:500–504. https://doi.org/10.1111/j.1365246X.1997.tb06594.x
Nagao T, Takeuchi A, Nakamura K (2011) A new algorithm for the detection of seismic quiescence: introduction of the RTM algorithm, a modified RTL algorithm. Earth Planets Space 63:315–324. https://doi.org/10.5047/eps.2010.12.007
Nakatani M (2020) Evaluation of phenomena preceding earthquakes and earthquake predictability. J Disaster Res 15:112–143. https://doi.org/10.20965/jdr.2020.p0112
Ogata Y (1992) Detection of precursory relative quiescence before great earthquakes through a statistical model. J Geophys Res 97:19845–19871. https://doi.org/10.1029/92JB00708
Ogata Y (2001) Increased probability of large earthquakes near aftershock regions with relative quiescence. J Geophys Res 106:8729–8744. https://doi.org/10.1029/2000JB900400
Ohtake M, Matsumoto T, Latham GV (1977) Seismicity gap near Oaxaca, Southern Mexico as a probable precursor to a large earthquake. Pure Appl Geophys 115:375–386. https://doi.org/10.1007/BF01637115
Page MT, van der Elst N, Hardebeck J, Felzer K, Michael AJ (2016) Three ingredients for improved global aftershock forecasts: tectonic region, timedependent catalog incompleteness, and intersequence variability. Bull Seismol Soc Am 106:2290–2301. https://doi.org/10.1785/0120160073
Reasenberg PA, Matthews MV (1988) Precursory seismic quiescence: a preliminary assessment of the hypothesis. Pure Appl Geophys 126:373–406. https://doi.org/10.1007/BF00879004
Scholz CH (2019) The mechanics of earthquakes and faulting, 3rd edn. Cambridge University Press, Cambridge
Sobolev GA, Tyupkin YS (1997) Lowseismicity precursors of large earthquakes in Kamchatka. Volcanol Seismol 18:433–446
Storchak DA, Harris J, Brown L, Lieser K, Shumba B, Verney R et al (2017) Rebuild of the Bulletin of the International Seismological Centre (ISC), part 1: 1964–1979. Geosci Lett 4:32. https://doi.org/10.1186/s405620170098z
Utsu T (1957) Magnitude of earthquakes and occurrence of their aftershocks. Zisin J Seismol Soc Jpn 10:35–45. https://doi.org/10.4294/zisin1948.10.1_35 (in Japanese)
Utsu T (1968) Seismic activity in Hokkaido and its vicinity. Geophys Bull Hokkaido Univ 20:51–75. https://doi.org/10.14943/gbhu.20.51 (in Japanese)
Wang Q, Schoenberg FP, Jackson DD (2010) Standard errors of parameter estimates in the ETAS model. Bull Seismol Soc Am 100:1989–2001. https://doi.org/10.1785/0120100001
Wessel P, Smith WHF (1991) Free software helps map and display data. Eos Trans AGU 72:445–446. https://doi.org/10.1029/90EO00319
Wiemer S, Wyss M (1994) Seismic quiescence before the Landers (M=7.5) and Big Bare (M=6.5) 1992 earthquakes. Bull Seismol Soc Am 84:900–916
Wiemer S, Wyss M (2000) Minimum magnitude of completeness in earthquake catalogs: examples from Alaska, the western United States, and Japan. Bull Seismol Soc Am 90:859–869. https://doi.org/10.1785/0119990114
Zechar JD, Jordan TH (2008) Testing alarmbased earthquake predictions. Geophys J Int 172:715–724. https://doi.org/10.1111/j.1365246X.2007.03676.x
Zhuang J, Ogata Y, VereJones D (2002) Stochastic declustering of spacetime earthquake occurrences. J Am Stat Assoc 97:369–380. https://doi.org/10.1198/016214502760046925
Zhuang J, Chang CP, Ogata Y, Chen YI (2005) A study on the background and clustering seismicity in the Taiwan region by using point process models. J Geophys Res 110:B05S18. https://doi.org/10.1029/2004JB003157
Acknowledgements
We thank Naoki Uchida, Takeshi Sagiya, Andrew Michael, and an anonymous reviewer for making valuable comments and Glenn Pennycook, MSc, from Edanz Group (https://enauthorservices.edanz.com/ac) for editing a draft of this manuscript. The source code of the program used for the spacetime ETAS model and stochastic declustering is available at http://bemlar.ism.ac.jp/zhuang/software.html. GMTSYSTEM (Wessel and Smith 1991) was used for mapping data.
Funding
This study was partly supported by the Ministry of Education, Culture, Sports, Science and Technology (MEXT) of Japan, under “The Second Earthquake and Volcano Hazards Observation and Research Program (Earthquake and Volcano Hazard Reduction Research)”. This study was also partly supported by the Tokio Marine Kagami Memorial Foundation.
Author information
Affiliations
Contributions
KK developed the detection method for seismic quiescence, created the program for analysis, and performed calculations. MN developed the statistical evaluation method for the seismic quiescence and was a major contributor in writing the manuscript. Both authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Additional file 1: Figure S1.
Temporal change in the minimum magnitude of completeness (M_{c}). Figure S2. ETAS parameters of each subcatalog. Figure S3. Seismicity with m_{b} ≥ 5.0 before and after great earthquakes with M_{w} ≥ 8.0.
Additional file 2.
It describes the section “Experiments 4 and 5 (dependence on the main shock magnitude)”, Tables S1 and S2, and Figures S4 and S5.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Katsumata, K., Nakatani, M. Testing the seismic quiescence hypothesis through retrospective trials of alarmbased earthquake prediction in the Kurile–Japan subduction zone. Earth Planets Space 73, 100 (2021). https://doi.org/10.1186/s4062302101418z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s4062302101418z
Keywords
 Earthquake prediction
 Seismic quiescence
 Crossvalidation
 Kurile–Japan subduction zone