Читать книгу Statistical Methods and Modeling of Seismogenesis - Eleftheria Papadimitriou - Страница 12
1.1. Introduction
ОглавлениеThe basis of statistical seismology is the probabilistic distribution of earthquake parameters. Unfortunately, the models of probabilistic distributions of earthquake parameters are mostly unknown, and those which are thought to be known are often disproved by more thorough testing.
On the other hand, increasing quality and decreasing prices of seismic monitoring systems, increasing density of seismic networks and the development of event picking tools cause the current seismic catalogs to be large and rising. This situation opens up the area for model-free approaches to probabilistic estimation and statistical inferences, which to be accurate, require considerable sample sizes.
The kernel density estimation is a model-free estimation of probability functions of continuous random variables. The estimation is carried out solely from sample data.
There is no attempt here to comprehensively present the kernel density estimation method. A detailed description and the discussion of the method can be found in the textbooks by Silverman (1986), Wand and Jones (1995), Scott (2015) and in a multitude of high-level research papers. The method is also implemented in Matlab, R and Python, among others, and in many statistical packages. In this chapter, the author presents how this method has been applied to selected problems of seismology. This presentation begins with a short and simplified theoretical introduction. The kernel density estimation has been fast developing from both theoretical and practical sides; hence, the techniques presented here do not aspire to be optimal. There is plenty of space for future modifications and developments.
In the case of a univariate random variable X, and a constant kernel, the kernel estimator of the actual probability density function (PDF) of X, fX(x), takes the form:
[1.1]
where {xi}, i = 1,.., n is the sample data, K(•) is the kernel function, which is a PDF symmetric about zero, and h is the bandwidth, whose value decides how much smoothing has been applied to the sample data.
In the presented seismological applications of the kernel density estimation, we use the normal kernel function:
For this kernel function, the kernel estimates of PDF and the cumulative distribution function (CDF, ) are, respectively:
[1.4]
where Ф(·) is the CDF of standard normal distribution.
As it will follow, the kernel method is used, among others, to estimate the magnitude distribution, whose distribution is exponential-like or light-tailed. Because of that, the tail values are sparse in a sample. Such sparsity can result in spurious irregularities in the estimate on tails, if a constant bandwidth is used. We can alleviate this problem by using an adaptive kernel with variable bandwidth. Because the estimates of magnitude distribution functions serve in the probabilistic seismic hazard analysis (PSHA), the quality of the estimate in the tail part is of utmost importance.
The adaptive kernel estimate of PDF used here takes the form:
and the estimate of CDF is:
[1.6]
where ωi are the local weights widening the bandwidth associated with the data points from the range where the data is sparse. The weights take the form:
where is the constant kernel PDF estimate in xi, equation [1.3], and g = The way in which the kernel estimate of PDF is composed and the difference between the estimation with a constant and a variable bandwidth are presented in Figure 1.1.
Figure 1.1. Buildup of the kernel estimate of PDF. Normal kernel functions (thick dashed lines) are associated with data points (triangles at abscissa). The PDF estimate (thin solid line) is the sum of the kernel functions. Whereas in the estimation with constant bandwidth, (a), all kernel functions are the same, in the estimation with variable bandwidth, (b), the kernels are wider in the regions of sparse data. For a color version of this figure, see www.iste.co.uk/limnios/statistical.zip
The final shape of the density estimate strongly depends on the value of h, which is demonstrated in the example in Figure 1.2. In the applications of the kernel estimators presented here, the criterion to select the bandwidth, h, is the minimizer of the mean integrated square error (MISE): where f is the actual (unknown) PDF, is its kernel estimate and E[•] denotes the expected value. Starting from MISE, Silverman (1986) derived the simplified score function of the form:
[1.8]
where K(•) is the kernel function, and Kijko et al. (2001) showed that for the normal kernel function, [1.2], the score function becomes:
[1.9]
and the bandwidth that minimizes M1(h) is the root of the equation:
When the random variable, for which the distribution functions are to be estimated, X, is defined over a finite or semi-finite interval, or its density is sharply zeroed outside a finite or semi-finite interval, the estimation of the distribution functions is modified according to Silverman (1986). Suppose that either X ∈ [x*, x*], or the PDF, fX(x), is not continuous in x* and x*, and fX(x) = 0 for x < x* and x > x*. To get the kernel estimates of PDF and CDF, the original data sample, {xi}, i = 1,.., n, is mirrored symmetrically around x* and x* resulting in the sample {2x* − xi, xi, 2x* − xi}, i = 1,.. n. Based on this sample, a density is estimated and the desired estimates of PDF and CDF are:
Figure 1.2. The dependence of density estimate on bandwidth, h. The actual distribution was the superposition of two normal distributions, 0.7 × NORM(0,1) + 0.3 × NORM(2,2). The exact PDF is presented in (a). The estimates were built from 500-element samples drawn from the actual distribution, using the normal kernel and constant bandwidths: (b) h = 0.1, (c) h = 0.25 and (d) h = 0.5. For a color version of this figure, see www.iste.co.uk/limnios/statistical.zip
When the interval [x*, x*] is semi-finite, then the original sample is mirrored around and the sample or the sample is used to estimate The desired estimates of PDF and CDF are:
In connection with the above-mentioned problems with probabilistic functional models of earthquake parameters, the flexibility of kernel density estimation makes it a perfect tool for modeling the probabilistic distributions of such parameters. In particular, the kernel estimation method responds well to the needs of the PSHA.
The mathematical model in PSHA can be formulated, e.g., as the following expression for the probability that the ground motion amplitude parameter, amp, at the point (x0, y0), during D time units will exceed the value a(x0, y0):
Pr[amp(x0, y0) ≥ a(x0, y0), D] =
[1.13]
where r(x0, y0) is the epicentral distance of an earthquake to the receiving point (x0, y0), M is the earthquake magnitude, N(D) is the number of earthquakes in D, Pr[amp(x0, y0) ≥ a(x0, y0)|M, r] is the probability that amp will exceed a due to the earthquake of magnitude, M, being distanced from the receiving point of r, fr is the PDF of epicentral distance, f(M|N(D) ≠ 0) is the probability density of earthquake magnitude, M, conditional upon earthquake occurrence in D and ℳ is its domain. fr is straightforwardly linked to the two-dimensional probability distribution of earthquake epicenters, fxy(x, y). fxy(x, y) and f(M|N(D) ≠ 0) represent the properties of seismic source (source effect), and Pr[amp(x0, y0) ≥ a(x0, y0)|M, r] represents the properties of the vibration transmission from the source to the receiving point (path and site effects).
The conditional probability of source magnitude, f(M|N(D) ≠ 0), can be evaluated from the probability mass function (PMF) of the number of earthquakes in D time units, Pr[N(D)=n], and the unconditional PDF, fM, and the CDF, FM, of magnitude:
Under the assumption that the seismic occurrence process is Poissonian, equation [1.14] becomes:
[1.15]
where λ is the rate of earthquake occurrence.
The seismic hazard is often only characterized by its source component, namely, by the probability that an earthquake of a magnitude greater than or equal to M occurs within D time units (the exceedance probability), R(M, D), or the expected average time between successive occurrences of earthquakes with magnitudes ≥ M (the mean return period), T(M). The exceedance probability reads:
If the seismic occurrence process is considered as Poissonian then:
and