Читать книгу Seismic Reservoir Modeling - Dario Grana - Страница 19
1.3.2 Multivariate Distributions
ОглавлениеIn many practical applications, we are interested in multiple random variables. For example, in reservoir modeling, we are often interested in porosity and fluid saturation, or P‐wave and S‐wave velocity. To represent multiple random variables and measure their interdependent behavior, we introduce the concept of joint probability distribution. The joint PDF of two random variables X and Y is a function fX,Y : ℝ × ℝ → [0, +∞] such that 0 ≤ fX,Y(x, y) ≤ +∞ and .
The probability P(a < X ≤ b, c < Y ≤ d) of X and Y being in the domain (a, b] × (c, d] is defined as the double integral of the joint PDF:
(1.21)
Given the joint distribution of X and Y, we can compute the marginal distributions of X and Y, respectively, as:
(1.22)
(1.23)
In the multivariate setting, we can also introduce the definition of conditional probability distribution. For continuous random variables, the conditional PDF of X∣Y is:
where the joint distribution fX,Y(x, y) is normalized by the marginal distribution fY(y) of the conditioning variable. An analogous definition can be derived for the conditional distribution fY∣X(y) of Y∣X. All the definitions in this section can be extended to any finite number of random variables.
An example of joint and conditional distributions in a bivariate domain is shown in Figure 1.4. The surface plot in Figure 1.4 shows the bivariate joint distribution fX,Y(x, y) of two random variables X and Y centered at (1, −1). The contour plot in Figure 1.4 shows the probability density contours of the bivariate joint distribution as well as the conditional distribution fY∣X(y) for the conditioning value x = 1 and the marginal distributions fX(x) and fY(y).
The conditional probability distribution in Eq. (1.24) can also be computed using Bayes' theorem (Eq. 1.8) as:
Figure 1.4 Multivariate probability density functions: bivariate joint distribution (surface and contour plots), conditional distribution for x = 1, and marginal distributions.
Figure 1.5 shows an example where the uncertainty in the prior probability distribution of a property X is relatively large, and it is reduced in the posterior probability distribution of the property X conditioned on the property Y, by integrating the information from the data contained in the likelihood function. In seismic reservoir characterization, the variable X could represent S‐wave velocity and the variable Y could represent P‐wave velocity. If a direct measurement of P‐wave velocity is available, we can compute the posterior probability distribution of S‐wave velocity conditioned on the P‐wave velocity measurement. The prior distribution is assumed to be unimodal with relatively large variance. By integrating the likelihood function, we reduce the uncertainty in the posterior distribution.
Figure 1.5 Bayes' theorem: the posterior probability is proportional to the product of the prior probability and the likelihood function.
We can also extend the definitions of mean and variance to multivariate random variables. For the joint distribution fX,Y(x, y) of X and Y, the mean μX,Y = [μX, μY]T is the vector of the means μX and μY of the random variables X and Y. In the multivariate case, however, the variances of the random variables do not fully describe the variability of the joint random variable. Indeed, the variability of the joint random variable also depends on how the two variables are related. We define then the covariance σX,Y of X and Y as:
The covariance is a measure of the linear dependence between two random variables. The covariance of a random variable with itself is equal to the variance of the variable. Therefore, and . The information associated with the variability of the joint random variable is generally summarized in the covariance matrix ∑X,Y:
(1.27)
where the diagonal of the matrix includes the variances of the random variables, and the elements outside the diagonal represent the covariances. The covariance matrix is symmetric by definition, because σX,Y = σY,X based on the commutative property of the multiplication under the integral in Eq. (1.26). The covariance matrix of a multivariate probability distribution is always positive semi‐definite; and it is positive definite unless one variable is a linear transformation of another variable.
We then introduce the linear correlation coefficient ρX,Y of two random variables X and Y, which is defined as the covariance normalized by the product of the standard deviations of the two random variables:
(1.28)
The correlation coefficient is by definition bounded between −1 and 1 (i.e. −1 ≤ ρX,Y ≤ 1), dimensionless, and easy to interpret. Indeed, a correlation coefficient ρX,Y = 0 means that X and Y are linearly uncorrelated, whereas a correlation coefficient |ρX,Y| = 1 means that Y is a linear function of X. Figure 1.6 shows four examples of two random variables X and Y with different correlation coefficients. When the correlation coefficient is ρX,Y = 0.9, the samples of the two random variables form an elongated cloud of points aligned along a straight line, whereas, when the correlation coefficient is ρX,Y ≈ 0, the samples of the two random variables form a homogeneous cloud of points with no preferential alignment. A positive correlation coefficient means that if the random variable X increases, then the random variable Y increases as well, whereas a negative correlation coefficient means that if the random variable X increases, then the random variable Y decreases. For this reason, when the correlation coefficient is ρX,Y = −0.6, the cloud of samples of the two random variables approximately follows a straight line with negative slope.
If two random variables are independent, i.e. fX,Y(x, y) = fX(x)fY(y), then X and Y are uncorrelated. However, the opposite is not necessarily true. Indeed, the correlation coefficient is a measure of linear correlation; therefore, if two random variables are uncorrelated, then there is no linear relation between the two properties, but it does not necessarily mean that the two variables are independent. For example, if Y = X2, and X takes positive and negative values, then the correlation coefficient is close to 0, but yet Y depends deterministically on X through the quadratic relation (Figure 1.6), and the two variables are not independent.
Figure 1.6 Examples of different correlations of the joint distribution of two random variables X and Y. The correlation coefficient ρX,Y is 0.9 and −0.6 in the top plots and approximately 0 in the bottom plots.