Download The Multivariate Normal Distribution • The univariate normal

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
The Multivariate Normal Distribution
• The univariate normal distribution can be characterized by:
X has the univariate normal distribution if its density
function fX (x) satisfies
log fX (x) = quadratic function of x.
• The multivariate normal distribution can be characterized
similarly.
1
Preliminaries
• The joint cumulative distribution function (cdf) of X1, X2, . . . , Xp
is
F1:p (x1, x2, . . . , xp) = Pr {X1 ≤ x1, X2 ≤ x2, . . . , Xp ≤ xp}
• If F (x1, x2, . . . , xp) is differentiable, the joint probability density function (pdf) is
∂ p F (x1 , x 2 , . . . , x p )
f1:p (x1, x2, . . . , xp) =
.
∂x1∂x2 . . . ∂xp
2
• For r < p, the marginal cdf of X1, X2, . . . , Xr is
F1:r (x1, x2, . . . , xr ) = Pr {X1 ≤ x1, X2 ≤ x2, . . . , Xr ≤ xr }
= Pr {X1 ≤ x1, X2 ≤ x2, . . . , Xr ≤ xr ,
o
Xr+1 ≤ ∞, . . . , Xp ≤ ∞
= F1:p (x1, x2, . . . , xr , ∞, . . . , ∞)
• The marginal pdf is
f1:r (x1, x2, . . . , xr )
=
Z ∞
−∞
...
Z ∞
−∞
f1:p x1, x2, . . . , xr , ur+1, . . . , up dur+1 . . . dup
3
Independence
• The variables X1, X2, . . . , Xr are indendepent of Xr+1, Xr+2, . . . , Xp
if
F1:p (x1, x2, . . . , xp) =
F1:r (x1, x2, . . . , xr ) × F(r+1):p xr+1, xr+2, . . . , xp
• If the pdf f1:p (x1, x2, . . . , xp) exists, then also
f1:p (x1, x2, . . . , xp) =
f1:r (x1, x2, . . . , xr ) × f(r+1):p xr+1, xr+2, . . . , xp
4
Conditional Pdf
• The conditional pdf of X1, X2, . . . , Xr , given
Xr+1 = xr+1, Xr+2 = xr+2, . . . , Xp = xp is
f1:r|(r+1):p x1, x2, . . . , xr |xr+1, xr+2, . . . , xp
=
f1:p x1, x2, . . . , xr , xr+1, xr+2, . . . , xp
f(r+1):p xr+1, xr+2, . . . , xp
• Note: the conditioning event has probability zero. In general,
this conditional pdf must be defined as a Radon-Nikodym
derivative. The expression may be derived as a limit, on the
assumption that the pdfs are all continuous and the denominator is non-zero.
5
Characterizing the Multivariate Normal Distribution
• (-2 times) the logarithm of the pdf is a quadratic:
− 2 log f1:p (x1, x2, . . . , xp) =
p
X
ai,ix2
i +2
i=1
p−1
X
p
X
ai,j xixj + 2
i=1 j=i+1
p
X
c i xi + d
i=1
• In matrix notation:
−2 log f (x) = x0Ax + 2c0x + d
6
• For f (x) to be integrable, it must converge to 0 as x → ∞,
so A must be positive definite.
• In particular, A−1 exists.
• Write b = A−1c. Then
−2 log f (x) = (x − b)0A(x − b) + d∗,
where d∗ = d − b0Ab.
• Equivalently
1
f (x) = K exp − (x − b)0A(x − b)
2
7
Standardization
• Choleski: A = LL0 for lower triangular L.
• Define a new multivariate random variable Y = L0(X − b).
• The pdf of Y is then
0−1
fY (y) = fX b + L y × J(y),
where J(y) is the Jacobian
∂ x J(y) = det
∂y
= | det L0−1|
q
=
det A−1
8
• That is,
1
fY (y) = K det A−1 exp − y0y
q
2
1
= K det A−1 exp − (y12 + y22 + . . . yp2)
2
q
q
p
Y
= K det 2π A−1
φ(yi),
i=1
where
2 /2
−1/2
−y
φ(y) = (2π)
e
is the standard (univariate) normal density.
9
√
• Since both√fY (y) and φ(y) integrate to 1, K det 2π A−1 = 1,
or K = 1/ det 2π A−1.
• So the pdf of X is
1
exp − (x − b)0A(x − b)
f (x) = √
2
det 2π A−1
1
• More importantly, Y1, Y2, . . . , Yp are mutually independent,
and each has the standard normal distribution.
10
Moments
• Definition: the expected value of a random vector or matrix
is the vector or matrix of elementwise expected values.
• Because Y1, Y2, . . . , Yp are mutually independent N (0, 1), their
mean vector is
E(Y) = 0
and their covariance matrix is
C(Y) , E [Y − E(Y)] [Y − E(Y)]
n
0
o
=I
11
• Since X = b + L0−1Y,
E(X) = b + L0−1E(Y) = b,
• Similarly
C(X) = L0−1C(Y)L−1 = (LL0)−1 = A−1
• If we write µ for
the pdf of X is
E(X)
and Σ for
C(X),
then A = Σ−1, and
1
exp − (x − µ)0Σ−1(x − µ) , n(x|µ, Σ)
fX (x) = √
2
det 2π Σ
1
12
Related documents