Exactly Central Limit: Multivariate Statistical Methods
Exactly Central Limit: Multivariate Statistical Methods
Figure: A normal density with mean µ and variance σ 2 and selected areas under
the curve
1 − 1
f (x1 , x2 ) = p e 2(1−ρ2 )
2πσ1 σ2 1 − ρ2
(x1 − µ1 )2 (x1 − µ1 )(x2 − µ2 ) (x2 − µ2 )2
× − 2ρ +
σ12 σ1 σ2 σ22
Figure: Two bivariate normal distributions (a) σ11 = σ22 and ρ12 = 0 (b)
σ11 = σ22 and ρ12 = 0.75
Abiyot. (JU) Multivariate statistical Methods 50
The Multivariate Normal Distribution Bivariate Normal Density
√
These ellipsoids are centered at µ and have axes ±c λi ei , where
where (λi , ei ) are the eigenvalue-eigenvector pairs of Σ.
This shows why Σ can be a measure of spread
The solid ellipsoid of x values satisfying
Let A0 = Iq 0q×(p−q) .Then xq×1 ∼ Nq (µq×1 , Σ11 )
4 xq×1 ⊥ x(p−q)×1 if and only if Σ12 = 0
Abiyot. (JU) Multivariate statistical Methods 54
The Multivariate Normal Distribution Further Properties of the Multivariate Normal Density
Further Properties...
The following are true for a normal vector X having a multivariate normal
distribution:
1 Linear combination of the components of X are normally distributed.
2 All subsets of the components of X have a (multivariate) normal
distribution.
3 Zero covariance implies that the corresponding components are
independently distributed.
4 The conditional distributions of the components are normal.
Conditional distribution
Nq (µ1.2 , Σ11.2 )
Note that the covariance does not depend on the value x(2) of the
conditioning variable.
Example: Bivariate case: x1 is the height of the son. x2 is the height
of the father.
1 x 1 µ σ12 ρσ1 σ2
x = , µ = , Σ =
x2 µ2 ρσ1 σ2 σ22
(2π) |Σ| 2 2
np n
ln{f (X1 , X2 , . . . , Xn )} = − ln(2π) + ln(|Σ|−1 |)
2 2
" ( n )#
1 −1 X
− tr Σ (Xi − x)(Xi − x)0 + n(x − µ)(x − µ)0
2 i=1
Likelihood
When the numerical values of the observations become available, they may
be substituted for the xj in the equation above. The resulting expression,
now considered as a function of μ and Σ for the fixed set of observations
x1 , x2 , . . . , xn , is called the likelihood.
One meaning of best is to select the parameter values that maximize the
joint density evaluated at the observations. This technique is called
maximum likelihood estimation, and the maximizing parameter values are
called maximum likelihood estimates.
n
1X
Σ̂ = (xi − x)(xi − x)0
n
i=1
n−1
= S
n
are the maximum likelihood estimators of µ and Σ, respectively. Their
1 Pn
observed value x and n i=1 (xi − x)(xi − x) are called the maximum
likelihood estimates of µ and Σ
Sampling distribution
Wishart distribution
Let X1 , · · · , XN be a random sample from Np (µ, Σ). Denote
N
X
A = (N − 1)S = (Xi − x)0 (Xi − x)
i=1
d Pn 0
Let n = N − 1. It can be shown that A = i=1 zi zi where
iid
zi ∼ Np (0, Σ)
Hence A ∼ Wp (n, Σ), a Wishart distribution with n d.f whose density
is given by
n−p−1
|A| 2 exp{− 12 tr (Σ−1 A)}
np n , (n ≥ p)
2 2 |Σ| 2 Γp ( n2 )
where
1
Γp (t) = π p(p−1)/4 Πpi=1 Γ(t − (i − 1))
2
is the multivariate gamma function
Abiyot. (JU) Multivariate statistical Methods 61
The Multivariate Normal Distribution Sampling distribution
Properties...
Partition
A11 A12 Σ11 Σ12
A= , Σ =
A21 A22 Σ21 Σ22
and
n(X − µ)0 S(X − µ) is approximately χ2p
for n − p large.