Statistical Foundaments
Statistical Foundaments
1 Expected Value
Let X = [x1 · · · xp ]T be a random p-dimensional vector.
Properties
i If a is a constant vector, then E(a) = a
E(a1 ) a1
.. ..
Proof. E(a) = . = . = a.
E(ap ap
ii E(a + X) = a + E(X).
Proof.
E(a1 + x1 ) a1 + E(x1 ) a1 E(x1 )
E(a + X) = .. .. .. ..
= = . + . .
. .
E(ap + xp ) ap + E(xp ) ap E(xp )
a) E(AX) = AE(X),
b) E(XA) = E(X)A.
Proof.
Pp Pp Pp
a1j xj
j E( j a1j xj ) j a1j E(xj )
E(AX) = E .. .. ..
= = = AE(X).
Pp . Pp . Pp .
j apj xj E( j apj xj ) j apj E(xj )
1
Proof.
x1 + y 1 E(x1 ) + E(y1 ) E(x1 ) E(y1 )
E(X + Y) = E ... = .. .. ..
= . + . = E(X)+E(Y).
.
xp + y p E(xp ) + E(yp ) E(xp ) E(yp )
2 Dispersion matrix
Definition. The variance-covariance matrix of X, or dispersion matrix,
denoted by Σ , is defined by
Note that
(x1 − µ1 )2
· · · (x1 − µ1 )(xp − µp )
(X − µ)(X − µ)T = .. ... ..
,
. .
2
(xp − µp )(x1 − µ1 ) · · · (xp − µp )
therefore
V (x1 )· · · cov(x1 , xp )
Σ= .. ... ..
.
. .
cov(xp , x1 ) · · · V (xp )
yT Ay ≥ 0 ∀y ∈ Rp .
2
yT Σy = yT E[(X − µ)(X − µ)T ]y,
= E[yT (X − µ) (X − µ)T y],
| {z } | {z }
1×1 1×1
Properties
i If b ∈ Rp is a constant vector, then D(b) = Σ = 0p×p .
Proof.
(b1 − b1 )(b1 − b1 ) · · · (b1 − b1 )(bp − bp ) 0 ··· 0
.. .. .. . .
D(b) = E = E .. . . . .. = 0p×p .
. . .
(bp − bp )(b1 − b1 ) · · · (bp − bp )(bp − bp ) 0 ··· 0
Proof.
D(b + X),
(b1 + x1 − b1 − µ1 )2 · · · (b1 + x1 − b1 − µ1 )( bp + xp − bp − µp )
= E .. ... ..
,
. .
2
bp + xp −
( bp − µp )(b1 + x1 − b1 − µ1 ) · · · bp + xp −
( bp − µp )
= D(X).
D(AX) = AΣAT .
Proof.
3
iv If X and Y are independent random variables, then
Proof. Before we prove this property, we shall prove that if X and Y are
independent then xi and yi are also independent for all (i, j) ∈ (1, · · · , p)2 .
That is to say, that the components of X and Y are also independent. Let
X and Y two independent random vectors, then
f (x1 , y1 )
f (x1 |y1 ) = ,
f (y1 ) marginal
probability
ZZ
1
= f (x1 , x2 , y1 , y2 )dx2 dy2 ,
f (y1 )
ZZ independence
1
= f (x1 , x2 )f (y1 , y2 )dx2 dy2 ,
f (y1 )
Z Z
1
= f (x1 , x2 )dx2 f (y1 , y2 )dy2 ,
f (y1 )
f (x1 )f (y1 )
= = f (x1 ).
f (y1 )
Then cov(xi , yj ) = 0, ∀(i, j) ∈ (1, · · · , p)2 . Now we have
cov(x1 + y1 , x1 + y1 ) · · · cov(x1 + y1 , xp + yp )
D(X + Y) = .. .. ..
.
. . .
cov(xp + yp , x1 + y1 ) · · · cov(xp + yp , xp + yp )
3 Covariance matrix
Let there be given two stochastic variables X = [x1 · · · xp ]T and Y = [y1 · · · yp ]T of
dimension p and q, respectively, and with mean values µ and ν.
Definition. The covariance matrix between X and Y is defined by
cov(x1 , y1 ) · · · cov(x1 , yq )
C(X, Y ) = E[(X − µ)(Y − ν)] = .. ... ..
.
. .
cov(xp , y1 ) · · · cov(xp , yq )
4
Properties
i C(X, X) = D(X).
iii If An×p and Bm×q are real matrix, then C(AX, BY) = AC(X, Y)B T .
Proof.
iv Let U and V two p and q-dimensional random vectors with mean γ and δ,
respectively. Then
Proof.
Proof.