Multi Variate Analysis
Multi Variate Analysis
é (as in the hypothesis), then det C = [Thy oi.
Since cj: > 0 for all i, then det@ > 0, and Cis non-singular. Note tht for each i the
cofactor of ci is the determinant of a lower triangular matrix, all of whose diagonal elements
are pacitive Hence all the diagonal elements of C-! are positive. Ifj > i, then the cofactor
of cj is (1) times the determinant of a lower triangular matrix with j - i diagonal
elements being zeros. Thus if O-! = (bj), then by = 0if j >i. QED.
LEMMA 8. If B is a positive-definite symmetric matriz, then there eists a lower triangular
matrit D which satisfies DD! = B.
Proof: Let B-¥" be a positive-definite matrix satisfying B-/?B-¥? = B+, and let BY? =
(B-?)-1, Note that BY? is positive definite, and B = BY/?B4/?, By Lemma thereexists a
lower triangular matrix C such that CBY/? is an orthogonal matrix. Then CBY?BYCt = J,
or B= C-Y(C')! = C-(0-1)', By Lemma 2, O- is lower triangular. Thus we may take
D-o". QED.
49LEMMA 4. The matriz A = XX¢ is positine definite with probability ane
Proof: Since n > p, X has rank p with probability one. Hence for all ¢ € RP\{0},¢'X # 0°
with probability one, which implies ¢*X.X'C > 0 with probability one. QED.
Notation: Let X} denote the ith row of X, ie., X' = (X ,). Note that by Lemma
8 of Section 1 the vectors X;,-+-,X, are linearly independent with probability one.
LEMMA 5. There ezists a p x p lower triangular matriz of random variables denoted by
Bo = (bi) such that BY is a measurable function of X1,-++,Xx only, and such that if
y= x,
Ye — Ox, + 0X,
Yo = PIX $+ + UPX,,
then Yi,-++,¥p form an orthonormal system in R” with probability 1. In addition b# > 0
with probability one for 1 0
yt
50with prohahility ane ‘Thus the inverse of Bol exists with probability one, and we may
define B by B = (B™")-1. We denote B = (83), and observe that, by Lemma 2, B is lower
triangular. We collect all this in the following lemma.
LEMMA 6. The following hold:
(1) Y¥=B°X,
(2) YY" = J, with probability one,
(3) X= BY,
(4) B= XY*,
(5) A= BBY, and
(©) bs = X1%.
Proof: (1) follows from Lemma 5. (2) is true because ¥,---,Yp are orthonormal with
probability one. (8) follows from (1). (4) follows from (2) and (3). (5) follows from (2) and
(4) and the definition of A. (6) follows from (4).
LEMMA T. For 2 Si S p, the sets of random vectors {¥4,-++ Yin} and {Xi,-+*, Xp} are
independent,
Proof: Recall that ¥j,--+,Y» form an orthonormal system, where ¥; (by the Gram-Schmidt
process that occurred to make ¥ out of X) is a function of X;,---,X; only. Since all np ran-
51dom variables in X are independent, then {Xi,+++,%p} aud {¥4,"-+)¥ie1) are independent,
QED.
LEMMA 8. For 2S iS p, the matriz
yf
yt
constitutes the first i—1 rows of a random orthogonal matriz, all of whose entries are random
variables that are independent of X;
Xp
Proof: This follows from Lemma 2 in Section 4 of Chapter 1 and from Lemma 7 above.
QED.
LEMMA 9. For 2
Yh
Proof: By Lemma 6(4), B= X¥* or Bt = YX'. The first i— 1 terms of the ith column of
both sides of this equation yield our result. QED.
.-1 are independent, and each
LEMMA 10. For 2 0] = 1.
Proof: By Lemmas 10 and 13, bir,+*+5
by arc independent. QED.
Also onc casily proves that by = 1/bi, Hence buy
LEMMA 18. The rows of B are stochastically independent.
Proof: Let us denote
xt yf
and Qi-1 =
Xa Yay
and let Bi) denote the first i—1 rows and first i- 1 columns of B-?. Because of the
fact that B- is lower triangular, it follows that Q;.1 = Bu',P.-1. But note that all of the
random variables in Bj-4 are Borel-measurable functions of those in 4.1, the first 1—1 rows
of X, and thus we may write Q;-1 = ¥(P-1), where # is a Borel-measurable function. Let
% ”
us agree on the following notation: if V = ( A ) is a random vector, and if v = ( i ) eR,
54then we shall denote [V < v] = (\[; Swi]. Let pi-a be an (i—1) x n matnix of Lnearly
fel
independent rows of real numbers. Then for x € R'“?, we have, for y > 0,
P([bix $ 21,+++, Bina S innybh S yl Pien = pins)
= PULQiaXi S xXIXi — XIQinQia% S vIlPina = pina)
= P({O(Pi-a) Xi $ x]IXEX: — XI0(Pi-1)'Y(Pi-a)%i S 9)
= [var [> oloat [xe sasltat,
where ¢ is the density of the A'(0,1) distribution and x2_i43(t) is the density of the x24"
distribution. Thus we see that the conditional joint distribution of the ith row of B namely,
baay+++ bi, given the values of the first i—1 rows of X,, is the same as that of the unconditional
distribution. We use this just-proved fact to prove that the rows of B are independent. Let
‘
ao ( ) so let ©, be any Borel set in R125 < ps Thos
P Mail € Ui) = £ i ‘wee
es a(e (Fr Amecai®s %))
ao (finest (ryecalXae*+ Ku) .
But by the fact proved above, if 2H') distribution.
Proof: Let X be a sample of size n on a Np(4,2) distribution. Then ¥ = JLX is a sample
of size n on a Nm (Hp, HS Hi) distribution. We may write
D=(X~ pI) - 214)
Then, clearly,
HDH' = (¥ - Hplt)(¥ - Hpty,
58Thue HDHt has the Wo(n, HSH!) distribution OED.
COROLLARY 1 TO LEMMA 1. If D W,(n,E),
is a partition of D where Dy, is an m xm submatriz, and if
En| En
Enl En
is a partition of © where Dy, is an m x m submatriz, then Dy has the Wm(n, Dy) distri-
bution.
Proof: Let H = (Jm|0) be an m x p matrix. Then Lemma 1 yields the result. QED.
COROLLARY 2 TO LEMMA 1. Ifh € RY is a constant vector, if ht # 0, and if D is
W,(n,), then ae has the x2-distribution.
Proof: By Corollary 1, h'Dh has the W3(n,h' Ch) distribution, i.e., htDh has the same
distribution as does DRL, ¥2, where ¥j,--+,Yq are independent, each being N(0,h* Dh).
Hence htDh/h' Dh has the x2-distribution. QED.
LEMMA 2. [fh is a p-dimensional random vector such that Ph # 0] = 1, if D is
W,(n,E), and ifh and D are independent, then
(i) h'Dh/h' Dh has the x2-distribution, and
39(i) b‘Dh/n‘E n and h are independent.
Proof: We shall prove this lemma only in the case where h has a joint absolutely continuous
distribution. By Corollary 2 to Lemma 1] and by Propositions 4 and 6 in Chapter 1 we have,
for z >0,
PUh'Db/b! Dh <2] = EUlpowmesrncsy)
Sf, Blew oun Sagalh = *)fale}ar
= (r'Dr/rt Sor < 2] ir
= fe, PeetDels! De 5 alfalrdd
= LL (Leto) soar = [rine
where x2(¢) denotes the density of the x2-distribution. This proves (i). In order to prove(ii),
we first note that in our proof of part (i) we showed that,
P({h'Dh/h! Sh < alfh =r) = [ xA(Hat
for all z > 0 and all r € R®. Hence, for all Borel sets A in R?,
P({h'Dh/h' Th S z)fh € A)
= Le P({h'Dh/ht yh < a]fh € Allh =r) fulr)dr
= hh »Plle'Dr/t' Dr < alle € Al) fale)ar
= J, PlletDe/e" Dr < 2) falter.
Now by the remark above, by the fact that P[h # 0] = 1, we see that by Corollary 2 to
Lemma 1 and for all x #0, and by (3),
P({t'Dr/r' Sr < 2) = Plth'Dh/h' Sh < z}).
60Hence
P({h'Dh/ht Soh < 2{h € A))
= | PUlb'Db/a! Dh < al) falrér
= P({h'Dh/h‘ Th < z}) I, Sa(r)dr
= P([h'Dh/h' Oh < 2})P(\h € Al).
QED.
‘We shall henceforth use the following notation for elements of the matrices D, D-?
D= (dij), = (4%), = (aij) and E~ = (0%).
LEMMA 8. If D is Wy(1,D), tien det D/UetE hus the oame distribution ao the product
ofp independent random variables whose distributions are x2, x2-15°*-sX2-pt1» and 0”? /d??
has the x2_,41-distribution.
Proof: Let C be a lower triangular matrix of constants such that = CCt, and define
‘hen, by Lemma 1, A is W,(n,C-? £(C’
A=C7D(C !), ie., Ais W,(n,1,). Thus
A has the same joint distribution as does XX', where X is a p x n matrix of independent
N(0,1) random vatiables. For distributional purposes we may define A = XX‘. Now with
respect to X let B be as defined in Section 2, i.e., B = (bij) is a pxp lower triangular matrix,
83, ie
)} are independent random variablee, by ie W{(0,1) if 5 <
Bylsisirs
xii: and A= BBt. Let us define T= CB. Clearly, T is lower triangular. We next note
61that
TT! = CBBYC' = CAC = CO“ D(C™)'C' = D,
ie, TT! = D. Thus
detD = det(TT*) = det(CBB‘C') = det(CC') det(BB')
aet() TT.
By Lemma 13 in Section 2 of Chapter 2, we obtain the first conclusion of our lemma. In order
to obtain the second conclusion, let Dp-1,Tp-1 Ap-1, Bp-1 Dp-1 and Cp-1 be (p—1) x (p—1)
matrices obtained from the first p—1 rows and the first p—1 columns of D,T, A, B,5 and C
= TysTp a Ap
= Cp-tAp-iCf.,. From these and
respectively. From relations established above it easily follows that Dy
Biy-1BhayZpea = CpsChaay Tyna = Cp-1Bpes and Dp
the fact that Ris lower triangular we get
det A det(BB') __(det BY?
Betas TB aba) ~ CEP
Hence
i = -dttA_ _ det Adet O det O' det Oy-1 det 1
oe = Tet Apa det Ayan det C det Of det 0,1 det Chay
det(CACt) _det(C,-1Ch.1)
FGrApaCha) det(CC4)
= detD det Dyan
CD ety
Note that D- = (a) where d'¥ = (cofactor of dji)/det D. Since det D,-1 is the cofactor
of dyp, it follows that d”? = (det D,-1)/det D. Similarly, o?? = (detD,-1)/det O. Hence
12, = a” /d?” which has the QED.LEMMA 4. Let D be W,(n,2), let h be a p-dimensional random vector with a joint
absolutely continuous distribution function, and assume that D and h are independent. Then
(i) RRS hes the x2_,4,-distribution, and
Gi) MEE and h are independent.
Proof: Let H be a px p orthogonal matrix of random variables whose pth row is (1/||hi|)h*
and such that all entries in H are measurable functions of h. We may write H = H(h). Let
hho € range (h), and let Ho = H(ho). Now Ho is an orthogonal px_p matrix of numbers. Note
that as a consequence of our hypotheses, H and D are independent. Define D* = HoDH} and
I" = HoDH§. By Lemma | and the fact that H and D are independent, the conditional
diotribution of HDH" given H — Hy ie the unconditional distribution of HoDH{, which is
W,(n, Ho X Hg). Let us denote
(dj), Do? = (d"4), O° = (0)
and D*-! = (o*4), where D* and [-* are as denoted above. By Lemma 3, o°??/d"?? has the
Xi -pax-distribution. Note that HyD-1H§ = D*-*, Now, for appropriate matrices H1, Ho, Hs
and Hy we have
HoD“H, =
He |pebrh§D-ho
63rm — pppphbTothy. Ao noted above, o°/d"™ hoo
Be has the same distribution function. Now let F(z)
be the distribution function for the x2_p41-distribution, and let g(-) be the joint density of
hh. Then, since h and D are independent we have
ht
he (ES
hg D7" ho
[2 RSE <2] oan
J, Fle)alho)dho = F(2),
Se] [b= ) g(by)dhy
1
)
in Lemma 2. QED.
which proves (i). The proof of (ji) follows the same steps for the proof of conclusion in
LEMMA 5. I/U is N(u,5), then (U — 4)! E7'(U — p) has the x2-distribution.
Proof: Since and therefore S--? are positive definite, there exists a positive definite matrix
XM? such that EVO“? = D-*. Ou easily verifies that L-2(U — gs) is Np(0, Jp).
Hence (U— 1)! 5-1(U~p) is the sum of squares of p independent W/(0, 1) random variables.
QED.
THEOREM 1. I/U is Nj(4,D), if D is Wp(r4D), ifp You might also like