An Introduction To Bayesian Inference in Ecometrics
An Introduction To Bayesian Inference in Ecometrics
Copyright � 1971 by John Wiley & Sons, Inc. Wiley Classics Library
Edition Published 1996. All rights reserved. Published simultaneously in
Canada. Reproduction or translation of any part of this work beyond that
permitted by Section 107 or 108 of the 1976 United States Copyright Act
without the permission of the copyright owner is unlawful. Requests for
permission or further information should be addressed to the Permissions
Department, John Wiley & Sons, Inc. Library of Congress Catalogue Card
Number: 70-156329 ISBN 0-471-98165-6 ISBN 0-471-16937-4 Wiley
Classics Library Edition Printed in the United States of America 10987654
to Agnes and our sons
124 ON ERRORS IN THE VARIABLES where [' -- (/80,/g), [' = (x, .,..., ,),
Y' = (Y', yo.'), yj' = (yjx, yo.,..., y), j -- 1, 2, and t is a n x 1 column vector,
with each element equal to one; that is, t' = (1, 1,..., 1). On taking the
logarithm of both sides of (5.32) and differentiating with respect to the
elements of , 0.x, and 0.2 we obtain (5.33) log/ 1 (Yx - ) + 1 [_/go. +/g(Y2
t/go)], d-'-- = 0.o. 0..-'. - log I n (5.34) 1 (y _ )'(Yx - ID, (5.35) Olog/ _n + 1
(yo. - 0.2 --' 0'2 0.2 3 As a necessary condition for a maximum, values of
the parameters must exist in the admissible parameter space, which sets
these derivatives equal to zero as well as the derivatives with respect to /go
and/g. On setting (5.33) equal to zero, we have 6 yx/3x' + (o./3?')(y. - ot)/ =
+ � = 1+' where 0 = 0�'/�'/02 �' and = (yo. -/ot)//. Substituting from
(5.36) in (5.34) and (5.35) and setting these derivatiyes equal to zero, we
have the following results' (5.37) --' (1 + 0) ' (yx - )'(Yx - ), 2 (5.38) o. =
(Yx - v)'(yx - ). 0.2 n(1 + 0) 2 These two equations can hold simultaneously
if, and only if, o. = x, For this problem we define the admissible parameSer
space as follows: 0 < **o. < o, i= 1,2; -o < < 0% i= 1,2,...,n; -oo < o, < oo;
ax 2 oo.o.; and o. oo.�'/oxo.. The reason for introducing this last condition
is explained in the text. x Equation 5.36 can be more fully appreciated by
writing (5.30) to (5.31) as Yx = ( + ux and (Y2 - o)/ = + uad. The elements
of (5.36) are then weighted averages of Yx{ and (Y2 - o)/18. For given t8o
and 18(0) this is an n-means problem with two observations per mean and
variances x 2 and .2/2. x7 This result was obtained by D. V. LindIcy,
"Regression Lines and the Linear Func- tional Relationship," J. Roy. Statist.
Sec. (Supplement) 9 (1947), 218-244. His inter- pretation of it, as well as
that of J. Johnston, Econometric Methods. New York: McGraw- Hill, 1963,
p. 152, is somewhat different from ours. CLASSICAL EVM: ML
ANALYSIS OF THE FUNCTIONAL FORM 125 However, in our
definition of the admissible parameter space, we explicitly stated that /g'-
0..o./0.. and thus the quantity o.= &../&o. falls in an inadmissible region of
the parameter space. Since this is so, the necessary first-order conditions for
a maximum of the likelihood function cannot be simultaneously'satisfied;
hence a maximum of the likelihood function does not exist in the
admissible region of the parameter space. xa Since there are basic
difficulties with the analysis of the model in (5.30) to (5.31) when all
parameters are assumed unknown, analyses have often gone forward under
the assumption that 4 = 0.o.'/0.x �' is known exactly? Under this
assumption a unique maximum of the likelihood function exists and thus
ML estimates can be obtained. To do so we write the likelihood function as
(5.39) 1 ( I(1, g, 0.o. ly, oc exp - 1 + (Y2 -/got -/g)'(yo, -/got - On
differentiating the logarithm of the likelihood function L with respect to the
unknown parameters and setting these derivatives equal to zero, we have
(5.40) L = .. 1 (y. _/got -/gg)'t = 0, (5.41) L 1 (y. _/got -/gg)'g = O, 0.2 2
(5.42) OL = 1 [4(Y - ) +/g(Y. - riot- I)1 0, and (5.43) OL 2n+ 1
[4(Yx-)'(Yx-)+(y. /got /gg)'(y. /got /g)] 0. From (5.40) (5.44) xs If we
combine the following diffuse prior pdf, P(18o, 18, , ax, *.) c l/*xo., 0 < **
< 0% i = 1, 2, and -oo < rio, fi, fi < 0% i = 1, 2,..., n, with the likelihood
function in (5.32), it is not difficult to show that the joint posterior pdf is
improper. x* In econometrics this precise knowledge is usually unavailable.
Below we present methods for utilizing less precise information 'about .
136 ON ERRORS IN THE VARIABLES with 0 < 4, < co and flL < fl < flu,
where flL and flu are the a priori bounds placed on fl. The denominator of
(5.78) has a minimum at o. -- mx./mn, the least squares quantity from a
regression of y2 on y, and thus were it not for the factors involving fl in the
numerator (5.78) would have a mode at fl., pro- vided fl < /. <- flu. The
factor (1 + tg�'/$) ' produces a modal value for larger in absolute value
than/s.4 The amount by which the modal value for fl is increased absolutely
will depend on what is assumed about 4,. If the prior pdf for 4, favors large
values, the modal value for fl will be close to/., abstracting from the
information about fl included in p(fl).4' To use (5.78) in practice prior pdf's
for fl and 4' must be assigned. With respect to the prior pdf for fl, p(tg),.tgr
<- fl < flu, we can assign a beta pdf of the following form *: (5.79) p(zla, b)
B(a, b5 z'-(1 - z) b- a, b > 0, = 0<z_< 1, where z = (/ -/D/(/gu -/gD, a and b
are prior parameters to be assigned by an investigator, and B(a, b) denotes
the beta function with arguments a and b. The pdf in (5.79) is a rather rich
one that accommodates the prior information that/gL </g <-/gu. If, for
example, a = b = 1, (5.79) gives us a uniform pdf for/g. With respect to 4,,
the prior pdf p.(4,) might be taken in the following inverted gamma form: 1
[ voSo\ 0 < 4, < (5.80) po.(4,lVo, So) cr 4,,--775 exp ,----], vo, So > 0.
where vo and So are prior parameters. Although (5.79) and (5.80) are not
the only forms of prior pdf's that can be used for the present problem, they
appear to be rich enough to be capable of representing available prior
information in a wide range of circumstances. Substituting from (5.79) and
(5.80) into (5.78), we have a bivariate posterior,. pdf which can be analyzed
using bivariate numerical integration tech- niques. Marginal posterior pdf's
for/g and 4' can be computed along with measures characterizing them. In
addition, it is possible to compute joint posterior regions for/g and 4,. In
order to illustrate application of the above techniques, data have bc
generated from the model Yx = : + ux, y{ = 2.0 + 1.0 + u4, i = 1, 2,..., 20, x
This can be interpreted heuristically as a correction for the inconsistency of
' Note that, given 4,, the value of/ which maximizes (1 + ta/4,)/(rnaa -
2/rnxa + is the ML estimate. Thus the modal value of (5.78) for given 4,
will be close to the ML estimate if pz(/) oc constant. � a See Appendix A
for a review of the properties of the beta pdf. 4 See Appendix A for
properties of the inverted gamma pdf. BAYESIAN ANALYSIS OF THE
FUNCTIONAL FORM OF THE EVM 137 under the following conditions.
The values of ux{ and uo., were drawn inde- pendently from normal
distributions with zero means and variances 4 and 1, respectively; that is 4,
= e.o./,x = �. The ,'s were drawn independently from a normal distribution
with mean t = 5 and variance r 2 = 16. The 20 pairs of observations are
shown in Table 5.1 and plotted in Figure 5.1. -5 -10 o Y2 i 20- o � -- 0 o o
o o �o o o o o o I I I l Yi 5 10 15 Figure 5.1 Plot of generated data. Table
5.1 GENERATED OBSERVATIONS i Yx Yat i Yz{ Ym 1 1.420 3.695 11
1.964 4.987 2 6.268 6.925 12 1.406 6.647 3 8.854 8.923 13 0.002 2.873 4
8.532 14.043 14 3.212 4.015 5 -5.398 -0.836 15 9.042 10.204 6 13.776
16.609 16 1.474 1.953 7 5.278 4.405 17 8.528 10.672 8 6.298 9.823 18
7.348 9.157 9 9.886 12.611 19 6.690 8.552 10 11.362 10.174 20 5.796
10.250 Yx = 5.587 Ya = 7.784 mxx = 19.332 maa = 17.945 mxa = 16.925
138 ON ERRORS IN THE VARIABLES Using the data in Table 5.1, first
conditional posterior pdf's for/g, given 4 '= 0.25 and 4-- 1.0, were computed
4 from (5.78) and are shown in Figure 5.2. It is seen that the location and
spread of the posterior pdf's are 5.0 4.0 3.0 2.0 0.25, y,y2) = 1.0, y, y2 ) 1.0
Figure 5.2 I 0.80 0.90 1.00 1.10 1.20 Conditional posterior pdf's for/, given
4,. 1.30 not too sensitive to what is assumed about 4. In this instance, when
4 is assumed to be equal to 0.25, the value used to generate the data, the
con- ditional posterior pdf for/, has its mode at/ = 1.02 close to the value
used to generate the data, namely, 1.0. With 4 assumed to be equal to 1.0,
the modal value of the conditional posterior pdf for/ is located at/ = 0.96.
These results contrast with what is obtained from an analysis of the model
under the assumption that a �' = 0, that is, no measurement error in y. In
this case, with a diffuse prior pdf on/o and/, the posterior pdf for/ is centered
at fl2x = mx2/mxx = 0.876, the least squares quantity. 46 Although the
conditional pdf for/, given 4, is of interest, it is often the case that we lack
prior information that is precise enough to assign a specific value to 4.
However, it may be possible to choose a prior pdf to 45 A uniform prior pdf
for/ with a rather large range was employed. � 6 A simple regression of
Ym on Yx results in .9m = 2.893 + 0.876yx, (0.674) (0.0948) where the
figures in parentheses are conventional standard errors. BAYESIAN
ANALYSIS OF THE FUNCTIONAL FORM OF THE EVM 139 represent
the information available about 4; for example, the inverted gamma pdf,
shown in (5.80), could be used for this purpose. To illustrate, let us assign
the following values to the parameters of (5.80)' Vo = 8 and = _L With these
values assigned, the prior pdf for 4 has its modal value SO 2 1 6' at 4 -
0.236, mean equal to 0.246, and variance equal to 0.0230 (standard
deviation = 0.152). 47 On inserting this prior pdf for for Pa(4) in (5.78) and
using the data shown in Table 5.1, we computed the joint posterior pdf for /
and 4 with a uniform prior for/. Also the marginal posterior pdf's for/ and
for 4 were computed. The results of these computations are shown in Figure
5.3. It is seen from the plots in Figure 5.3 that the marginal posterior 5.0 4.0
2.0 1.0 10.0 (. 8.0 II 6.0 4.0- 2.0- 0 ,-,, 0.80 1.0 1.2 1.4 0 O. 10 0.20 0.30
0.40 Figure 5.3 Marginal posterior pdf's for/ and 4' based on generated data
and utilizing inverted gamma prior pdf for 4,. pdf's for/g and 4 are
unimodal. As regards the pdf for 4, it appears that the information in the
particular sample employed has reduced the modal value from 0.236 in the
prior pdf to about 0.18 in the posterior pdf. The posterior pdf for/g has a
modal value at about 1.03, a value close to that used in generating the data.
We have gone ahead conditional on g = , an assumption that obviates the
need for a distributional assumption regarding the :,'s. On the other hand, *
See Appendix A for algebraic expressions for the modal value, mean,
variance, etc., of the inverted gamma pdf.
CHAPTER VII Time Series Models: Some Selected Examples Most, if not
all, economic analyses involve time series data. Thus it is important to have
good techniques for the analysis of time series models. In this chapter we
take up the analysis of selected time series models to demonstrate how the
Bayesian approach can be applied in the analysis of time series data. It will
be seen that if the likelihood function is given and if we have a prior pdf the
general principles of Bayesian analysis, presented in Chapter 2, apply with
no special modifications required. This is indeed fortunate, since it means
that our general principles are applicable to time series problems as well as
to others. 7.1 FIRST ORDER NORMAL AUTOREGRESSIVE PROCESS
The model assumed to generate our observations, y' = (yx, y,..., YT) is '
(7.1) y =/g + tgyt- + ut, t = 1, 2,..., T where/x and/g are unknown parameters
and u is a disturbance term. We assume that the ut's are normally and
independently distributed, each with zero mean and common variance ,o..
As regards initial conditions, we shall first go ahead conditional on given
Y0, the observation for t = 02 With these assumptions the likelihood
function is (7.2) l(, rio., *IY, y0) oc- exp Z (Y*- fi - fi2yt )2 T ZO - with the
summation extending from t = 1 to t = T. The distinction often made
between time series and cross-section data does not invalidate the statement
made above, since cross-section data are in fact observations on time series
variables pertaining to individual units in a cross section. Overlooking the
time series nature of cross-section data can at times lead to serious errors in
analyzing data. 2 Here we use the subscript t to denote the tth value of a
variable. a See the discussion of initial conditions presented in connection
with the problem of autocorrelation in regression analysis, presented in
Chapter 4, for other possible assumptions. 186 FIRST ORDER NORMAL
AUTOREGRESSIVE PROCESS 187 As regards a prior pdf for the
parameters, we shall assume that our information is diffuse and represent
this in the usual way, namely, 1 (7.3) p(/,/o., *) oc-, with -oo </ < oo, -oo </
< oo, and 0 < , < oo. Note that we do not restrict/ to be within the interval -
1 to + 1 and thus the analysis applies to both the explosive and
nonexplosive cases. 4 In fact, our posterior pdf for will reflect what the
sample information has to say about whether the process is or is not
explosive. On combining (7.2) and (7.3), the posterior pdf for the
parameters is (7.4) P(15, IY, Yo) oc -F'i exp 22 (Yt - tg - fiYt-)o. , where 15'
= (/,/o.), which is in a form exactly similar to that obtained in our analysis
of the simple normal regression model in Section 3.1. To obtain the
marginal posterior pdf for 15 we integrate (7.4) with respect to, which
yields P(151Y, Yo) oc [Y. (y,-/ -/o. yt_0�'] (7.5) oc [vs + (15 - )'H(15 -
where v = T- 2, (7.6) H= Ey_ ( r (7.7) = Y- Y-,.] \Y. Y- Yd and vs a = (Yt -
fix - aYt-x) a. It is seen from (7.5) that the joint posterior pdf for fix and a is
in the bivariate Student t form with mean given by (7.7), the least squares
quantity. This fact permits us to make inferences about d a quite readily. In
particular, the marginal posterior pdfs for will each be in the form of
univariate Student t pdf. Explicitly, the quantities will each be distributed as
a Student t variable with v = T- 2 degrees of frdom. Further, by integrating
(7.4) with respect to , the marginal pdf for is given by 1 [ s (7.8) P(lY, yo)
wire = r - 2 and = - - Of course, if we have information that the process is
nonexplosive and wish to use it, the prior pdf in (7.3) can be altered to
incorporate this information. See below for an example.
zzz
366 APPENDIX A where fi2 = //2 '. The measure yo. assumes a value of
zero for the UN pdf; pdf's for which yo. = 0 are called mesokurtic, those for
which y2 > 0 are leptokurtic, and those for which 72 < 0, platykurtic. As
already pointed out, "... it was thought that leptokurtic curves were more
sharply peaked, and platykurtic curves more flat-topped, than the normal
curve. This, however, is not necessarily so and although the terms are useful
they are best regarded as describing the sign of y2 rather than the shape of
the curve. "5 A.2 THE UNIVARIATE STUDENT t (US t)Pdf A random
variable, , is distributed in the US t form if, and only if, it has the following
pdf: (A. 14) p(xlO, h,O= I'[(v + 1)/2] (�) [ h r(1/2) r(/2) 1 +-(x - 0) �-
where -oo < 0< oo, 0 < h < o%0< randwhere I' denotes the gamma function.
This pdf has three parameters, 0, h, and v. From inspection of (A. 14) it is
seen that the US t pdf has a single mode at x = 0 and is sym- metric about
the modal value x = 0. Thus x -- 0 is the median and mean (which exists for
v > lmsee below) of the US t pdf. The following expressions give the odd-
and even-order moments about the mean: (A.15) Io.r- ---- E( -- O) �'r - = f
, and 2r -- E( - O) 2r = (A.16) (x - o) �'- p(xl o, h, 0 dx = O) r= 1,2,..., v >
2r-l, j.o (x - o) 2' p(xl o, h, 0 dx r(1/2) r0/2) > 2r. As shown below, for the
existence of the (2r - 1)st moment about 0 we must have v > 2r - 1.
Similarly, for the existence of the 2rth-order moment about 0, v must satisfy
v > 2r. Given the symmetry of the US t pdf about x -- 0, all existing odd-
order moments in (A. 15) are zero. In particular, E(. - 0) = 0 so that E. = 0
which exists for v > 1. � With respect to the even-order M. G. Kendall and
A. Stuart, The Advanced Theory of Statistics, Vol. I. London: Griffin, 1958.
New York: Hafner, p. 86. However, for many distributions encountered in
practice a positive ya does mean a sharper peak with higher tails than if the
distribution were normal. o For v = 1 the US t pdf is identical to the Cauchy
pdf for which the first- and higher- order moments do not exist. (A. 17) and
PROPERTIES OF SEVERAL IMPORTANT UNIVARIATE PDF'S 367
moments in (A.16) the second- and fourth-order moments are given by 7 1
v /o. -= E(� - 0) �' = for v > 2, (A.18) /z4 -= E(: - 0) 4 = (v - 2)(v - 4) ' for
v > 4. Given that v > 2, the variance/o. exists and is seen to depend on v and
h. When v > 4, the fourth-order moment/4 exists and, as with/., depends on
just v and h. Since the US t pdf is symmetric about x = 0, the measures of
skewness, discussed in connection with the UN pdf, are all zero, provided,
of course, that moments on which they depend exist. With respect to
kurtosis, we have from (A. 17) and (A. 18) (A. 19) 4 3 =3 1 = -5' v>4. Thus,
for finite v the US t pdf is leptokurtic (yo. > 0), probably because it has
fatter tails than a UN pdf with mean 0 and variance 1/h. As v gets large, the
US t pdf assumes the shape of a UN pdfwith mean 0 and variance/zo. = 1/h.
8 We can obtain the standardized form of the US t pdf from (A.14) by the
following change of variable (^.20) t= v7 (x- 0), -o < t < o. Using (A.20),
we obtain F[(v+ 1)/2] ( (A.21) pOlO = Vl5 r-2) 1 + , --00 < t < 00, which,
with v > 0, is the proper normalized standard US t pdf. This pdf has a single
mode at t = 0 and is symmetric about this point. From (A.20) the moments
of f can be obtained from those of g - 0 when they exist. Proofs of
Properties of the US t Pdf To establish properties of the US t pdf we need
the following results from the calculus: 7 In obtaining these results,
repeated use is made of the relationship I'(q + 1) = qI'(q), q > 0, a
fundamental property of the gamma function. o For v about 30 these two
pdf's are similar; (A.17) and (A.18) show explicitly how and tz4 for the US
t are related to the corresponding moments of the limiting normal pdf; that
is, ta = 1/h and t4 = 3/h u.
(A.22) where 368 APPENDIX A 1. If f(v) I is continuous for a < v < oo and
lim_.o v'f(v) = A, a finite constant, for r > 1, then f If(v)[ dv < oo; that is,
the integral converges absolutely. 9 2. If g(v) is continuous for -oo < v < b
and lim_._.o (-v) r g(v) = c, a constant, for r > 1, then lb_ oo Ig(v)l dv < oo.
3. The relation connecting the beta function, denoted B(u, v), and the
gamma function is x� r(u) r(v) B(u,v) = P(u+ v)' 0 < u,v < o% (A.23) B(u,
v) = f2 x-X(1 - x) - dx, 0 < U,V <00. With these results stated, we first show
that for v > 0 the US t pdf in (A.21) is a proper normalized pdf. xx We note
thatp(tlv) > 0 for -oo < t < 0% and letting t' = t/x/v we can write (A.21) as -
-00 < t' < 00, (A.24) p(t'10 = a , (1 q- , with from the result given in (A.22).
Now make the change of variable z = 1/(1 + t�'), 0 < z < 1, to obtain '
(A.25) P(Z10 = B , zV"-x(1 - z)-6 dz T' 0<z<l. Noting that fp(t'[O dr' = 2 f
p(zlO dz, we have L [ (A.26) p(t'l) dt'= B , z,=-(1 - z)- & = 1, since the
integral on the rhs of (A.26) is just B(1/2, v/2), provided that > O. This
condition is required in order to have f p(t'10 dt' < [see (1) and (2)1. o Note
that I If()l dv < implies lf(v)dv <m; that is, absolute convergen implies
simple convergence. See, for example, D. V. Widder, Advanced Calculus.
New York: Prentice-Hall,. 1947, p. 271. See also p. 273 if. for proofs of (1)
and (2). xo Equation A.22 implies B(u, v) = B(v, u). xx Since (A.21) is
obtained from (A.14) by the change of variable in (A.20), showing that
(A.21) is a proper normalized pdf will imply that (A. 14) also has this
property. x= From z = 1/(1 + t'=), I&ldt'l 2t'/(1 + t'=? and t '= = (1 - z)/z;
thus [dt']&[ = z-%(1 - z)- /2. PROPERTIES OF SEVERAL IMPORTANT
UNIVARIATE PDF'S 369 The results for the odd-order moments in (A. 15)
can be established easily by considering (A.27) f-oo t�'r-X p(t'lv) dt', -oo <
t' < 0% with t' = t/5/' = x/-h (x - 0). For the integral in (A.27) to converge we
need 2r - 1 < by application of (1) and (2). Thus, if 2r - 1 < , the (2r - 1)st
odd-order moment exists and is zero by virtue of the symmetry of p(t'lO
about t' = 0. The expression for the even-order moments in (A. 16) is most
easily obtained by evaluating (A.28) f_ t '"r p(t'lO dt', -oo < t' < oo, with t'=
(x- 0). For the integral in. (A.28) to converge we need v > 2r. If this
condition is satisfied, we use the transformation z = 1/(1 + t''), or t '' = (1 -
z)]z [see (A.24)] to obtain L t'*P(t'lv) dt'= [B(,)] fo () 'zt (1- z) -6 dz (A.29)
= B , z't"-'-(1 - z) - dz B(v/2 - r, r + 1/2) P(r + 1/2) r0/2 - r) = = , >2r. B(1/2,
r(1/2) This gives the 2rth moment ofp(t'lv). From x - 0 = t' we obtain the
2rth moment ofp(x I 0, h, ), la9.r = E('- O) 'r P(r + 1/2) r(v/2- r) () r = r(1/2)
p(./2) , v > 2r, which is just (A. 16). A.3 THE GAMMA (G) AND X �'
Pdf's As the name implies, the G pdf is closely linked to the gamma
function. A random variable, ., is distributed according to the G distribution
if, and only if, its pdf is given by X a- 1 (A.30) p(xly , a) = p(a)y e -x", 0 <
x < oo, where a and y are strictly positive parameters; that is, a, y > 0. From
the form of (A.30) it is seen that y is a scale parameter. When a > 1, the pdf
has
370 n'PENDIX n a single mode 3 at x = y(a - 1). For small values of a the
pdf has a long tail to the right. As grows in size for any given value of y, the
pdf becomes more symmetric and approaches a normal form. The G pdf can
be brought into a standardized form by the change of variable z = x/y,
which results in 1 (A.31) p(zl) = p()z-e -, 0 < z < . From the definition of the
gamma function it is obvious that (A.31) is a proper normalized pdf and
that moments of all orders exist. The moments about zero, denoted by /, are
given by (A.32) r' = f z p(zla) & = r(r + a) P(a) ' r = 1,2, .... From (A.32) we
have for the first four moments about zero x (A.33) '=; a'=(1 +); o'=(2+a)(1
+); and ' = (3 + )(2 + )(1 + ). From these results it is seen that is the mean of
the G pdf and, surprisingly, also its variance. x* Further, for the third and
fourth moments about the mean we have = 2 and = 3(2 + ). Collecting these
results, we have (A.34) x' =, =, =2 and = 3(2+). Given that 0 < < , the
skewhess is always positive. Since the mode is located at z = - 1 for 1, we
have for Pearson's measure of skewness Sk = (mean-mode)/ = 1/. Clearly,
as grows in size, this measure of skewhess approaches zero. As regards
kurtosis, ya = 4/a a - 3 = 6/a, which also approaches zero as grows large.
That Sk 0 and ya 0 as is, of course, connoted with the fact that the G pdf
assumes a normal form as .x The X a pdf is a special case of the G pdf
(A.30) in which a = v/2 and y = 2; that is, the X pdf has the following
formX7: xV12 - i e- x12 (A.35) P(xlv) -- r(v/2)' 0 < x < xa For 0 < < 1 the
pdf has no mode. x4 In obtaining the expressions for the moments below,
we use the relation r(1 + q) = q I'(q) repeatedly. x5 The variance tzo. is in
general related to moments about zero by the following relation- ship: 2 =
to.' - t '2. For the G pdf 2 = (1 + ) - 2 = . x6 Moments, etc., for the
unstandardized G pdf in (A.30) are readily obtained from those for the
standardized pdf (A.31) by taking note of z = x/y. x7 Often (A.35) is written
with x = X 2. PROPERTIES OF SEVERAL IMPORTANT UNIVARIATE
PDF'S 371 where 0 < v. The parameter v is usually referred to as the
"number of degrees of freedom." Since (A.35) is a special case of (A.30),
the standardized form is given by (A.31) with replaced by v/2. Also, the
moments of the standardized form are given by (A.33) and (A.34), again
with = v/2. Since the standardized variable z is related to the unstandardized
variable x by z = x/y = x/2, moments of the unstandardized X a pdf in
(A.35) can be obtained easily from moments of the standardized X �' pdf.
For the reader's convenience we present the moments associated with
(A.35): (A.36) /x'= v; /a = 2v; /3 = 8v; and A very important property of the
X a pdf is that any sum of squared inde- pendent, standardized normal
random variables has a pdf in the X �' form; that is, if = g' + ga a +... + gff,
where the g's are independent, standard- ized normal random variables, then
� has a pdf in the form of (A.35) with V -- .19 A.4 THE INVERTED
GAMMA (IG) Pdf's The IG pdf is obtained from the G pdf in (A.30) by
letting y equal the positive square root of 1 Ix; that is, y = [ '[ and thus y2 =
1 Ix. With this change of variable the IG pdf is 2� 2 (A.37a) P(YlY, ) =
p(a)yy2+ e-, 0 < y < where y, a > 0. Since this pdf is encountered frequently
in connection with prior and posterior pdf's for a standard deviation, we
rewrite (A.37a), letting = y, = v/2, and y = 2/rs ' to obtain (A.37b) p(lv, s) =
2 (___) vo. 1 r(7/2) ,+ e -'�'�, 0 < < where v, s > 0. The pdf in (A.37b) has
a single mode at the following value �-x of (A.38) *moo = S[v--'-'-} '
Clearly, as v gets large, *moo--> s. o These are obtained from (A.34) with =
v/2 and z = x/2, where x is the X �' variable in (A.35), z is the standardized
variable in (A.31), and is the parameter in (A.31). x0 See M. G. Kendall
and A. Stuart, op. cit., pp. 246-247, for a proof of this result. 2o. Since
(A.37a) is obtained from a proper normalized pdf by a simple one-to-one
differentiable change of variable, it is a proper normalized pdf. 2 This is
easily established by taking the log of both sides of (A.37b) and then
finding the value of a for which log p(alv, s) achieves its largest value.
372 a, PENDIX A The moments of (A.37b), when they exist, are obtained
by evaluating the following integral: (A.39) btr' = c f; er-(+ X)e-S2m'2 de,
where 2 On letting y = vs'/2e ', (A.39) can be expressed as (A.40) [v$9.\ (r-
v)l v. f; i r' = �C..-. I y (v-r)19'- Xe- dy. The integral in (A.40) is the
gamma function. For it to converge we must have (A.41) v - r > 0, which is
the condition for existence of a moment of order r. Inserting the definition
of c in (A.40), we have F[(v - r)/2] (v_) rt' (A.42) //= F(v/2) , v > r, a
convenient expression for the moments about zero. The first four moments
are F[(v - 1)/2] [v\ (A.43) m '= [5] s, > 1, 1)() us ' (A.44) r(v/2) s�'= 2' > 2,
- 3)/2 ! % (A.45) I'(v/2) s3' v > 3, and I'(v/2 - 2) �'s4 = 4) s4' v > 4. (A.46)
/4'= ((v]) (v- 2)(v- It is seen from (A.43) that the mean/x' is intimately
related to the parameter s. As v gets large, x' s, which is also approximately
the modal value for large v (see above). In the calculus it is shown that with
0 < q < , as q , qO- (q + a)/(q + b) I for a and b finite. PROPERTIES OF
SEVERAL IMPORTANT UNIVARIATE PDF'S As regards moments about
the mean, '3 we have 373 t 2 = t2 -- /1 2, (A.47) vs ' - v - 2 /x'" v>2, i (A.48)
tza =/zo' - /x/o. -/x'o, v > 3, and (A.49) tz4 =/z' - 4tz'/o - 6/q'atza -/x', v > 4.
These formulas are useful when we have to evaluate higher order moments
of the IG pdf. The Pearson measure of skewhess for the IG pdf is given by
mean - mode Sk= (A.50) - L' - > 2. Since this measure is generally positive,
the IG pdf is skewed to the right. Clearly, as v gets large Sk O. For small to
moderate v the IG pdf has a rather long tail to the right? A.5 THE BETA
(B) Pdf A random variable, �, is said to be distributed according to beta
distri- bution if, and only if, its pdf has the following form: (A.51) p(xla, b,
c) = cB(a, b) 1 - , 0 < x < c, where a, b, c > 0 and B(a, b) denotes the beta
function, shown in (A.23), with arguments a and b. It is seen that the range
of the B pdf is 0 to c. By a change of variable, z = x/c, we can obtain the
standardized B pdf, 1 z_X( 1 _ z),_ x (A.52) p(zla, b) = B(a, b--} , 0 < z < 1,
which has a range zero to one. Some properties of (A.52) follow. That
(A.52) is a proper normalized pdf is established by observing that the pdf is
non-negative in 0 < z < 1 and that B(a, b) = o x z-X(1 - z) -x dz, aa See M.
G. Kendall and A. Stuart, op. cit., p. 56, for formulas connecting moments
about zero with moments about the mean. a Since it is rather
straightforward to obtain the pdf for a n, n = 2, 3,..., from that for in (A.37b)
and to establish its properties, we do not provide these results.
376 APPENDIX A The IB pdf has a single mode if b > 1 at b-1 (A.65)
Umoa = a + 1' Then Pearson's measure of skewness for the IB pdf is (A.66)
Sk= b/(a- 1)- (b- 1)/(a + 1) 2b+a-l[ a-2 j] = a+l 5(a+b-1 ' b>l and a>2, which
is positive and shows that the IB pdf usually has a long tail to the right.
Last, we can obtain an important alternative form of (A.60) by letting u =
y/c, with 0 < c, to yield 1 (y/cy '-' 0 < y < co, (A.67) P(Yla?b'c) = cB(a,b)(1
+ y/c) '+" where a, b, c > 0. Since y -- cu and we have already found the
moments associated with (A.60), the moments for (A.67) are directly
available from (A.61) to (A.64), that is the rth moment about zero is c%.'
with/r' given in (A.61). It will be seen in the next section that the F and US t
pdf's are special cases of (A.67). A.6 THE FISHER-SNEDECOR F PDF A
random variable, , is said to have an F distribution if, and only if, it has a
pdf in the following form' (A.68) p(xJux, vO = B ,' V'o.! (1 + (vdv.)x) (" +
where vx, v. > 0. It is seen that (A.68) is a special case of the IB pdf in
(A.67), where a = d2, b = vx/2, and c -- ./vx. The parameters vx and . are
usually referred to as degrees of freedom and (A.68) is called the F pdf with
x and . degrees of freedom. If vx/2 > 1, the F pdf has a single mode at
(A.69) vo.v/2- 1 Xmoa ,, vd2 + 1 The moments of the F pdf can, of course,
be obtained directly from those associated with the IB pdf shown in (A.62)
t.o (A.64). For easy reference we PROPERTIES OF SEVERAL
IMPORTANT UNIVARIATE PDF'S 377 list the moments of the F pdf: ,
(A.70) I = vx %/2 - 1" > 1, (A.71) /" = (v,./2- 1)(v,./2- 2)' > 2, and so on.
The variance of the F pdf is (Lt ' (d2)[( + 0/2 - 11 v. (A.72) = w/ (d2-
1)2(,,./2- )' 5 > 2. We.now review relations of the F pdf to several other
well-known pdf's. 1. If in the F pdf in (A.68) rx = 1 and we let t 2 -- x, the F
pdf is trans- formed to a standardized US t pdf with . degrees of freedom. 2.
If x and . are independent random variables with X 2 pdf's that have v and
% degrees of freedom, respectively, then = (x/vO/(z'a/vo.) has an F pdf with
vx and vo. degrees of freedom, provided vx, vo. > 0. 3. If � has the F pdf
in (A.68), then, when % --> co, the random variable vx� will have a Xo.
pdf with vx degrees of freedom. 4. If � has the F pdf in (A.68), then, when
vx --> co, the random variable v./ will have a X �' pdf with va degrees of
freedom. 5. If � has the F pdf in (A.68), then, when % --> co with vx = 1,
the random variable / will have a standardized UN pdf. 6. If and ao. are
independent random variables with IG pdf's in the form of (A.37b) and
parameters vx, sx, and vo., so., respectively, the random variable � =
(ax�Jx�')/(ao.�'/so.�) will have an F pdf with va and vx degrees of
freedom. Proposition (1) is established by making the change of variable t
�. = x in (A.68) and noting that with vx = 1 the resulting pdf is precisely
the standard- ized US t pdf with vo. degrees of freedom. Proposition (2) is
established by noting that the joint pdf for x and 3o. is 'p(zl, z2lvl, v2) -----
kzm-Xzata-Xe-% +a )/2, 0 < zx, z2 < co, with 1/k = 2%+a)t�'P(vx/2)P(vd2
). Now let v = zx/zo. and y = (zx + zO/2, which implies that zx = 2vy/(v +
1) and za = 2y/(v + 1). The Jacobian of this transformation is 4y/(v + 1) 2
and thus the pdf for v and y is 9vx/2 - 1 y (vx +va)12-le-lt 0 < v, y, <co. p(v,
ylvx, %) = 2% +va)/2k (1 + v)% +va)/2 , On integrating with respect to y,
the result is p(vl, = + rod2, vd2) (1 + v)% +'-)'" 0 < v < co. Finally, on
letting x = (vdvx)v, we obtain the F pdf in (A.68).
382 A??�NDIX B and (B. 14b) Cov(l[2) = ]11 - Y"I2Y'22 - lye21' The
marginal pdf for xo. can be obtained from (B. 12) by integrating with
respect to the elements of x. Since x appears just in the first factor on the rhs
of (B.12), and this is in the form of a normalized MN pdf, integrating the
first factor with respect to the elements of xx yields 1. Thus the second
factor on the rhs of (B. 12) is the marginal pdf for xo.. It is a MN pdf with
mean vector 0o. and covariance matrix 6 (B.15) Cov(a.) = - By similar
operations the marginal pdf for x is found to be MN, with mean and
covariance matrix (Vx - Last, consider linear combinations of the elements
of ; that is (B.16) $x = Lx, where ' is an rn x 1 vector of normal random
variables with a MN pdf, as shown in (B.2), and Lx is a n x m matrix of
given quantities with n < rn of rank n. Then $x is an n x 1 vector whose
elements are linear combinations of the elements of . If n < m, write I.
(u.17) = = \wo.! \LJ with the matrix L an rn x m nonsingular matrix. Then E
= LE = L0, and we can write (B.18) - L0 = L(' - 0). Then, on noting that the
Jacobian of the transformation from ' to in (B.18) is Iz-l, we can obtain the
pdf for ' (B.19) p(w10, Y, L) = IZ'Zl - (2n)m/2 exp [-�(w - LO)'L-X'Z-XL-
X(w - L0)], we have Z x2 = -5.115.1222-1 and (Zxx) -x = 51xx -zaaa-xaz.
Thus Vxx -x = 5'.11 -- ,12Z22-1E21 and Vxx- x Vx2 = -5'.12'22-1. Since V
= Z-x, we have and thus Z.xVx2 4-22[/22 = I and Z2xVxx + E22Vl = 0.
The second of these relations yields Zx = -Z2Vo. x Vn -x which, when
substituted in the first, yields Z.o.: (V.2 - V2xVn-xV12) -x PROPERTIES
OF SOME MULTIVARIATE PDF'S 383 which is a MN pdf with mean
vector L0 and covariance matrix LZL'. Thus $ = L has a MN pdf. If we
partition w, as shown in (B.17), the marginal pdf for wx will be multivariate
normal with mean vector Lx0 and covariance matrix LxZLx', an application
of the general result associated with (B.12) which gives the marginal pdf for
a MN pdf. B.2 THE MULTIVARIATE STUDENT (MS) t Pdf A random
vector ': (x, o.,..., m) has elements distributed according to the MS t
distribution if, and only if, they have the following pdf: (B.20) p(x10, V, ,
rn) = + m)/2]l + (x - - - < x, < o,i= 1,2,...,m, where v > 0, V is an rn -oo < 0,
< oo, i= 1,2, is PD, the MS t pdf has symmetric about x = 0, v > 1, as shown
below. x rn PDS matrix, and 0' = (0x, 0o.,..., 0), with ..., m. Since the
quadratic form (x - 0)'V(x - 0) a single mode at x = 0. Further, since the pdf
is 0 is the mean of the MS t pdf which exists for The symmetry about 0
implies that odd-order moments about 0, when they exist, will all be zero.
The matrix of second- order moments about the mean exists for v > 2 and is
given by V- [v/(v - 2)]. To establish that (B.20) is a proper normalized pdf
we note that it is positive in the region for which it is defined. If we let
(B.21) x - 0: Cz, where C is an rn x rn nonsingular matrix such that C' VC =
Ira, the pdf for z, an rn x 1 vector, is * (B.22) p(zl,m) = v�'I'[( + m)/2] + -o
< z < oo, i- 1, 2,..., m, the standardized form of the MS t pdf. We show that
(B.22) is a normalized pdf 6 by making the following change of variables
from zx, zo.,..., Zm to 7 Note from C' VC: Ira, I VI A = I Cl - . The
Jacobian associated with the transforma- tion in (B.21) is [C[ and thus
[VI� times the JacobJan is equal to 1. Another way of showing this is to
observe that p(zlv ) can be written as p(zlv ) = p(z,z(,-x), v) p(Jrn_llZ(m_2),
v)... p(Jllv), where z&_y = (zx, z.,. . ., Zm-), j = 1, 2, ..., m - 1. Each factor
is in the form of a US t pdf and can be integrated by using the results of
Appendix A.
388 APPENDIX B and (B.47) Var(.) = v>2, since H -- V/v. For the
conditional pdf in (B.45) we have E(xl,.) = 0x - Hxx-XHx,.(x. - (B.48) = 0x
- Vxx -x Vx.(x. - 0.), m2 + v > 1, and (B.49) Var(l=) = 1 (1 + Q.)H - m.+v-2
_ v (1 + Q.)Vxx -, m. + v > 2, -m.+v-2 where Q. = (x. - 0.)'(H.. - H.xHxx-
XHx.)(x. - 0.). Next consider linear combinations of the elements of a
random vector i with a MS t pdf, as shown in (B.40): (B.50) S = Li, where
L is an m x m nonsingular matrix. Then ES = L0. The Jacobian of the
transformation in (B.50) is IL-l and thus from (B.40) the pdf for S is P[(v +
m)/2] (B.51) p(w10, F, v) = IFlV2[1 + (w - mO)'F(w- mo)] rrm.i,(/2) '
where F = L-X'HL-x. Thus the elements of S have a MS t pdf with mean L0
and covariance matrix equal to (u - 2)-XLH-XL' = [/( - 2)]LV-XL '. The
marginal and conditional pdf's associated with (B.51) are easily obtained by
using (B.44) and (B.45) and, of course, will be in the MS t form. A single
linear combination of the elements of , say S, the first element of S, will
have a marginal US t pdf. Last, as is apparent from many examples cited,
the MS t is related to the MN and IG pdf's. Consider the joint pdf (B.52)
p(x, d0, V, 0 = g(xl 0, ', v) where g(x10, ,, V) denotes an m-dimensional
MN pdf with mean 0 and covariance matrix V-Zo and h(d0 denotes an IG
pdf with parameters v > 0 and s = 1 [see (A.37b)]; that is, (B.53) p(x, crl0,
V, v) k IVIV ( 1 [v + (x - 0)'V(x - 0)]}, = crm++ iexp PROPERTIES OF
SOME MULTIVARIATE PDF'S 389 where k is the normalizing constant.
Then, on integrating (B.53), with respect to , 0 to oo, we have (B.54) p(x10,
v, 0 = k'lvl[ + (x - 0)'V(x - 0)] -(m+,,,'., which is precisely in the form of
(B.20), with k' the normalizing constant. B.3 THE WISHART (W) Pdf The
m(m + 1)/2 distinct elements of an m x m PDS random matrix = {j} are
distributed according to the W distribution if, and only if, they have the
following pdf: (.55) where k- x p(AliZ, ,, m) = k izl,,. exp {- � tr Z-XA},
IAI > o, = 2 '%r'"(-x)t* 1-l[Lx P[( + 1 - i)/2], m < v, and Y. = [rqj], an m x m
PDS matrix. The pdf in (B.55) is defined for the region given by IAI > 0.
We denote the pdf in (B.55) by W(Y., v, m). Some properties of the W pdf
are listed below. 1. If ix, [.,..., [ are m x 1 mutually independent random
vectors, each with4 MN pdf, zero mean vector, and common PDS m x m
covariance matrix Y., the distinct elements of A = '5', where : = (., .,.,..., .,),
have a W(Y., v, m) pdf. Note that 55' = Y.= x [[' has diagonal elements
given by Y[__ x g,/', j = 1, 2,..., m, and off-diagonal elements given by Y.=
x 2sg{, j 4= k = 1, 2,..., m. Thus '/v = , the sample covariance matrix, and
the distinct elements of have a W [(1/05;, v, m] pdf. 2. The distinct elements
of a random matrix J, with the W(Y., v, m) pdf in (B.55), have the following
means, variances, and covariances: (B.56) E = (B.57) Var &j = v(% ' + and
(B.58) Cov(, at) = v(rhe% + cr,%e). Let us partition A and Y.
correspondingly as = { A= mx(Axx Ax., A_ \ A'x A''] ' m - mx\A.x A..] m -
m \Za
390 APPENDIX B where, in each instance, the (1, 1)th submatrix is of size
m x rn and the (2, 2)th submatrix is of size rn - rn x rn - rn. Further, let and
-1 -1 A11.. = (All - Ai.A.. A.l) -- A ll Z�g. = (Zn - Zl,.Z,.,.-lZ,.1) -1 = Z n.
Then the following properties of the W(Y, v, m) pdf are known to hold: 5 3.
The joint pdf for the distinct elements of Axx is W(Zxx, v, m 0. 4. The joint
pdffor the distinct elements of A.. is W [Zn.., v - (m - m0, 5. The marginal
pdf for r. = (B.59) h(rxalt, xa, v) = kx(1 - rxaa)('-a)/a(1 - pxo.�'),/�. with
kx = [(v - + i)], = and x5 f/ ay . (B.60) /(r) = (cosh y - tr)" (B.59) gives the
pdf for a sample correlation coefficient based on v pairs of observations
drawn independently from a bivariate normal pdf with zero means and 2 x 2
PDS covariance matrix Z. Property 1 is a fundamental relationship between
the MN pdf and the W pdf. It is established as follows? The joint pdf for v
normal, mutually independent rn x 1 vectors, ix, [a,..., iv, each with zero
mean vector and common PDS covariance matrix, Z, is (B.61) p(ZIS;, , m)
= 12;I (2rr),m/o. eXp {- � tr Z- xZZ'}, where Z = (z, z.,..., z0, an rn x v
matrix with rn < u. Now make the following transformation Z = TK, where
K is an rn x v matrix such that KK' -- Ira; that is, K is a semiorthogonal
matrix, and T is an rn x rn lower x$ The following well known properties
have been listed in S. Geisser, "Bayesian Estima- tion in Multivariate
Analysis," Ann. Math. Statistics, 36, 150-159 (1965). xo The "cosh"
function is defined by cosh u = (e u + e-U)/2. 't The derivation follows that
presented in S. N. Roy, Some Aspects of Multivariate Analysis. New York:
Wiley, 1957, p. 33. triangular matrix: PROPERTIES OF SOME
MULTIVARIATE PDF'S (B.62) T = ttxx 0 ... t21 t22 ' ' ' , tml tm9. ' ' ' mm
391 with ]T[ = 1-lP=x t, > 0. Note that KK' =Im places m(m + 1)/2
constraints on the elements of K. Thus there are really only urn - m(rn +
1)/2 inde- pendent elements in K. Choose an independent set of elements
from those in K, say (kxx, kxo., . . . , kx,,_. ), (k.x, k.., . . . , k..v_ 0, . . . ,
(kin, k., . . . , k.v_), and call this set K. Thus we can regard the
transformation Z = TK subject to KK' = Im as equivalent to the
transformation from the um elements of Z to the rn(rn + 1)/2 elements of T
and the vrn - rn(rn + 1)/2 elements of K. Then, on substituting Z = TK in
(B.61), we have (B.63) p(T, K,[Z, , m) = Jlzl (2rr)o. exp {- � tr 52- xrT'},
where J denotes the Jacobian of the transformation from the elements of Z
to those of T and K. To obtain an explicit expression for J we use the
following result TM: If y = f(x, x.,..., xv, xv + ,. �., xv+) for i = 1, 2,..., p,
where the xs's, j = 1, 2, ..., p + q, are subject to q constraints, f(xx, x.,..., xv,
xv + x, ...,x v+) = 0 for i=p + 1,p +2,...,p+q, then x� the JacobJan J
associated with the transformation from xx, x,.,..., xv to yx, y,.,..., Yv is
(B.64) J = I (f'f"''"fv'fv+"'"fv+q) + I(fv+,...,fv+q) l :. :. 0(4+. In applying this
result to the present problem, Z = TK takes the place of y = f and KK' - Im
= 0 takes the place off = 0. Further, the elements of K are to be associated
with xx, xo.,..., xv, whereas the remaining elements of K, denoted Ko, are to
be associated with xv + x,..., xv + . Then the Jacobian in (B.60) is [ cq(Z,
_KK') o(gg'- Ira) The explicit expression for the numerator of (B.65) is a�
(B.66) [O(Z, _KK')] = 2 t;_x. This result from the calculus is presented in S.
N. Roy, op. eit., p. 165. o It is assumed that the usual conditions for the
existence of the Jacobian, including the nonvanishing of the numerator and
denominator in (B.64), are satisfied. =o Cf. Roy, op. cit., pp. 170-174.
396 a,P�NDtX B Jacobian of the transformation is [A[- <' + x>oo and thus
(B.71) can be expressed in terms of A = G-X' p(AIH, , m) = k :i,+i exp {- i
tr AH}, IAI > 0. (B.73) lml,ll exp {- tr AH), IA[ > 0, with k given in
connection with (B.71). If, in (B.73), we define Z -x = H, it is seen that
(B.73) is in precisely the form of the W(Z, v, m) pdf shown in Property 2
can be established by noting that Gxx = (Axx - AxA-XAx)-x = A7. As
shown in the preceding section, if A has a W pdf, Axx. also has a W pdf.
Then Property 1 of the IW pdf can be employed to obtain the pdf for Gxx =
A; from that for Axx.. Given this result, the special case in which Gxx is a
scalar, gn, leads to the result in (B.72) which is Property 3. Property 4 is an
immediate consequence of Property 3. B.5 THE GENERALIZED
STUDENT t (GSt) The pq elements of a p x q random matrix, = {iru}, have
the GSt dis- tribution if, and only if, they have the following pdfa': i (B.74)
P(TIP , Q,n) = k iQ + T,PTl,t,., -oo < tj < oo, where k - = rr "t" I-IL-x P[(n -
p - i + 1)J2]Jl-I[-x P[(n - i + 1)J2], n > p + q - 1, and P and Q are PDS
matrices of sizes p x p and q x q, respec- tively. For convenience we denote
the pdf in (B.74) as T(P, Q, O, n), where ao To show that the Jacobian is
IAI-(re+x> write AG = I. Then (cA/cO)G + A(cG/cO) 0 or aG/00 = -
G(OA/OO)G. If 0 = au, we have ag/aau = -gg for ,/, i, and j = 1, 2,..., rn,
with/ < and j <_ i, since G and A are symmetric matrices and the trans-
formation from the elements of G to those of A involves just rn(rn + 1)/2
distinct elements of G. On forming the Jacobian matrix and taking its
determinant, we have IG] '+x = ]AI -('+x). See T. W. Anderson, op. tit., p.
162, for a derivation of this Jacobian which relies on properties of the W
pdf. Also on pp. 348 to 349 Anderson provides the Jacobian of the
transformation G-- A = G -x for the general case in which G and A are not
symmetric, a result not applicable to the present case in which G and A are
symmetric. ax Some call this pdf the matrix t pdf. See J. M. Dickey,
"Matricvariate Generalizations of the Multivariate t Distribution and the
Inverted Multivariate t Distribution," Ann. Math. Statistics, 38, 511-518
(1967); S. Geisser, "Bayesian Estimation in Multivariate Regression," cit.
supra; G. C. Tiao and A. Zellner, "On the Bayesian Estimation of
Multivariate Regression," tit. supra; and the references cited in these works
for further analysis of this pdf. a*. The pdf in (B.74) was encountered in
Chapter 8 in connection with the analysis of the multivariate regression
model. With a diffuse prior pdf the posterior pdf for the regression
coefficients was found to be p(B[ Y) oc JS + (B - ,)'X'X(B - 2)J-,,i9.. If we
let S - 12, P X'X, and T = B --/}, p(BJ Y) is exactly in the form of (B.74).
PROPERTIES OF SOME MULTIVARIATE PDF'S 397 0 appears to denote
that the mean of (B.74), by symmetry, is a zero matrix. Some properties of
(B.74) follow: 1. The pdf in (B.74) can be obtained as the marginal
distribution of p(G, T) = px(G)p.(TIG ), where px(G) denotes an inverted
Wishart pdf and p.(TIG ) denotes a multivariate normal pdf. Let qx q. Px P.
and with qx + q. = q and Px + P. = P and These quantities appear in the
properties of (B.74). �� 2. If T =' (Tx, T0, the conditional pdf for Tx,
given T., is a GSt pdf with parameters (p-x+ TaQaa-XTa,)-x, Qxx.a, T.Qaa-
XQax, n. The mean is Ta Qo. a - X Q ax. 3. If T' = (Xx, XO, the conditional
pdf for Xx, given Xa, is a GSt pdf with parameters Px, Q + Xa'Pa..X., Px-
xPxaXa, n. The mean is Pxx - Px .X'.. 4. If T = (Tx, T0, with Txp x qx and
T.p x q., the marginal pdf for is GSt with parameters P, Q,.,., 0, n - 5. If T' =
(Xx, X0, with Xxpx x q and X,.p,. x q, the marginal pdf for X. is GSt, with
parameters P=,..x, Q, 0, m - p. 6. If in (2) Tx is a p x 1 vector, the
conditional pdf for T, given T., is in the multivariate Student t form.
Similarly, if in (4) T. is a p x 1 vector, it has a marginal pdf in the
multivariate Student t form. 7. With T = (h, t=,..., t), (B.75) p(T) -
P(h)P(t4h)p(tolh, and each of the pdf's on the rhs of (B.75) is in the form of
a multivariate Student t pdf. To establish Property 1 write the IW pdf for the
q(q + 1)/2 distinct elements of G as (B.76) Iw(GIQ,,,,p) = exp{- � tr aa
Some of the following properties established in Chapter 8 draw on the
results in papers by J. M. Dickey, S. Geisser, and G. C. Tiao and A. Zellner,
cited in the preceding footnote.
402 Repeat the procedure results to get Ax ff(x) dx -' 3" (Yo + 4y + 2y. +...
+ 2y,_. + 4y,_ + y,). This is Simpson's rule. APPENDIX C for [x., x4], [x4,
x0],..., [x_.,x] and sum the FORTRAN Programs for Simpsoh's Rule 1.
SimpsoWs Rule 1 In this program the user specifies the function to be
integrated by means of a function-defining statement. He enters the limits of
integration on a data card along with the value of a parameter which we call
TOL. The program then computes a first approximation to the integral and
then does Simpson's rule for 2, 4, 8, 16,... subintervals (at each step, only
the functional values corresponding to the new x-values are actually
computed). After a specified number of subintervals has been reached (16
in the example below) the result from the current computation of Simpson's
rule is compared with the result of the preceding computation. If the
absolute difference between the two is less than the value of TOL,
computation stops and the answer is printed out. If not, the number of
subintervals is doubled and Simpson's rule is computed again and the
procedure is repeated. If the user is not certain of the limits of integration
(e.g., he is approximating (-o% oo) by a finite interval), he simply enters
another data card which specifies the new limits of integration. The obvious
advantage to Simpson's Rule I is that it provides a good idea of the accuracy
of the results for most functions. Also, the ease of changing the limits of
integration is desirable at times. Furthermore, it requires very little
programming effort on the part of the user. On the other hand, use of the
function-defining statement makes it difficult to change the function within
a given run of the program (parameters of the function may, of course, be
changed as shown in the example below). Also, computer time adds up
rapidly as the number of subintervals increases. Finally, this procedure does
not extend easily to bivariate or multivariate integration.' Simpson's Rule I
is illustrated in the computation of the area under several normal densities.
Program statements are shown in Table C.1. 1. After the user's ID card
comes the XEQ card and the DIMENSION statement in which all arrays
are declared. 2. The next statement, which must appear before any
executable state- ments, is the function-defining statement. The following
rules are pertinent to construction of function-defining statements.
FORTRAN PROGRAMS FOR NUMERICAL INTEGRATION 403 (a)
The name of the function must have four to seven characters, the last of
which is F and the first of which is a letter other than X. The name must not
be the same as any of the computer's built-in functions. (b) There may be as
many arguments as the user wishes but they may not be subscripted in this
statement. (c) The following are functions on most computers that may be
useful: SQRTF(X) = LOGF(X) -- loge x EXPF(X) -- e NEXPF(X) = e -
ABSF(X) = Ixl GAMMAF(X) = P(x) LGAMAF(X) = loge P(x) FLOATF(I)
converts a number without a decimal to one with a decimal. Consult a
manual for the particular computer that is being used for information
regarding limitations on the arguments for these functions. 3. Statements I
to 4. Statement 1 instructs the computer to print 'the heading in Statement 2.
Statement 3 instructs the computer to read values of the variables listed
from a data card. They are punched according to the FORMAT of Statement
4. 4. Statements 5 to 8. These statements define the values of the arguments
(other than X) of FUNCF. 5. Statement 9. This statement calculates the
width of the subinterval for the first iteration. 6. Statements 10 to 15. These
statements perform the initial computa- tions for Simpson's rule. 7.
Statements 16 to 20. These statements finish the first computation of
Simpson's rule. 8. Statements 21 and 22. Statement 21 doubles the number
of sub- intervals and Statement 22 halves the width of the subinterval. 9.
Statement 23. This statement instructs the computer to repeat the above if
(the new value of) N is less than 16 and to go on if N is greater than or
equal to 16. 10. Statements 24 to 26. These statements compute Simpson's
rule again. 11. Statement 27. This statement tells the computer to continue
to State- ment 28 if N is less than or equal to 4000 but to go to Statement 39
if N is greater than 4000. This sets a limit on the amount of computation to
be done. 12. Statement 28. This statement computes the current result of
Simpson's rule (actually, three times the result). 13. Statement 29. In this
statement the difference between the current result of the Simpson's rule
computation and the preceding result is compared.
408 APPENDIX C with TOL. If the difference is greater than TOL, the
computer goes to Statement 30. If it is less than or equal to TOL, it goes to
Statement 35. 14. Statements 30 to 34. These statements perform the initial
computations for the next computation of Simpson's rule. 15. Statements 35
to 38. Statement 35 computes the final result and State- ment 36 prints out
the results according to the FORMAT of Statement 37. Statement 38 sends
the computer to Statement 41, which is the last statement in the DO-loops
started in Statements 5 and 7. 16. Statements 39 to 40. These statements
print out the results (D(1) = three times the last result of Simpson's rule) in
the event that N is greater than 4000. 17. Statement 42. This statement
sends the computer back to Statement 3 which instructs the computer to
read another data card and proceed. If there are no more data cards,
computation stops. 18. Data Cards. These cards are punched according to
FORMAT Statement 4. The second card provides a check on the
computations done with the first data card by computing the tail area. 2.
Simpson's Rule H In this program the computation of Simpson's rule is
performed by the subprogram FUNCTION FUNC1 (UP, SL, MM).
Whenever FUNC1 is used, the user specifies the upper limit of integration
(UP), the lower limit (SL), and the number of subintervals (MM). The user
writes his own main program to compute the values of the integrand at the
endpoints of all the subintervals and to store them in an array (called W in
the example below). After the computation of the W array is completed the
next statement is of the form ANSWER = FUNC1 (.,., .). This is all that is
required to perform the computation of Simpson's rule. If another integral is
desired, the values of this integrand at the endpoints of its subintervals are
computed and placed in the W array and FUNC1 is used again. Naturally
the values of UP, SL, and MM may be changed each time. The main
advantage of Simpson's Rule II is that several different integrals may be
computed easily in one program. Also, as shown below, it can be extended
easily to multivariate integrations. However, it requires more pro-
gramming on the part of the user and within a single run provides no idea of
the accuracy of the results. Simpson's Rule II is illustrated in the
computation of the area under several normal densities and its first and
second moments. Program statements are shown in Table C.2. 1. After the
XEQ card and the DIMENSION statement the statement COMMON W is
included. It occupies a similar position in the FUNC1 FORTRAN
PROGRAMS FOR NUMERICAL INTEGRATION 409 subprogram. This
tells the computer that W in the subprogram is the same array as in the main
program. 2. Statements 1 and 2. Statement 1 prints out the title shown in
FORMAT Statement 2. 3. Statements 3 to 7. Values of SIGMA and XBAR
are computed here as well as the constant of the integrand� 4. Statements
8 to 12. A loop is set up to compute the various values of Z, the variable of
integration, and to compute the various integrantis. In Statement 9 we have
picked the lower limit of integration Z = -30. and the width of the
subinterval = . 30. The upper index of the DO (Statement 8) tells us the
number of subintervals (201 - 1 = 200)� These facts imply that the upper
limit of integration is + 30. 5. Statement 13. FUNC1 is used to compute the
area. 6. Statements 14 to 19. The values of the X (first moment) array are
placed in the W array and FUNC1 performs the integration for the first
moment (Statements 14 to 16). Similarly, the computation for the second
moment is done. 7. Statements 20 and 21. The results are printed out.
Statement 20 ends the loops started in Statements 3 and 5. After Statement
20 is completed, the computer returns to Statement 5 and the Statements 7
to 20 are executed again for the new value of ZBAR. After this is
completed, it returns to Statement 3 and does Statements 5 to 20 for the new
value of SIGMA. When this is done the computation stops and the program
ends via CALL EXIT and END. 8. FUNCTION FUNC1 subprogram
follows the main program as shown. Suppose we wished to check the
accuracy of our results by doubling the number of subintervals. The way the
program is now written we have lost the original entries in the W array and
would thus have to recompute them. We still,have the values of X and Y,
however. We can do the following: set up new arrays, XX and YY. Transfer
X(I) to XX(2 � I) and Y(I) to YY(2 � I). Then we.need only compute
XX(3), XX(5),... XX(401) and YY(3), YY(5), � .. YY(401). (i.e., only the
new values) and perform the integration by trans- ferring XX and YY to W.
The same procedure could be followed for the original values in the W
array by transferring them to another array. C.3 BIVARIATE
INTEGRATION General Here the problem is of the form
424 Holt, C., 322 Hood, W. C., 64,258 Houthakker, H. S., 159 Huang, D.S.,
240 Hume, D., 2 AUTHOR INDEX Modigliani, F., 145,322 Moore, E. H.,
77 Moore, H., 1 Morgan, J. N., 37 Muth, J. F., 322 James, W., 115 Jeffreys,
H., vii, 2,4, 5, 7, 8, 9, 11, 12, 20, 31, 33, 35, 40, 42, 43, 44, 45, 46, 47, 48,
49, 50, 51, 52, 53, 58,153, 190, 217,219, 225,226,228,255, 256,289,292,
302, 304,305,306, 360, 395 Johnson, R. A., 32 Johnston, J., 124 Jones, R.,
145 Kakwani, N. C., 243 Kalaba, R., 348 Kaufman, G. M., 75, 23,3
Keeping, E. S., 375 Kendall, M. G., 114, 115,118, 119, 127, 164,
176,357,366,371, 373,384 Kenney, J. F., 375 Kiefer, J. C., 127,333 Klein,
L. R., 194 Kmenta, J., 177,240, 243,289,327 Koopmans, T. C., 64,258
Koyck, L., 200 Kullback, S., 256 Kuznets, S., 5 Laub, P.M., vii Le Cam, L.,
31 Leenders, C. T., 271 LindIcy, D. V., vi, vii, 20, 21, 31, 33, 40, 53,
58,124, 129,141,298,299, 300, 301,303,304, 318 Lu, J. Y., 96 Lucas, R.,
208,314 Luce, R. D., 292 Mach, E., 5 Madansky, A., 126 Maddala, G. S.,
208, 314 Marsaglia, G., 128,281 Martin, J. E., 96,203 Meiselman, D.,
261,306,323 Miller, M. H., 145 Minhas, B., 169 Nerlove, M., 248 Neyman,
J., 31,115,118, 119 Noble, B., vi Ockham, W., 8 Orcutt, G. H., 190 Park, C.
J., vii, 203 Pearson, K., 1, 39,365,370, 373, 374,376 Peers, H. W., 286
Penrose, R., 77 Pierce, G. S., 5 Plackett, R. L., 42 Pointare, H., 5 Popper, K.
R., 3 Prescott, E. C., 345,346,347,349,350, 351,352,353,354 Press, S. M.,
vii, 75,279, 330, 331 Price, R., 11 Raiffa, H., vii, 21, 58, 75,292, 375,387
Ramsey, J. B., 163 Rao, C. R., 77,329 Reichenbach, H., 2 Reiersol, O., 128,
130 Renyi, A., 20 Revankar, N. S., vii, 176,177,179,183 Richard, J. F., vii
Richardson, D. H., 147 Roberts, H. V., vii, 276 Robinson, E. A. G., 159
Rothenberg, T. M., 239,258,271 Roy, S. N., 390, 391,392 Samuelson, P. A.,
194, 195 Sankar, U., vii, 131,133,154, 169 Savage, L. M., vii, 14, 46,226
Sawa, T., 147 Schlaifer, R., vii, 21, 58, 75,375,387 Scott, E., 118, 119
Shannon, C. E., 43 Sharma, D., vii Simon, H. A., 322 Smffnov, N. V., 27,
64 AUTHORINDEX 425 Sobel, E., 146 Solow, R., 169,314,315 Stein, C.
M., 115,140 Stone, M., vii, 26 Stuart, A., 114, 115,118, 119, 127,164, 176,
357,366,371,373,384 Summers, R., 281 Swamy, P. A. V. B., vii, 145 Tan,
W. Y., 129 Theil, H., 100, 194, 272, 322 Thornber, H., vii, 26, 95,169,
171,172, 190, 201,307,312 Tiao, G. C., vii, 27, 46, 96, 98, 112, 129, 224,
229,252, 396,397 Tocher, K. D., 230 Varga, R. S., 209 Wallace, N., 208,
314 Watson, G. S., 146 Watts, D. G., 320, 333 Weaver, W., 43 Welch, B. L.,
286 Widder, D. V., 368 Winkour, H. S., Jr., 190 Wolfowitz, J., 127 Wright,
R. L., 154 Zarembka, P., 164 Zellner, A., 75,96, 98, 112, 131,133,145, 154,
176,177, 179, 183, 191,194, 203,207,210, 211,224,229,233, 240, 243,252,
261,267,272,289, 312, 320, 327,331,333,336, 396,397
Subject Index Absolute error loss function, 24, 25, 333, 334 Adaptive
control solution, 340ff Aimon distributed lag technique, 221 Appraisal of
alternative control solutions, 322-324, 334-336, 343-344, 351-353
Approximate posterior pdf's, 33, 34, 47, 96, 101-106, 110-112, 238,240
Asymptotic expansions, 110-112 Autocorrelation, 86-97 Autoregressive
process, first order, 186-191, 216-220 fast order with incomplete data, 191-
193 second order, 194-200 Average risk, 26 Bayes biographical note, 11
Bayes-Laplace procedure, 41, 42 Bayesian estimator, 26-27 Bayes'
Theorem, 10-11, 13-14 several sets of data, 17-18 Beta function, 36%368
Beta pdf, 373-375 Beta prime pdf, 375-376 Binomial distribution, 3840
Bivariate moment-cumulant inversion formulas, 112 Box:Cox analysis of
transformation, 162ff Certainty equivalence approach, 322ff Chi-square
pdf, central, 370-371 noncentral, 156 Chow's model of U.S. economy, 350-
351 Cobb-Douglas production function, 69, 84, 162, 177,182, 289
Comparing hypotheses, 291ff Conditional posterior pdf, 21 Confidence
intervals and regions, 27-28 Constant elasticity of substitution production
function, 162, 169ff Constant returns to scale, 69, 173 427 Control
problems, adaptive control solution, 340.343 adaptive decision rules, 352
certainty equivalence solution, 322ff cost of changing control variable, 324-
325,330-331 here and now solution, 338, 343 linear decision rules, 351-352
money multiplier model, 323-3 24 monopoly problem, 325-327 multiperiod
problems, 344ff multiple regression process, 327-331, 336-343 multivariate
regression process, 331-333 one period problems, 320ff perfect information
decision rules, 352 sensitivity of solutions to form of loss function, 333-336
sequential updating solution, 338-340, 343 simple regression process,
320.325, 333-336 simultaneous equation process, 346ff two period
problem, 336-343 Convergence of integrals, 368 Correlation coefficient,
228 Covariance matrix, diffuse prior pdf, 225-227 posterior pdf in
multivariate regression, 227 Deductive inference, 24 Degrees of belief and
probability, 8-10 Direct probability, 13 Distributed lag models, 220ff Almon
technique, 221 application to consumption function, 207ff, 312-317
generalizations, 213ff Solow's family, 314-317 Drze's results on
identification, 257-258 Dynamic properties of models, 194ff