Dips Academy Statistics Notes
Dips Academy Statistics Notes
0
Conditional Distribution For jointly continuous random variables x and Y,
Sex vis)=[fnx(els)de vx such that f(x) >0
4.4. Conditional Expectation
We can also ask what the expected behaviour of one random variable is,
given knowledge of the value of a second random variable and this gives
tise to the idea of conditional expectation.
Definition 4.10 (Conditional Expectation): The conditional expectation in
discrete and continuous cases corresponds to an expectation with respect 10
the appropriate conditional probability distribution:
Diserete
E[Y|X=s]= DvP (Y= 91K =x)
Continuous
ELL = 2]=[° ohne (V1 = 2),
Note that before x is known to take the value x, E[}’| X] is itself a random
variable being a function of the random variable x . We might be interested
in the distribution of the random variable E[Y| x).
Theorem 4.1 (Tower Property of Conditional Expectation): For any two
random variables X, and X,
sR Nae LL Nw DIOLS, Ps 0 RHE ak PATRI
wee wn aoucsdet com pee 20
(a marran ms
Scanned with CamScanneree
ELEN 1 42))=EL%]
Exercise 44.1: Suppose that © ~ Ufo, 1} and (x |@)~8in(2, @)
Find B[4|@] and hence or otherwise show that E[X]=1
4.5. Conditional Expectations of Functions of Random Variables
By extending the theorem on marginal expectations we can relate the
conditional and marginal expectations of functions of random variables (in
particular, their variances).
Theorem 4.2: (Marginal Expectation of a ‘Transformed Random
Variables): For any random variables X; and X;, and for any function
40
{EL A(%)) X2]]= ELA)
Theorem 4.3 (Marginal Variance): For any random variables X, and x,
Var( X,) = [ Var(%; |2)]}+Var(
M1 4%2])
46, Independence of Random Variables
Whilst the previous sections have been conceméd with the information that
‘one random variable carries about another, it would seem that there must be
pairs of random variables which each provide no information whatsoever
about the other. It is, for example, difficult to imagine that the value obtain
when a die is rolled in Coventry will tell us much about the outcome of a
coin toss taking place at the same time in Lancaster.
There are two equivalent statements of a property termed stochastic
independence which capture precisely this idea The following two
definitions are equivalent for both discrete and continuous random variables
Definition 4.11: (Stochastic Independence): Definition 1 Random
variables X,,X2,...X, are stochastically independent iff
nals) “FTF 0)
Definition 2: Random variables. ),,,....%, are stochastically.
independent iff
foe s)=T1 f(s)
If X; and X; are independent then their conditional densities are equal to
their marginal densities
4.7. Covariance and Correlation
Having established that sometimes one random variable does convey
information about another and in other cases knowing the value of a random
variable tells us nothing useful about another random variable itis useful to
have mechanisms for characterising the relationship between pairs (or larger
‘groups) of random variables
Definition 4.12 (Covariance and Correlation): Covariance: For random
variables ’ and ¥ defined on the same probability space
THAT, Ft Paria Sara Haat Khar
rau lafaaipracanvcom: Webate: wor istndctesom
Scanned with CamScanner
Joint and Conditional Distributions
Re Dab 11016, Ph (268759, Cas MTR vam
ee
Put Your Own Notes48.
Ty
Coo. Y] FLX He (Hr)
EL] -ney
Correlation: For random variables x
probability space
Cov[.x,¥]
and Y defined on the same
CovfX,7]
ola] Wvar[x]Ver[?]
provided that ¢y >0 and oy >0.
): Let X and Y have finite
Theorem 4A (Cauchy-Schwart Tneqt
second moments. Then
f°]
2X]=1 for some constant ¢
e[xy]) =|e[x yp se[ x
( elevy’ sel,
with equality if and only if PLY
‘Transformation of Random Variables:
(x)
Theorem 4.5 (Distribution of a Function of a Random Variable
be a random variable and Y= g(X) where g is injective (ic. it maps at
‘most one x to any value y). Then
)
de
dy
fel») = fe(27"09)}
given that (2"'(y)) exists and (g°'(y)) >0 vy or (g"(y)) <0 vy. If g
is not bijective (one-to-one) there may be values of for which there exists
no x such that y= g(x). Such points clearly have density zero,
When the conditions of this theorem are not satisfied it is necessary to be a
little more careful. The most general approach for finding the density of a
transformed random variable is to explicitly construct the distribution
function of the transformed random variable and then to use the standard
approach to turn the distribution function into a density (this approach is
discussed in Larry Wasserstein’s “All of Statistics”)
Exercise 4.8.1: Let X be distributed exponentially with parameter 0, that
ee x20
Il eco
Find the density function of
ne a ee
ae 0
Gi) y=x"%, poo
0. for x<0
(ii) ¥=e(X) with g(X)=|x for Osxst
1 forx>t
(
Deli 0016, Ps (12657527, Cal 99918) 59117, RROD
Scanned with CamScanner2> ips ccademy
Theorem 4.6 (Probability Integral Transformation): If X is a random UT
variable with continuous Fy (x), then U = Fy (X) is uniformly distributed Pat Your Own Notes
over the interval (0, 1). ——
Conversely if U is uniform over (0,1), then X =F! (U) has distribution
funetion Fy
49. Moment-Generating-Function Technique
‘The following technique is but one example of a situation in which the
‘moment generating function proves invaluable.
Function of Vail. For Y= g() compute
iy-efe"] -efee()]
m(pe ele" }=e[ 4]
If the result is the MGF of a known distribution then it will follow that ¥
hhas that distribution,
Sums of Independent random variables. For ¥ =F, Xj, where the X; are
independent random variables for which the MGE exists V—A<1You might also like