Introduction To Probability Theory
Introduction To Probability Theory
K. Suresh Kumar
Department of Mathematics
Indian Institute of Technology Bombay
LECTURES 24-25
Now we will state a result which is useful, its proof is beyond the scope
of this course.
Theorem 0.1 (Continuity Theorem) Let Xn , X be random variables on
(Ω, F, P ) such that,
lim ΦXn (t) = ΦX (t), t ∈ R.
n→∞
fY |X (y|x) = P (Y = y|X = x) if P (X = x) 6= 0
= 0 if P (X = x) = 0 .
Intuitively, fY |X means the pmf of Y given the information about X. Here
information about X means knowledge about the occurrence (or non occur-
rence) of {X = x} for each x. One can rewrite fX|Y in terms of the pmfs as
follows.
f (x, y)
fY |X (y|x) = if fX (x) 6= 0
fX (x)
= 0 if fX (x) = 0 .
0.2. CONDITIONAL PDF OF CONTINUOUS RANDOM VARIABLES 3
Definition of fX|Y is similar. Also one can relate fY |X and fX|Y through the
Bayes theorem as follows.
fX|Y (x|y)fY (y)
fY |X (y|x) = ,
fX (x)
where fY denote the marginal distribution of Y .
Definition of fY |X is similar and also we can see that for x, y with fX (x), fY (y) 6=
0,
fX|Y (x|y)fY (y)
fY |X (y|x) = , (0.1)
fX (x)
this can be treated as a variation of Bayes theorem for densites.
G(0, y) = P {X = 0, Y ≤ y}
= P {Y ≤ 0, Y ≤ y}
Ry
= −∞ f (z)dz if y ≤ 0
1
2 if y > 0.
Similarly
0R if y ≤ 0
G(1, y) = y
0 f (z)dz if y > 0.
Now
dG(0, y) f (y) if y < 0
=
dy 0 if y > 0,
dG(1, y) 0 if y < 0
=
dy f (y) if y > 0.
0.3. CONDITIONAL PMF/PDF: FOR A COMBINATION OF RANDOM VARIABLES5
Hence
dG(x, y)
fY |X (y|x) = 2
dy
2f (y) if y < 0, x = 0 or y > 0, x = 1
=
0 if otherwise,
Also
fY |X (y|x)fX (x)
fX|Y (x|y) =
fY (y)
fY |X (y|x)
=
2f (y)
1 if y < 0, x = 0 or y > 0, x = 1
=
0 if otherwise,
Example 0.2 Consider the random experiment given as follows.
• Pick a point at random from the interval (0, 1) say x
• Now pick another point at random from the interval (0, x).
Find how is the second point distributed.
Let X denote the position of the first point and Y denote the position of
the second point.
Then X be uniform random variable over (0, 1) and the conditional pdf
of Y given X is given by
1 0<x<1
fX (x) =
0 otherwise .
Note that the pdf of Y given X = x is U (0, x), i.e.
( 1
0<y<x<1
fY | X (y | x) = x
0 otherwise .
Also ( 1
0<y<x<1
f (x, y) = fX (x)fY | X (y | x) x
0 otherwise .
Hence
Z ∞ Z 1
1
fY (y) = f (x, y)dx = dx = −ln y, 0 < y < 1 .
−∞ y x
6
Definition 9.4. If X, Y are random variable. Then for −∞ < a < b < ∞,
we define
X
P (a ≤ Y ≤ b | X = x) = fY |X (y|x) if Y is discrete
y∈DY ∩[a, b]
Z b
= fY |X (y|x)dy if Y is continuous.
a
Note that
1
f (x, y) = , 0 ≤ x2 + y 2 < R, = 0 other wise.
πR2
Also for −R < x < R,
Z ∞
fX (x) = f (x, y)dy
−∞
Z √R2 −x2
1
= √ dy
− R2 −x2 πR2
√
2 R 2 − x2
= .
πR2
0.4. CONDITIONAL PROBABILITIES 7
For 0 ≤ x2 + y 2 < R,
f (x, y)
fY |X (y|x) =
fX (x)
1
= √ .
2 R 2 − x2
Hence for −R < x < R,
√
Z R2 −x2
1
P Y > 0X = x = √ dy
0 2 R 2 − x2
1
= .
2
Example 0.4 Let X, Y be i.i.d. exp(λ). Find P ({X > 1|X + Y = 2}).
Let f denote the joint pdf of X, Y and g denote the joint pdf of X, Z =
X + Y . Then
f (x, y) = λ2 e−λ(x+y) , x, y > 0, = 0 otherwise.
Recall that
1
g(u, v) = f ((u, v)A−1 ),
|detA|
when (U, V ) = (X, Y )A and f, g denote respectively the pdf of (X, Y ) and
(U, V ). Using this, it follows that
g(x, z) = λ2 e−λz , 0 < x < z, = 0 otherwise.
Hence using the definition of fX|Z , we get for 0 < x < z,
g(x, z)
fX|Z (x, z) =
fZ (z)
λ2 e−λz
= Rz
0 g(x, z)dx
λ2 e−λz
=
zλ2 e−λz
1
= .
z
Also fX|Z (x|z) = 0 otherwise. Now
Z ∞
P (X > 1|X + Y = 2) = fX|Z (x|2)dx
1
Z 2
1 1
= dx = .
1 2 2
8
x1 x2 xn+1 n
P {La1 > x1 , · · · Lan+1 > xn+1 } = 1 − − − ··· − ,
a a a +
Now for x1 + x2 > a, it is clear that P {La1 > x1 , La2 > x2 } = 0. Hence
x1 x2
P ({La1 > x1 , La2 > x2 }) = 1 − − .
a a +
m t m−1
f (t) = 1− , 0 ≤ t ≤ a.
a a
0.5. CONDITIONAL EXPECTATION 9
Hence
Z a
P {La1 > x1 , · · · Lam+1 > xm+1 } = P {La1 > x1 , · · · Lam+1 > xm+1 |La1 = x}
0
m x
(1 − )m−1 dx
Z aa a
x2 xm+1 m−1
= (1 − − ··· − )
x1 a−x a−x +
m x
(1 − )m−1 dx
a a
m a
Z
= (a − x − x2 − · · · − xm+1 )m−1
+ dx
a m x1
m (a−x1 −x2 −···−xm+1 )+ m−1
Z
= t dt
am 0
1
= (a − x1 − x2 − · · · − xm+1 )m+.
am
This completes the proof by induction.
The above example gives the distribution of the random partition (using
uniform distribution )of interval [0, a). Also note that Lak gives the kth order
statistc of uniform disrtibution on [0, a).
Set
Z = X +Y .
P (Y = y | Z = n) = 0 if y ≥ n + 1 .
For y = 0, 1, 2, · · · , n
P {Y = y, X + Y = n}
P (Y = y|Z = n) = .
P {X + Y = n}
Now
n
X
P {X + Y = n} = P {X = x, Y = n − x} = (n + 1)p2 (1 − p)n .
x=0
P {Y = y, X + Y = n} = P {Y + y, X = n − y}
= P {Y = y}P {X = n − y}
= p(1 − p)y p(1 − p)n−y .
Therefore
1
P (Y = y|Z = n) = .
n+1
i.e.,
1
z+1 if y = 0, 1, 2, · · · , n
fY |Z (y|z) = .
= 0 otherwise
Now
n
X X 1 1 n(n + 1) n
E[Y |Z = n] = yfY |Z (y|n) = k = = .
y
n+1 n+1 2 2
k=1
When X and Y are discrete random variables, E[Y |X] is defined using
conditional pmf of Y given X. Hence we define E[Y |X] when X and Y are
continuous random variables with joint pdf f in a similar way as follows.
Now we state a very useful theorem.
0.5. CONDITIONAL EXPECTATION 11
We state our final theorem in this chapter which is about the conditioning
method we mentioned in the begining of the chapter and is our destination
for this chapter.
Example 0.7 Let X, Y be continuous random variables with joint pdf given
by
6(y − x) if 0 ≤ x ≤ y ≤ 1
f (x, y) = .
0 otherwise
Find E[Y |X = x] and hence calculate EY . Note that
Z 1
fX (x) = 6(y − x)dy = 3(x − 1)2 , 0 ≤ x ≤ 1
x
Example 0.8 Let X ∼ uiniform (0, 1) and Y ∼ uniform (0, X). Find
E(X + Y )2 .
Note that what is given is fX (x) = 1, 0 < x < 1, = 0 otherwise and the
conditional pdf
1
fY |X (y|x) = , 0 < y < x, = 0 otherwise.
x
Hence
Z 1
2
E(X + Y ) = E[(x + Y )2 |X = x]dx
0
Z 1Z ∞
= (x + y)2 fY |X (y|x)dydx
0 −∞
1Z x
(x + y)2
Z
= dydx
0 0 x
Z 1 Z 2x
1 7
= z 2 dzdx = .
0 x x 9