Week 3
Week 3
2. Consider an experiment whose sample space is Ω. For each event A of the sample space (Ω, F),
we assume that a number P (A) is defined and satisfies the following three conditions:
(i) 0 ⩽ P (A) ⩽ 1.
(ii) P (Ω) = 1.
(iii) For any events A1 , A2 , . . . that are mutually exclusive, then
(∞ )
∪ ∑∞
P An = P (An ) .
n=1 n=1
We refer to P (A) as the probability of the event A. The triple (Ω, F, P ) is called a probability
space.
3. Properties of Probability.
(i) If A ⊂ B, then P (A) ⩽ P (B).
(ii) P (Ac ) = 1 − P (A).
(iii) P (A ∪ B) = P (A) + P (B) − P (A ∩ B).
∑n
(iv) P (∪ni=1 Ai ) = i=1 P (Ai ) if Ai are mutually exclusive.
∑∞
(v) Boole’s inequality: P (∪∞ i=1 Ai ) ⩽ i=1 P (Ai ).
Then
P (A) = lim P (Ai ) .
i→∞
Page 1
SUStech Probability and Some Calculations Applied Stochastic Process
P (B | A)P (A)
P (A | B) = .
P (B)
9. If {X ⩽ x} ∈ F for all x ∈ R, we say X is measurable with respect to F, and people often use
the shorthand notation X ∈ F .
10.For a random variable X we define σ(X) to be the smallest sigma field which make X measurable
with respect to σ(X). We read it as ”the sigma field generated by X ′′ . For random variables X, Y , we
say that X is Y -measurable if X ∈ σ(Y ).
Example. Let Ω = {a, b, c} and A = {{a, b, c}, {a, b}, {c}, ∅}, and we define X, Y, Z as follows:
ω X Y Z
a 1 1 1
b 1 2 7
c 2 2 4
Page 2
SUStech Probability and Some Calculations Applied Stochastic Process
which always makes sense, but may be ∞. If X is a general random variable, then
EX = EX + − EX −
which is the expectation formula of the continuous random variable. If F is a step function with jumps
at {xi , i ⩾ 1}, then
∫ ∑ ∑
g(x)dF (x) = g (xi ) (F (xi ) − F (xi −)) = g (xi ) f (xi ) .
R i⩾1 i⩾1
ϕ is the mgf since we can obtain all moments from it.In fact, by Taylor’s formula, we have
∑
∞
ϕ(n) (0)
ϕ(t) = tn ,
n=0
n!
Page 3
SUStech Probability and Some Calculations Applied Stochastic Process
and
( ) ∑
∞
(tX)n ∑
∞
E (X n )
ϕ(t) = E e tX
=E = tn .
n=0
n! n=0
n!
Thus,
E (X n ) = ϕ(n) (0).
14.
V (X) = E(V (X | Y )) + V (E(X | Y )).
Homework
13. The dice game craps is played as follows. The player throws two dice, and if the sum is seven or
eleven, then she wins. If the sum is two, three, or twelve, then she loses. If the sum is anything else,
then she continues throwing until she either throws that number again (in which case she wins) or she
throws a seven (in which case she loses). Calculate the probability that the player wins.
Solution:
Note that
i 2 3 4 5 6 7 8 9 10 11 12
1 2 3 4 5 6 5 4 3 2 1
P 36 36 36 36 36 36 36 36 36 36 36
Let Ai be the event that throw i at the first time and win the game. Then
{
2
9
, i = 7, 11
P (Ai ) =
0, i = 2, 3, 12
Now we calculate it for other i. Denote by p the probability of getting i at each trial, r the
probability of not i and 7 . Then,
∑
∞
P (Ai ) = P (Ai , win in n + 1 trials )
n=1
∑∞
p2 p2
= prn−1 p = = .
n=1
1−r 1
6
+p
Hence
1
, i = 4, 10
36
P (Ai ) = 2
, i = 5, 9
45
25
, i = 6, 8
396
∑
12
P ( win ) = P (Ai ) ≈ 0.49
i=2
30. Let X be a Poisson random variable with parameter λ. Show that P (X = i) increases mono-
tonically and then decreases monotonically as i increases, reaching its maximum when i is the largest
Page 4
SUStech Probability and Some Calculations Applied Stochastic Process
Solution:
0, y ≤ −1
[ ]
y3
F (y) = 3
y− + 2
, −1 < y < 1
4 3 3
1, y≥1
43. An urn contains n + m balls, of which n are red and m are black. They are withdrawn from the
urn, one at a time and without replacement. Let X be the number of red balls removed before the first
black ball is chosen. We are interested in determining E[X]. To obtain this quantity, number the red
balls from 1 to n. Now define the random variables Xi , i = 1, . . . , n, by
1, if red ball i is taken before any black ball is chosen
Xi =
0, otherwise
Solution:
(b)
( )
∑
n ∑
n
E(X) = E Xi = E (Xi )
i=1 i=1
∑
n
1
= ( consider i th red and all black)
i=1
m+1
n
=
m+1
Page 5
SUStech Probability and Some Calculations Applied Stochastic Process
(a)
∑
∞ ∑
∞
E[X] = P (X ≥ n) = P (X > n)
n=1 n=0
Note that,
∑
∞
X= In .
n=1
( )
∑
∞ ∑
∞
E(X) = E In = E (In )
n=1 n=1
∑
∞
= P (X ≥ n)
n=1
∑∞
= P (X > n).
n=0
(b) Let
1, m ≤ Y
Jm =
0, m > Y
Similarly,
∑
∞
Y = Jm .
m=1
Page 6
SUStech Probability and Some Calculations Applied Stochastic Process
Then [( ) ] ( )
∑
∞ ∑
∞ ∑
∞
E(XY ) = E In Y =E In Y = E (In Y )
n=1 n=1 n=1
[ ( ∞ )]
∑
∞ ∑
= E In Jm
n=1 m=1
∑∞ ∑∞
= E (In Jm )
n=1 m=1
∑∞ ∑ ∞
= P (X ≥ n, Y ≥ m)
n=1 m=1
67. Calculate the moment generating function of the uniform distribution on (0, 1). Obtain E[X] and
Var[X] by differentiating.
Solution: Note that ∫
( ) 1
1( t )
ϕ(t) = E etX = etx dx = e −1 ,
0 t
Then
1∑ 1 m ∑ 1 1 n
∞ ∞
ϕ(t) = t = t .
t m=1 m! n=0
n! n + 1
Hence,
1
ϕ(n) (0) = .
n+1
Thus,
1 1
ϕ′ (0) = , ϕ′′ (0) = .
2 3
Hence
1
E(X) = ϕ′ (0) =
2
( ) 1
Var(X) = E X 2 − (E(X))2 = ϕ′′ (0) − (ϕ′ (0)) =
2
12
Supplementary exercises
1. (Some Probability Identities). Let A1 , A2 , . . . , An denote events and define the indicator
variables Ij , j = 1, . . . , n by
1 if Aj occurs
Ij =
0 otherwise.
Letting
∑
n
N= Ij ,
j=1
then N denotes the number of the Aj , 1 ≤ j ≤ n, that occur. A useful identity can be obtained by
noting that
1 if N = 0
(1 − 1)N =
0 if N > 0.
Page 7
SUStech Probability and Some Calculations Applied Stochastic Process
Hence, if we let
1 if N > 0
I=
0 if N = 0,
then yield ( )
∑
n
N
1−I = (−1)i
i=0 i
or ( )
∑
n
N
I= (−1)i+1
i=1 i
Taking expectations
[( )] [( )]
N N
E[I] = E[N ] − E + · · · + (−1) n+1
E (1)
2 n
However,
E[I] = P {N > 0}
= P { at least one of the Aj occurs }
(n )
∪
=P Aj
j
and [ ]
∑
n ∑
n
E[N ] = E Ij = P (Aj ) ,
j=1 j=1
[( )]
N
E = E [number of pairs of the Aj that occur]
2
[ ]
∑∑
=E Ii Ij
i<j
∑∑
= E [Ii Ij ]
i<j
∑∑
= P (Ai Aj )
i<j
Page 8
SUStech Probability and Some Calculations Applied Stochastic Process
− · · · + (−1) n+1
P (A1 A2 · · · An ) .
and so [ { }]
∑
N [ ]
N
ΨY (t) = E exp t Xt = E (ΨX (t))
i
Evaluating at t = 0 gives
E[Y ] = E[N E[X]] = E[N ]E[X]
Page 9
SUStech Probability and Some Calculations Applied Stochastic Process
and [ ] [ [ ]]
E Y 2 = E N (N − 1)E 2 [X] + N E X 2
[ ]
= E[N ] Var(X) + E N 2 E 2 [X]
Hence,
[ ]
Var(Y ) = E Y 2 − E 2 [Y ]
= E[N ] Var(X) + E 2 [X] Var(N )
3. A miner is trapped in a mine containing three doors. The first door leads to a tunnel that
takes him to safety after two hours of travel. The second door leads to a tunnel that returns him to
the mine after three hours of travel. The third door leads to a tunnel that returns him to his mine
after five hours. Assuming that the miner is at all times equally likely to choose any one of the doors,
let us compute the moment generating function of X, the time when the miner reaches safety. Let Y
denote the door initially chosen. Then
[ ] 1 ( [ tX ] [ ] [ ])
E etX = E e | Y = 1 + E etX | Y = 2 + E etX | Y = 3 . (2)
3
Now given that Y = 1, it follows that X = 2, and so
[ ]
E etX | Y = 1 = e2t .
Also, given that Y = 2, it follows that X = 3 + X ′ , where X ′ is the number of additional hours
to safety after returning to the mine. But once the miner returns to his cell the problem is exactly as
before, and thus X ′ has the same distribution as X. Therefore,
[ ] [ ]
E etX | Y = 2 = E et(3+X)
[ ]
= e3t E etX .
Similarly,
[ ] [ ]
E etX | Y = 3 = e5t E etX .
Page 10
SUStech Probability and Some Calculations Applied Stochastic Process
4. The Matching Problem At a party n people put their hats in the center of a room where the
hats are mixed together. Each person then randomly selects one. We are interested in the mean and
variance of X-the number that select their own hat. To solve, we use the representation
X = X1 + X2 + · · · + Xn ,
where
1 if the i th person selects his or her own hat
Xi =
0 otherwise
Now, as the i th person is equally likely to select any of the n hats, it follows that P {Xi = 1} = 1/n,
and so
E [Xi ] = 1/n,
( )
1 1 n−1
Var (Xi ) = 1− =
n n n2
Also
Cov (Xi , Xj ) = E [Xi Xj ] − E [Xi ] E [Xj ] .
Now,
1 if the i th and j th person both select their own hats
Xi Xj =
0 otherwise,
and thus
E [Xi Xj ] = P {Xi = 1, Xj = 1}
= P {Xi = 1} P {Xj = 1 | Xi = 1}
1 1
=
nn−1
Hence,
( )2
1 1 1
Cov (Xi , Xj ) = − = 2
n(n − 1) n n (n − 1)
Therefore,
E[X] = 1
and ( )
n−1 n 1
Var(X) = +2
n 2 n2 (n − 1)
= 1.
Page 11
SUStech Probability and Some Calculations Applied Stochastic Process
Thus both the mean and variance of the number of matches are equal to 1. (See (6) for an
explanation as to why these results are not surprising )
5. Matching problem restated
Suppose in the matching problem (4), that those choosing their own hats depart, while the others
(those without a match) put their selected hats in the center of the room, mix them up, and then
reselect. If this process continues until each individual has his or her own hat, find E [Rn ] where Rn is
the number of rounds that are necessary.
We will now show that E [Rn ] = n. The proof will be by induction on n, the number of individuals.
As it is obvious for n = 1, assume that E [Rk ] = k for k = 1, ., n − 1. To compute E [Rn ], start by
conditioning on M , the number of matches that occur in the first round. This gives
∑
n
E [Rn ] = E [Rn | M = i] P {M = i}.
i=0
Now, given a total of i matches in the initial round, the number of rounds needed will equal 1
plus the number of rounds that are required when n − i people remain to be matched with their hats.
Therefore,
∑
n
E [Rn ] = (1 + E [Rn−i ]) P {M = i}
i=0
∑
n
= 1 + E [Rn ] P {M = 0} + E [Rn−i ] P {M = i}
i=1
∑n
= 1 + E [Rn ] P {M = 0} + (n − i)P {M = i}
i=1
Pn = P (E) = P (E | M )P (M ) + P (E | M c ) P (M c )
Page 12
SUStech Probability and Some Calculations Applied Stochastic Process
Clearly, P (E | M ) = 0, and so
n−1
Pn = P (E | M c ) .
n
Now, P (E | M c ) is the probability of no matches when n − 1 people select from a set of n − 1
hats that does not contain the hat of one of them. This can happen in either of two mutually exclusive
ways Either there are no matches and the extra person does not select the extra hat (this being the
hat of the person that chose first), or there are no matches and the extra person does select the extra
hat. The probability of the first of these events is Pn−1 , which is seen by regarding the extra hat as
”belonging” to the extra person Since the second event has probability [1/(n − 1)]Pn−2 , we have
1
P (E | M c ) = Pn−1 + Pn−2
n−1
and thus, from Equation (1.5.4),
n−1 1
Pn = Pn−1 + Pn−2 ,
n n
or, equivalently,
1
Pn − Pn−1 = − (Pn−1 − Pn−2 )
n
However, clearly
1
P1 = 0, P2 = .
2
Thus, from Equation (1 5.5),
P3 − P2 = − (P2 −P
3
1)
= − 3!1 or P3 = 1
2!
− 1
3!
,
P4 − P3 = − (P3 −P
4
2)
= 1
4!
or P4 = 1
2!
− 1
3!
+ 4!1 ,
Page 13
SUStech Probability and Some Calculations Applied Stochastic Process
∑n
then the number of matches, i=1 Xi , can be regarded as the number of successes in n trals when each
is a success with probability 1/n. Now, whereas the above result is not immediately applicable because
these trials are not independent, it is true that it is a rather weak dependence since, for example,
P {Xi = 1} = 1/n
and
P {Xi = 1 | Xj = 1} = 1/(n − 1), j ̸= i.
Hence we would certainly hope that the Poisson limit would still remain valid under this type of
weak dependence. The results of this example show that it does.
Page 14