FE-606 - Unit 2-Revision of Probability and Statistics
FE-606 - Unit 2-Revision of Probability and Statistics
S = f1, 2, 3, 4, 5, 6g
S = f(H, H ) , (H, T ) , (T , H ) , (T , T )g
S = [0, ∞)
F. Guta (CoBE) FE 606 September, 2023 6 / 167
Any subset A of the sample space S is known as an
event. Some examples of events are:
In the experiment of ‡ipping a coin A = fHg is the
event that a head appears on the ‡ip of the coin
In the experiment of ‡ipping two coin
A = f(H, H ) , (H, T )g is the event that a head appears
on the …rst coin.
i). 0 P (A) 1
ii). P (S ) = 1
iii). For any sequence of events A1 , A2 , ... that are
mutually exclusive, i.e., An \ Am = ? when n 6= m,
then
F. Guta (CoBE) FE 606 September, 2023 10 / 167
∞
S
P An = ∑∞
n=1 P (An )
n =1
We refer to P (A) as the probability of the event A.
Example (2.1)
In the die rolling experiment, if we supposed that all
six numbers were equally likely to appear, then the
probability of getting an even number would equal
1 1 1 1
P (f2, 4, 6g) = P f2g + P f4g + P f6g = + + = .
6 6 6 2
Theorem (2.1)
If A and B are two events in a sample space S, then
P (A [ B ) = P (A) + P (B ) P (A \ B )
Proof.
Using the Venn diagram,
A [ B = (A \ B 0 ) [ (A0 \ B ) [ (A \ B )
F. Guta (CoBE) FE 606 September, 2023 12 / 167
Proof.
P (A [ B ) = P (A \ B 0 ) + P (A0 \ B ) + P (A \ B )
However, A = (A \ B 0 ) [ (A \ B )
P (A) = P (A \ B 0 ) + P (A \ B )
Similarly, B = (A0 \ B ) [ (A \ B )
P (B ) = P (A0 \ B ) + P (A \ B )
Finally,
= P (A) P (A \ B ) + P (B ) P (A \ B ) + P (A \ B )
= P (A) + P (B ) P (A \ B )
Example (2.2)
Suppose that we toss two coins, and suppose that
we assume that each of the four outcomes in the
sample space
S = f(H, H ) , (H, T ) , (T , H ) , (T , T )g
is equally likely and hence has probability 41 .
F. Guta (CoBE) FE 606 September, 2023 14 / 167
Example (2.2 continued. . . )
Let A is the event that the …rst coin falls heads, and
B is the event that the second coin falls heads, i.e.,
A = f(H, H ) , (H, T )g , B = f(H, H ) , (T , H )g
The probability of A [ B is
P (A [ B ) = P (A) + P (B ) P (A \ B )
1 1 1 3
= + =
2 2 4 4
P (A [ B [ C ) = P ((A [ B ) [ C )
= P (A [ B ) + P (C ) P ((A [ B ) \ C )
= P (A ) + P (B ) P (A \ B ) + P (C ) P ((A \ C ) [ (B \ C ))
∑ P (Ai \ Aj \ Ak \ Al ) +
i <j <k <l
+ ( 1)n +1 P (A1 \ A2 \ \ An )
Example (2.3)
Roll a fair dice once and let
A = feven numberg ,
F. Guta (CoBE) FE 606 September, 2023 18 / 167
Example (2.3 continued. . . )
B = f1, 2, 3, 5g.
1 4
P (A \ B ) = , and P (B ) =
6 6
So we obtain that,
1/6 1
P ( Aj B ) = =
4/6 4
P (A) = P ( Aj B ) P (B ) + P ( Aj B 0 ) P (B 0 )
Example (2.4)
A bank is considering extending credit to a new
customer and is interested in the probability that
the client will default on the loan. Based on
historical data, the bank knows that there is a 5%
P (A) = P ( Aj B ) (B ) + P ( Aj B 0 ) P (B 0 )
= (0.05) (0.3) + (0.005) (0.7) ' 0.0185.
Bi \ Bj = ?, for i 6= j
P (A \ B ) = P (A) P (B )
F. Guta (CoBE) FE 606 September, 2023 24 / 167
By Equation (1.5) this implies that A and B are
independent if
P ( Aj B ) = P (A)
Example (2.5)
Consider a sequence of n independent trials, each of
1
which has a probability n of being a “success”.
What is the probability of zero successes in n trials?
What if the number of trials is doubled?
The probability is simply:
F. Guta (CoBE) FE 606 September, 2023 26 / 167
Example (2.5 continued. . . )
n
P (failure on trial 1,. . . ,failure on trial n) = ∏ P (failure on trial i )
i =1
n
1 1
= 1 'e
n
2n
P (failure on trial 1,. . . ,failure on trial n) = ∏ P (failure on trial i )
i =1
2n
1 2
= 1 'e
n
A = (A \ B ) [ (A \ B 0 )
P (A) = P (A \ B ) + P (A \ B 0 )
F. Guta (CoBE) FE 606 September, 2023 28 / 167
P (A) = P ( Aj B ) P (B ) + P ( Aj B 0 ) P (B 0 ) (2.3)
From the preceding two expressions it follows:
P ( Aj B ) P (B )
P ( Bj A) =
P ( Aj B ) P (B ) + P ( Aj B 0 ) P (B 0 )
Example (2.6)
In answering a question on a multiple-choice test a
student either knows the answer or guesses. Let p
be the probability that she knows the answer and
1 p the probability that she guesses. Assume that
a student who guesses at the answer will be correct
F. Guta (CoBE) FE 606 September, 2023 29 / 167
Example (2.6 continued. . . )
with probability 1/m, where m is the number of
multiple-choice alternatives.
P (A \ B )
P ( Bj A) =
P (A)
P ( Aj B ) P (B )
=
P ( Aj B ) P (B ) + P ( Aj B 0 ) P (B 0 )
p
=
p + (1/m) (1 p )
pm
=
1 + p ( m 1)
A = [ni=1 (A \ Bi )
P (A \ B 0 )
P B0 A =
P (A)
P ( Aj B 0 ) P (B 0 )
=
P ( Aj B 0 ) P (B 0 ) + P ( Aj B ) P (B )
0.14 0.99
= ' 0.94.
0.14 0.99 + 0.88 0.01
F. Guta (CoBE) FE 606 September, 2023 35 / 167
2.1.5 Random Variables
De…nition (2.2)
For an arbitrary set A 2 F de…ne IA (ω ) = 1 if
ω 2 A and 0 otherwise. This is called an indicator
random variable.
F. Guta (CoBE) FE 606 September, 2023 38 / 167
Example (2.8)
Suppose that our experiment consists of tossing two
fair coins. Letting X denote the number of heads
appearing, then X is a random variable taking on
one of the values 0, 1, 2 with respective probabilities
1
P fX = 0g = P f(T , T )g =
4
1
P fX = 1g = P f(H, T ) , (T , H )g =
2
1
P fX = 2g = P f(H, H )g =
4
P fX = 1g = P fHg = p
P fX = 2g = P f(T , H )g = (1 p) p
F. Guta (CoBE) FE 606 September, 2023 40 / 167
Example (2.9 continued. . . )
P fX = 3g = P f(T , T , H )g = (1 p )2 p
..
.
P fX = x g = P f(T , T , : : : , T , H )g = (1 p )x 1
p, x 1
∞
P ([∞
x =1 fX = x g) = ∑ P fX = x g
x =1
∞
= ∑ (1 p )x 1
p
x =1
p
= =1
1 (1 p )
F. Guta (CoBE) FE 606 September, 2023 41 / 167
Example (2.10)
Suppose that our experiment consists of seeing how
F (x ) = ∑ pi
fi,Xi xg
p (x ) = P fX = xg
p (xi ) > 0, i = 1, 2, : : :
p (x ) = 0, all other values of x
F. Guta (CoBE) FE 606 September, 2023 48 / 167
Since X must take on one of the values xi , we have
∞
∑ p (xi ) = 1.
i =1
The cumulative distribution function F can be
expressed in terms of p (x ) by
F (x ) = ∑ p (xi )
fi :xi xg
Example (2.11)
It is known that any item produced by a certain
machine will be defective with probability 0.1,
independently of any other item. What is the
probability that in a sample of three items, at most
one will be defective?
3 3
= 0.10 (1 0.1)3 0
+ 0.11 (1 0.1)3 1
' 0.972.
0 1
p (x ) = (1 p )x 1
p, x = 1, 2, : : : (2.8)
n!
P fX = x g = p x (1 p )n x
x ! (n x ) !
x n x
n! λ λ
= 1
x ! (n x ) ! n n
n (n 1) (n x + 1) λ x (1 λ /n )n
=
nx x ! (1 λ /n )x
Example (2.12)
Consider an experiment that consists of counting
the number of α-particles given o¤ in a one-second
interval by one gram of radioactive material.
F. Guta (CoBE) FE 606 September, 2023 62 / 167
Example (2.12 continued. . . )
If we know from past experience that, on the
average, 3.2 such α-particles are given o¤, what is a
good approximation to the probability that no more
than two α-particles will appear?
The number of α-particles given o¤ will be a
Poisson r.v with parameter λ = 3.2. Hence the
desired probability is
Example (2.13)
Calculate the cdf of a r.v uniformly distributed over
Rx
(α, β ). Since F (x ) = ∞f (u ) du, we obtain
8
>
>
>
> 0, x α
>
>
>
<
F (x ) = x α
>
> β α, α <X <β
>
>
>
>
>
: 1, X β
F. Guta (CoBE) FE 606 September, 2023 68 / 167
2.3.2 Exponential Distribution
FY (y ) = P fY yg = P fαX + β yg
y β y β
= P X = FX
α α
F. Guta (CoBE) FE 606 September, 2023 72 / 167
Z (y β )/α n o
p1 1 x µ 2
FY ( y ) = 2π
exp 2 σ dx
∞ σ
( )
2
d 1 1 y (α µ + β )
FY ( y ) = p exp , ∞<y <∞
dy ασ 2π 2 ασ
E (X ) = ∑ xp (x )
fx :f (x )>0g
1 1 1 1 1 1 7
E (X ) = 1 +2 +3 +4 +5 +6 = .
6 6 6 6 6 6 2
Example (2.15)
Find E (X ) when X is binomially distributed with
parameters n and p.
n
( n 1) !
= np ∑ (x 1) ! ( n x ) !
p x 1 (1 p )n x
x =1
n
n 1 x 1
= np ∑ x 1
p (1 p )n x
x =1
n 1
n 1
= np ∑ k
p k (1 p ) (n 1) k
k =0
= np [p + (1 p )]n 1
= np
F. Guta (CoBE) FE 606 September, 2023 76 / 167
Example (2.16)
What is the expected value of a geometric r.v with
parameter p.
∞ ∞
E (X ) = ∑ xp (1 p) x 1
=p ∑ x (1 p )x 1
x =1 x =1
∞ ∞
= p ∑ x (1 p )x 1
=p ∑ xqx 1
x =1 x =1
where q = 1 p.
!
∞ ∞
d d
E (X ) = p ∑ dq (qx ) = p dq ∑ qx
x =1 x =1
F. Guta (CoBE) FE 606 September, 2023 77 / 167
Example (2.16 continued. . . )
d q 1
E (X ) = p dq =p = p1 .
1 q (1 q )2
Example (2.17)
Find E (X ) if X is a Poisson r.v with parameter λ .
∞ ∞ ∞
λx λx λx
E (X ) = ∑ xe λ
x!
= ∑ xe λ
x!
=e λ
∑
x =0 x =1 x =1 (x 1) !
∞ ∞
λx 1 λk
= λe λ
∑ = λe λ
∑ = λe λ λ
e =λ
x =1 (x 1) ! k =0 k !
Example (2.18)
Let X be exponentially distributed with parameter
λ . Find E (X ).
Z ∞
λx
E (X ) = xλ e dx
0
Example (2.19)
Let X be a gamma r.v with parameters α, and λ .
Find E (X ).
Z ∞ 1 Z ∞
λ e λ x (λ x )α Γ (α + 1) λ e λ x (λ x )α
E (X ) = x dx = dx
0 Γ (α ) λ Γ (α ) 0 Γ (α + 1)
αΓ (α ) α
= = .
λ Γ (α ) λ
Writing x as (x µ ) + µ yields
Z ∞ Z ∞
1 1
( x σ µ ) dx + µ p1
2 1 2
( x σ µ ) dx
E (X ) = p (x µ)e 2 e 2
σ 2π ∞ σ 2π ∞
Letting y = x µ leads to
Corollary (2.1)
If a and b are constants, then
E (aX + b) = aE (X ) + b
Example (2.21)
Let X be normally distributed with parameters µ
and σ 2 . Find var (X ).
Recalling that E (X ) = µ, we have that
h i
2
var (X ) = E (X µ)
Z ∞
1 (x µ )2 /2σ 2
= p (x µ )2 e dx
σ 2π ∞
Substituting y = (x µ ) /σ yields
F. Guta (CoBE) FE 606 September, 2023 86 / 167
Example (2.21 continued. . . )
2 R∞
var (X ) = pσ y 2 e y 2 /2 dy
2π ∞
Example (2.22)
Calculate var (X ) where X represents the outcome
when a fair die is rolled.
F. Guta (CoBE) FE 606 September, 2023 87 / 167
Example (2.22 continued. . . )
As previously noted in Example 2.15, E (X ) = 27 .
Also,
1 1 1 1 1 1 91
E X 2 = 12 + 22 + 32 + 42 + 52 + 62 = .
6 6 6 6 6 6 6
Hence,
2
2 291 7 35
var (X ) = E X E (X ) = =
6 2 12
FY (y ) = P fY yg = F f∞, yg
p (x, y ) = P fX = x, Y = yg
F. Guta (CoBE) FE 606 September, 2023 90 / 167
The probability mass function of X may be obtained
from p (x, y ) by
Similarly,
P fX 2 Ag = P fX 2 A, Y 2 ( ∞, ∞)g
Z ∞Z Z
= f (x, y ) dxdy = fX (x ) dx
∞ A A
F. Guta (CoBE) FE 606 September, 2023 92 / 167
Z ∞
where fX (x ) = f (x, y ) dy.
∞
Similarly, the probability density function of Y is
given by
Z ∞
fY (y ) = f (x, y ) dx
∞
Since
Z x Z y
F (x, y ) = P fX x, Y yg = f (u, v ) dvdu
∞ ∞
di¤erentiation yields
∂2
F (x, y ) = f (x, y )
∂ x∂ y
F. Guta (CoBE) FE 606 September, 2023 93 / 167
A variation of Proposition 2.1 states that if X and
where
F. Guta (CoBE) FE 606 September, 2023 95 / 167
Example (2.23 continued. . . )
8
< 1, if the i th man selects his own hat
Xi =
: 0, otherwise
and so,
1
E (X ) = E (X1 ) + E (X2 ) + + E (XN ) = N =1
N
P fX x, Y yg = P fX xg P fY yg (2.12)
p (x, y ) = pX (x ) pY (y )
f (x, y ) = fX (x ) fY (y )
E [h (x ) g (y )] = E [h (x )] E [g (y )]
F. Guta (CoBE) FE 606 September, 2023 98 / 167
Proof.
Suppose that X and Y are jointly continuous. Then
Z ∞Z ∞
E [h (x ) g (y )] = h (x ) g (y ) f (x, y ) dxdy
∞ ∞
Z ∞Z ∞
= h (x ) g (y ) fX (x ) fY (y ) dxdy
∞ ∞
Z ∞ Z ∞
= h (x ) fX (x ) dx g (y ) fY (y ) dy
∞ ∞
= E [h (x )] E [g (y )]
= E [XY XE (Y ) YE (X ) + E (X ) E (Y )]
= E (XY ) E (X ) E (Y ) E (X ) E (Y ) + E (X ) E (Y )
= E (XY ) E (X ) E (Y )
1
f (x, y ) = e (y +x /y )
, 0 < x, y < ∞
y
Solution
a). To show that f (x, y ) is a joint density function we
F. Guta (CoBE) FE 606 September, 2023 101 / 167
Solution
need to show it is nonnegative, which is immediate,
Z ∞Z ∞
and that f (x, y ) dxdy = 1.
∞ ∞
Z ∞Z ∞ Z ∞Z ∞
1
f (x, y ) dxdy = e (y +x /y ) dxdy
∞ ∞ 0 0 y
Z ∞ Z ∞
y 1
= e e x /y dx dy
0 0 y
Z ∞
= e y dy = 1
0
Consequently,
cov (X , Y ) = E (XY ) E (X ) E (Y ) = 2 1 ( 1) = 1
Properties of Covariance
cov (X , Y + Z ) = E [X (Y + Z )] E (X ) E (Y + Z )
F. Guta (CoBE) FE 606 September, 2023 107 / 167
cov (X , Y + Z ) = E [XY + XZ ] E (X ) [E (Y ) + E (Z )]
= E (XY ) E (X ) E (Y ) + E (XZ ) E (X ) E (Z )
= cov (X , Y ) + cov (X , Z )
i). E X = µ.
ii). var X = σ 2 /n.
F. Guta (CoBE) FE 606 September, 2023 110 / 167
iii). cov X , Xi X = 0.
Proof.
Parts (i ) and (ii ) are easily established as follows:
1 n
n i∑
E X = E (Xi ) = µ
=1
!
2 n 2 n
1 1
var X =
n
var ∑ Xi =
n ∑ var (Xi ) =
i =1 i =1
!
n
1 1 σ2
=
n
cov (Xi , Xi ) + cov
n
∑ Xj , Xi n
j 6=i
σ2 σ2
= =0
n n
Example (2.25)
Compute the variance of a binomial r.v X with
parameters n and p.
F. Guta (CoBE) FE 606 September, 2023 112 / 167
Solution
Since a binomial r.v represents the number of
successes in n independent trials when each trial has
a common probability p of being a success, we may
write
X = X1 + X2 + + Xn
But
and thus
var (X ) = np (1 p)
F. Guta (CoBE) FE 606 September, 2023 114 / 167
It is often important to be able to calculate the
distribution of X + Y from the distributions of X
and Y when X and Y are independent.
Suppose that X and Y are continuous, X having
pdf f and Y having pdf g.
Example (2.26)
Let X and Y be independent Poisson r.v with
0 k n, we have
n
P fX + Y = n g = ∑ P fX = k, Y = n kg
k =0
F. Guta (CoBE) FE 606 September, 2023 117 / 167
Solution
P fX + Y = ng = ∑nk =0 P fX = k g P fY = n kg
k
n
λ n2 k
= ∑e k!
λ1 λ 1
e λ2
(n k )!
k =0
e (λ 1 +λ 2 ) n n!
=
n!
∑ k ! (n k )!
λ k1 λ n2 k
k =0
n
(λ 1 +λ 2 ) ( λ 1 + λ 2 )
= e
n!
Example (2.27)
If X and Y are independent gamma r.v with
parameters (α, λ ) and (β , λ ), respectively, compute
the joint density of U = X + Y & V = X / (X + Y ).
F. Guta (CoBE) FE 606 September, 2023 121 / 167
Solution
The joint density of X and Y is given by
fX ,Y (x, y ) = fX (x ) fY (y )
λe λ x (λ x )α 1 λ e λ y (λ y )β 1
=
Γ (α ) Γ (β )
λ α +β λ (x +y ) α 1 β 1
= e x y
Γ (α ) Γ (β )
∂ h1 ∂ h1 ∂ h2 ∂ h2
= v, = u, =1 v, = u
∂u ∂v ∂u ∂v
F. Guta (CoBE) FE 606 September, 2023 122 / 167
Solution
Finally, we see that
fU ,V (u, v ) = fX ,Y (uv, u (1 v )) u
λ α +β λu 1 1
= e (uv )α (u (1 v ))β u
Γ (α ) Γ (β )
λ α +β λ u α +β 1 α 1 1
= e u v (1 v )β
Γ (α ) Γ (β )
1
λe λu
(λ u )α +β Γ (α + β ) α 1 1
= v (1 v )β
Γ (α + β ) Γ (α ) Γ (β )
Γ (α + β ) α 1 1
fV (v ) = v (1 v )β , 0<v <1
Γ (α ) Γ (β )
d d tX
MX0 (t ) = E e tX = E e = E Xe tX
dt dt
F. Guta (CoBE) FE 606 September, 2023 125 / 167
Hence,
MX0 (0) = E (X )
Similarly,
d
MX00 (t ) = E Xe tX = E X 2 e tX
dt
and so
MX00 (0) = E X 2
and p :
n
n x
MX (t ) = E e tX = ∑ e tx x
p (1 p )n x
x =0
n
n
∑
x n
= pe t (1 p )n x
= pe t + 1 p
x =0 x
F. Guta (CoBE) FE 606 September, 2023 127 / 167
Example (2.28 continued. . . )
n 1
Hence, MX0 (t ) = n (pe t + 1 p) pe t
and so MX0 (0) = E (X ) = np
n 2 2 n 1
MX00 (t ) = n (n 1) pe t + 1 p pe t + n pe t + 1 p pe t
and so
E X 2 = MX00 (0) = n (n 1) p2 + np
var (X ) = E X 2 E (X )2
= n (n 1) p2 + np n2 p 2
= np (1 p)
Example (2.29)
MGF of the Poisson distribution with mean λ
∞ λλx
tx e
MX (t ) = E e tX
= ∑e x!
x =0
F. Guta (CoBE) FE 606 September, 2023 129 / 167
Example (2.29 continued. . . )
x
(λ e t )
MX (t ) = e λ
∑∞
x =0 x! =e λ eλ et = exp λ (e t 1)
Di¤erentiation yields
MX0 (t ) = λ e t exp λ e t 1
E X 2 = MX00 (0) = λ 2 + λ
F. Guta (CoBE) FE 606 September, 2023 130 / 167
Example (2.29 continued. . . )
var (X ) = E X 2 E (X )2 = λ
Example (2.30)
MGF of the Gamma Distribution with Parameters
α, and λ
Z ∞ λx 1
tX tx λ e (λ x )α
MX (t ) = E e = e dx
0 Γ (α )
Di¤erentiation of MX (t ) yields
αλ α α ( α + 1) λ α
MX0 (t ) = α +1
, MX0 (t ) =
(λ t) (λ t )α +2
Hence,
α α ( α + 1)
E (X ) = MX0 (0) = , E X 2 = MX00 (0) =
λ λ2
F. Guta (CoBE) FE 606 September, 2023 132 / 167
Example (2.30 continued. . . )
The variance of X is thus given by
var (X ) = E X 2 E (X )2
α ( α + 1) α 2 α
= =
λ2 λ λ2
Example (2.31)
MGF of the Normal Distribution with Parameters µ
and σ 2
The MGF of a standard normal r.v Z is obtained as
follows.
F. Guta (CoBE) FE 606 September, 2023 133 / 167
Example (2.31 continued. . . )
R∞ tz p1 e z 2 /2 dz
MZ (t ) = E e tZ = ∞e 2π
Z ∞
1 2
= p e (z 2tz )/2 dz
∞ 2π
Z ∞
t 2 /2 1 2 2
= e p e (z t ) /2 dz = e t /2
∞ 2π
By di¤erentiating we obtain
1
MX0 (t ) = µ + t σ 2 exp t µ + t 2 σ 2
2
2 1 1
MX00 (t ) = µ + tσ 2 exp t µ + t 2 σ 2 + σ 2 exp t µ + t 2 σ 2
2 2
E X 2 = MX00 (0) = µ 2 + σ 2
var (X ) = E X 2 E (X )2 = σ 2
Solution
MGF of X + Y is found as follows:
h i
MX +Y (t ) = E e t (X +Y ) = E e tX e tY = MX (t ) MY (t )
1 1
= exp t µ 1 + t 2 σ 21 exp t µ 2 + t 2 σ 22
2 2
1
= exp t ( µ 1 + µ 2 ) + t 2 σ 21 + σ 22
2
n
∑ (Xi
2
= µ )2 n X µ
i =1
E S2 = σ 2
F. Guta (CoBE) FE 606 September, 2023 140 / 167
We will now determine the joint distribution of the
sample mean X = ∑ni=1 Xi /n and the sample
variance S 2 when the Xi have a normal distribution.
To begin we need the concept of a chi-squared
random variable.
De…nition (2.4)
If Z1 , ..., Zn are independent standard normal r.v,
then the random variable ∑ni=1 Zi2 is said to be a
chi-squared random variable with n degrees of
freedom.
F. Guta (CoBE) FE 606 September, 2023 141 / 167
We shall now compute the moment generating
Z ∞
2 2 1 2
E e tZi = e tzi p e zi /2 dzi
∞ 2π
Z ∞
1 2
= p e (1 2t )zi /2 dzi
∞ 2π
Z ∞
1 2 2 1
= σ p e zi /2σ dzi , where σ 2 = (1 2t )
∞ σ 2π
1 /2
= σ = (1 2t )
Hence,
F. Guta (CoBE) FE 606 September, 2023 142 / 167
n n
n /2
E exp t ∑ Zi2 = ∏ E exp tZi2 = (1 2t )
i =1 i=
Proof.
We give a proof for the case where X is continuous
with density f .
Z ∞ Z a Z ∞
E (X ) = xf (x ) dx = xf (x ) dx + xf (x ) dx
0 0 a
F. Guta (CoBE) FE 606 September, 2023 148 / 167
Proof. Z a Z ∞
E (X ) xf (x ) dx a f (x ) dx = aP fX ag
0 a
E (X )
E (X ) aP fX ag P fX ag
a
Example (2.33)
Suppose we know that the number of items
produced in a factory during a week is a r.v with
F. Guta (CoBE) FE 606 September, 2023 150 / 167
Example (2.33 continued. . . )
mean 500.
Solution
Let X be the number of items that will be produced
in a week.
F. Guta (CoBE) FE 606 September, 2023 151 / 167
Solution
a). By Markov’s inequality,
E (X ) 500 1
P fX 1000g = = .
1000 1000 2
σ2 1
P fjX 500j 100g =
1002 100
Hence
1 99
P fjX 500j 100g 1 =
100 100
F. Guta (CoBE) FE 606 September, 2023 152 / 167
Solution
and so the probability that this week’s production
will be between 400 and 600 is at least 0.99.
Theorem (2.1: Strong Law of Large Numbers)
Let X1 , X2 , ... be a sequence of independent r.v
having a common distribution, and let E (Xi ) = µ.
Then, with probability 1, X ! µ, n ! ∞.
Example (2.34)
Suppose that a sequence of independent trials is
performed. Let E be a …xed event and denote by
F. Guta (CoBE) FE 606 September, 2023 153 / 167
Example (2.34 continued. . . )
P (E ) the probability that E occurs on any
particular trial. Let
8
< 1, if E occurs on the i th trial
Xi =
: 0, if E does not occurs on the i th trial
X1 + X2 + + Xn
! E (X ) = P (E )
n
F. Guta (CoBE) FE 606 September, 2023 154 / 167
Theorem (2.2: Central Limit Theorem)
Let X1 , X2 , ... be a sequence of independent,
identically distributed r.v, each with mean µ and
variance σ 2 . Then the distribution of
X µ d
p ! N (0, 1) , as n ! ∞.
σ/ n
Solution
Since the binomial is a discrete r.v, and the normal
p X jY ( xj y ) = P fX = xj Y = yg
P fX = x, Y = yg p (x, y )
= =
P fY = yg pY (y )
F. Guta (CoBE) FE 606 September, 2023 158 / 167
for all values of y such that P fY = yg > 0.
Similarly, the conditional probability distribution
function of X given that Y = y is de…ned, for all y
such that P fY = yg > 0, by
F X jY ( xj y ) = P fX xj Y = yg = ∑ p X jY ( aj y )
a x
Finally, the conditional expectation of X given that
Y = y is de…ned by
E [ X j Y = y ] = ∑ xP f X = x j Y = y g = ∑ xp X jY ( x j y )
x x
p X jY ( xj y ) = P fX = xj Y = yg = P fX = xg
Example (2.36)
If X and Y are independent Poisson random
variables with respective means λ 1 and λ 2 ,
calculate the conditional expected value of X given
F. Guta (CoBE) FE 606 September, 2023 160 / 167
Example (2.36 continued. . . )
that X + Y = n.
Solution
Let us …rst calculate the conditional probability mass
P fX = x, X + Y = ng
P f X = x j X + Y = ng =
P fX + Y = n g
P fX = x, Y = n x g
=
P fX + Y = n g
P fX = x g P fY = n x g
=
P fX + Y = ng
λ1
E fX j X + Y = ng = n .
λ1 +λ2
Example (2.37)
Suppose the joint density of X and Y is given by
8
>
>
>
< 4y (x y )e (x +y ) , 0 < x < ∞, 0 y x
f (x, y ) =
>
>
>
: 0, otherwise
Compute E [ X j Y = y ].
F. Guta (CoBE) FE 606 September, 2023 164 / 167
Solution
The conditional density of X , given that Y = y, is
given by
f (x, y )
f X jY ( x j y ) =
fY (y )
4y (x y )e (x +y )
= Z ∞
4y (x y )e (x +y ) dx
y
(x y )e x
= Z ∞
(x y )e x dx
y
(x y )
= (x y )e , x >y
Z ∞
(x y )
E [XjY = y] = x (x y )e dx
y
Z ∞
= (s + y ) se s ds
Z0∞ Z ∞
2 s s
= s e ds + yse ds
0 0
= 2+y
F. Guta (CoBE) FE 606 September, 2023 166 / 167
Note 1: Law of iterated expectations:
E (X ) = E [E ( X j Y )]