Department of Mathematics Indian Institute of Technology Guwahati

Download as pdf or txt
Download as pdf or txt
You are on page 1of 3

DEPARTMENT OF MATHEMATICS

INDIAN INSTITUTE OF TECHNOLOGY GUWAHATI


MA225 Probability Theory and Random Processes July - November 2017

Duration: 60 Minutes Quiz – II : Solutions Maximum Marks: 15


Answer all the questions. Support your conclusions with precise explanations. NS
In the sequel, all the conventions and notations are as used in the class. Answers given without proper
justification will be awarded zero marks.

1. The joint probability mass function of the lifetimes X and Y of two connected components in a
machine can be modeled by
e−2

, x = 0, 1, 2, . . . , y = x, x + 1, x + 2, . . .

f (x, y) = x!(y − x)!
0, otherwise

What is the joint probability mass function of X and Y − X? Are X and Y − X independent? What
is the correlation between X and Y ? 3
Solution: Let V = Y − X. The joint probability mass function of X and V is
( −2
e
fX,V (x, v) = P {X = x, V = v} = P {X = x, Y = v + x} = x!v! , x, v = 0, 1, 2, . . .
0, otherwise 1 mark
P∞ e−2 −2 −1
Now, the marginal distributin of X is fX (x) = v=0 x!v! = ex! e = ex! for x = 0, 1, 2, . . . (zero
e−2 e−1
otherwise) and the marginal distributin of V is fV (v) = ∞
P
x=0 x!v! = v! for v = 0, 1, 2, . . . (zero
otherwise). Since fX,V (x, v) = fX (x)fV (v), the random variables X and V = Y −X are independent.
1 mark
P∞ P∞ e−2
To compute E(XY ), rather than using E(XY ) = x=0 y=x xy x!(y−x)! , it can be computed from
E(XY ) = E[X(Y − X + X)] = E(X)E(Y − X) + E(X 2 ), using the independence of X and Y − X.
Since both X and Y − X are P oisson(1) random variables and hence has mean 1 and variance 1, it
follows immediately that E(XY ) = 1 + 2 = 3. Also, it implies that E(Y ) = E(Y − X) + E(X) = 2
and V ar(Y ) = V ar(Y − X) + V ar(X) = 2. This gives ρX,Y = √12 . 1 mark

2. The continuous random variables X and Y have the joint probability density function
( x 1 2
e− 2 x(1+y ) , x, y > 0
f (x, y) = π
0, otherwise

What is the marginal probability density function of X? Show that the random variables Y X and

X are independent. [Fact: Γ(1/2) = π ] 3
1 R∞ 1 2
Solution: The marginal PDF of X, for x > 0, can be obtained as fX (x) = πx e− 2 x 0 e 2 xy dy.
2 p −(1/2)x R ∞ e−z 1 1 √
Putting xy2 = z, we get fX (x) = x2 e π 0
√ dz = √1 x 2 e− 2 x (since Γ(1/2) =
z 2π
π). The
PDF equals zero for x ≤ 0. 1 mark


∂(x,y)
Now, let U = Y X and V = X. Hence, X = V , Y = √UV . Since ∂(u,v) = √1 , the joint PDF of U
v
u 1 1 √ − 1
v − 1 2
u
and V is fU,V (u, v) = fX,Y (v, √v ) √v = π ve 2 e 2 for u, v > 0 and fU,V (u, v) = 0 otherwise.

1 mark

1
− 1 u2 R ∞ √ 1
Now, the marginal PDF of U is fU (u) = e π2 0 ve− 2 v dv, for u > 0 (and zero otherwise). Putting
1 2√ R ∞ t2 e− 12 t2 1 2√ q
e− 2 u e− 2 u 1 2
2
v = t , we get fU (u) = π 2π 0 √

= π 2π.1 = π2 e− 2 u [using normal density].
√ 1
Similarly, we can obtain the marginal PDF of V as fV (v) = √12π ve− 2 v [using gamma density], for
v > 0 (and zero otherwise).
Since fU,V (u, v) = fU (u)fV (v), for all u, v, U and V are independent. [Note: U is distributed as |Z|
with Z a standard normal distribution, and V is distributed as gamma with parameters 3/2 and 1/2]
1 mark

3. If the moment generating function of a random variable X is given by


1 1 −t 1 3t 3 5t
MX (t) = + e + e + e , t ∈ R,
3 6 8 8

then find E(X) and V ar(X). 2


Solution: We know that MX (t) = E(etX ) (if it exists in some neighbourhood of the origin) and
dn
n
dtn MX (t) t=0 = E(X ).

Thus
d 1 3 15 25 1 mark
E(X) = MX (t) t=0 = − + + =
dx 6 8 8 12
and
d2 1 9 75 32
E(X 2 ) =

2
MX (t) t=0 = + + = .
dx 6 8 8 3
Therefore  2
32 25 911
V (X) = E(X 2 ) − E 2 (X) = − = 1 mark
3 12 144
4. Let X1 , X2 , . . . , X50 be a collection of i.i.d. random variables, each having exponential distribution
with parameter α > 0. Find the distribution of S = X1 + X2 + · · · + X50 . 3
Solution: The MGF of exponential distribution with parameter α(> 0) is given by
Z ∞
α
MX (t) = α etx e−αx dx = , t < α. 1 mark
0 α−t
The MGF of the random variable S is given by
P50
MS (t) = E[etS ] = E[et i=1 Xi
]
50
= [MX1 (t)] , since Xi ’s are IID 1 mark
 50
α
= , t<α
α−t
 50
1
= , t<α
1 − αt

which is the MGF of a gamma RV with parameter (50, α). Therefore, by the uniqueness property of
MGF, S ∼ Gamma(50, α). 1 mark

2

5. Let {Xn } be a sequence of random variables with E(Xn ) = 8 and V ar(Xn ) = 1/ n, for n ∈ N.
Prove or disprove: {Xn } converges in probability to 8. 2
Solution: For {Xn } to converge in probability to a constant c, we must have, for every  > 0,
P {|Xn − c| > } → 0 as n → ∞. 1 mark
By Chebyshev’s inequality, for any  > 0, we have

V ar(X)
P {|Xn − E[Xn ]| > } ≤ .
2
Applying this here, we get
1
P {|Xn − 8| > } ≤ √ → 0 as n → ∞
n2

and hence Xn converges to 8 in probability. 1 mark

6. Let {Yn , n ≥ 0} be a sequence of uncorrelated random variables with common mean 0 and common
variance σ 2 . Define Xn = aYn + (1 − a)Yn−1 , for n ∈ N and 0 ≤ a ≤ 1. Prove or disprove: {Xn } is
a covariance stationary process. 2
Solution: Since Yn ’s have mean 0 and variance σ 2 and uncorrelated, we have

Cov(Xn , Xm ) = E[Yn Ym ] − E[Yn ]E[Ym ] = 0

and this implies that (


σ2, n = m
E[Yn Ym ] =
0, n 6= m

To check whether Xn is covariance stationary, we consider

E[Xn ] = aE[Yn ] + (1 − a)E[Yn−1 ] = 0 1 mark


and

Cov(Xn , Xm ) = E[Xn Xm ]
= a2 E[Yn Ym ] + a(1 − a)(E[Yn Ym−1 ] + E[Yn−1 Ym ]) + (1 − a)2 E[Yn−1 Ym−1 ]

2 2 2 2
 a σ + (1 − a) σ , |n − m| = 0

= a(1 − a)σ 2 , |n − m| = 1

 0, |n − m| > 1

which is a function of n − m only (and not dependent on n or m). Therefore, the process {Xn , n ≥ 0}
is a covariance stationary process. 1 mark

You might also like