MATH2901 Part 1 Term 2 2020
MATH2901 Part 1 Term 2 2020
Jeffrey Yang
August 2020
Most of the examples were taken from the slides Rui Tong (2018 StatSoc
Team) made for that year’s revision seminar.
All examples are presented at the end so that it isn’t as obvious which
techniques/methods should be used.
It is recommended that you refer to the official lecture notes when quoting
definitions/results.
4 Inequalities
From the axioms, we are able to derive the following fundamental results:
1 If A1 , A2 , ..., Ak are mutually exclusive,
k k
!
[ X
P Ai = P(Ai ).
i=1 i=1
2 P(∅) = 0.
3 For any A ⊆ Ω, 0 ≤ P(A) ≤ 1 and P(Ā) = 1 − P(A).
4 If B ⊂ A, then P(B) ≤ P(A). Hence, if B occurs → A occurs, then
P(B) ≤ P(A).
These results can be used without proof.
Given that B has occurred, the total probability for the possible results of
an experiment equals P(B). Visually, we see that the only outcomes in A
that are now possible are those in A ∩ B.
Jeffrey Yang (UNSW Society of Statistics/UNSW Mathematics
MATH2901Society)
Revision Seminar August 2020 8 / 116
Independence
Recall that for any two events A and B, we have P(A ∩ B) = P(A|B)P(B).
Thus, A and B are independent if and and only P(A|B) = P(A) (or
equivalently, P(B|A) = P(B)). Intuitively, this means that knowing event A
has occurred does not give us any information on the probability of event B
occurring (and vice versa).
Independence implies pairwise independence but not vice versa. Can you
think of an example of where pairwise independence does not imply
independence?
Jeffrey Yang (UNSW Society of Statistics/UNSW Mathematics
MATH2901Society)
Revision Seminar August 2020 10 / 116
The Multiplicative Law
P(A1 ∩ A2 ∩ A3 ) = P(A3 ∩ A2 ∩ A1 )
= P(A3 |A2 ∩ A1 )P(A2 ∩ A1 )
= P(A3 |A1 ∩ A2 )P(A2 |A1 )P(A1 ).
You might have noticed that the Additive Law resembles the
Inclusion-exclusion principle. The following quote from Wikipedia sheds
some light on this:
4 Inequalities
The following two properties are important and apply to all discrete random
variables:
1 fX (x) ≥ 0 for all x ∈ R
P
all x fX (x) = 1
2
The following two properties are important and apply to all continuous
random variables:
1 f (x) ≥ 0 for all x ∈ R.
X
R∞
−∞ fX (x)dx = 1.
2
def X X
EX = E[X ] = x · P(X = x) = x · fX (x),
all x all x
Definition (Variance)
Let µ = E[X ]. Then the variance of X , denoted Var[X ], is defined as
Var[X ] = E[X 2 ] − µ2 .
Var[X + a] = Var[X ]
Var[aX ] = a2 Var[X ].
µX (u) = E[e uX ].
The following result regarding the r th moment of X , E[X r ], shows why the
moment generating function is called as such.
Result (r th Moment of a random variable)
Let X be a random variable. Then in general, we have
(r )
E[X r ] = mX (0)
for r = 0, 1, 2, ....
Jeffrey Yang (UNSW Society of Statistics/UNSW Mathematics
MATH2901Society)
Revision Seminar August 2020 26 / 116
Properties of Moment Generating Functions
Result (Uniqueness)
Let X and Y be two random variables all of whose moments exist. If
mX (u) = mY (u)
for all u in a neighbourhood of 0 (i.e., for all |µ| < for some < 0), then
Result (Convergence)
Let {Xn : n = 1, 2, ...} be a sequence of random variables, each with
moment generating function mXn (u). Furthermore, suppose that
4 Inequalities
Here, ”∼” has the meaning ”is distributed as” or ”has distribution”.
Results
If X ∼ Bin(n, p) then
n x n−x , x = 0, ..., n,
x p (1 − p)
1 f (x) =
X
2 E[X ] = np,
3 Var(X ) = np(1 − p).
Jeffrey Yang (UNSW Society of Statistics/UNSW Mathematics
MATH2901Society)
Revision Seminar August 2020 33 / 116
Geometric Distribution
Results
If X ∼ Geom(p), then
1 fX (x; p) = p(1 − p)x−1 , x = 1, 2, ...
2 E[X ] = p1 ,
1−p
3 Var(X ) = p2
.
e −λ λx
fX (x; λ) = P(X = x) = , x = 0, 1, 2, ...
x!
A common abbreviation is X ∼ Poisson(λ).
Definition
A random variable X is said to have an exponential distribution with
parameter β > 0 if X has density function:
1 −x/β
fX (x; β) = e , x > 0.
β
Results
1 E(X ) = β,
2 Var(X ) = β 2 ,
3 Memoryless property: P(X > s + t|X > s) = P(X > t).
Jeffrey Yang (UNSW Society of Statistics/UNSW Mathematics
MATH2901Society)
Revision Seminar August 2020 36 / 116
Uniform Distribution
Results
If X ∼ Unif(a, b), then
1
1 fX (x; a, b) = b−a , a < x < b,
a+b
2 E(X ) = 2 ,
2
3 Var(X ) = (b−a)
12 .
The Gamma Function extends the factorial function to the real numbers.
Definition (Gamma Function)
The Gamma function at x ∈ R is given by
Z ∞
Γ(x) = t x−1 e −t dt.
0
Results
1 Γ(x) = (x − 1)Γ(x − 1),
2 Γ(n) = (n − 1)!,, n = 1, 2, 3...
√
3 Γ( 21 ) = π,
R ∞ m −x
0 x e dx = m! for m = 0, 1, 2, ....
4
Result
For all x, y ∈ R,
Γ(x)Γ(y )
B(x, y ) = .
Γ(x + y )
Definition (Φ)
For all x ∈ R, Z x
1 2 /2
Φ(x) = √ e −t dt.
2π −∞
Results
If X ∼ N(µ, σ 2 ),
1 E(X ) = µ,
2 Var(X ) = σ 2 ,
Result
If Z ∼ N(0, 1) then
Result
If X ∼ N(µ, σ 2 ) then
X −µ
Z= ∼ N(0, 1).
σ
e −α/β x α−1
fX (x; α, β) = ,x > 0
Γ(α)β α
X ∼ Gamma(α, β).
Results
If X ∼ Gamma(α, β) then
1 E(X ) = αβ,
2 Var(X ) = αβ 2 .
Moreover, Y has an Exponential distribution iff Y ∼ Gamma(1, β).
Jeffrey Yang (UNSW Society of Statistics/UNSW Mathematics
MATH2901Society)
Revision Seminar August 2020 43 / 116
Beta Distribution
x α−1 (1 − x)β−1
fX (x; α, β) = , 0 < x < 1.
B(α, β)
fX ,Y (x, y ) = P(X = x, Y = y ),
FX ,Y (x, y ) = P(X ≤ x, Y ≤ y )
(P P
P(X = u, Y = v ) (X discrete)
= R y R x v ≤y
u≤x
Result
1 If
PX and P Y are discrete radom variables, then
all x all y fX ,Y (x, y ) = 1.
2 If
R∞ X and
R ∞ Y are continuous random variables then
−∞ −∞ fX ,Y (x, y )dxdy = 1.
P(X = x, Y = y ) fX ,Y (x, y )
fX |Y (x|y ) = P(X = x|Y = y ) = = .
P(Y = y ) fY (y )
Similarly,
fX ,Y (x, y )
fY |X (y |x) = P(Y = y |X = x) = .
fX (x)
Result
X
P(Y ∈ A|X = x) = fY |X (y |X = x).
y ∈A
Shorthand: fY |X (y |x).
Result
Z b
P(a ≤ Y ≤ b|X = x) = fY |X (y |x)dy .
a
where (P
2 x 2 P(X = x|Y = y )
E(X |Y = y ) = R ∞all x 2
−∞ x fX |Y (x|y )dx.
Definition (Independent)
Random variables X and Y are independent if and only if for all x, y
fX ,Y (x, y ) = fX (x)fY (y ).
Result
Random variables X and Y are independent if and only if for all x, y
fY |X (x, y ) = fY (y )
Result
If X and Y are independent
FX ,Y (x, y ) = FX (x) · FY (y ).
Jeffrey Yang (UNSW Society of Statistics/UNSW Mathematics
MATH2901Society)
Revision Seminar August 2020 53 / 116
Covariance
Definition (Covariance
The covariance of X and Y is
Cov(X , Y ) measures how much X and Y vary about their means and also
how much they vary together linearly.
Results
1 Cov(X , X ) = Var(X )
2 Cov(X , Y ) = E(XY ) − µX µY
3 If X and Y are independent, then Cov(X , Y ) = 0.
Definition (Correlation)
The correlation between X and Y is
Cov(X , Y )
Corr(X , Y ) = p .
Var(X ) · Var(Y )
Results
1 | Corr(X , Y )| ≤ 1
2 | Corr(X , Y )| = 1 if and only if P(Y = a + bX ) = 1 for some constants
a, b.
Result
For discrete X ,
X
fY (y ) = P(Y = y ) = P[h(X ) = y ] = P(X = x).
x:h(x)=y
Result
For continuous X , if h is monotonic over the set {x : fX (x) > 0} then
dx
fY (y ) = fX (x)
dy
dx
= fX (h−1 (y ))
dy
Result
For a continuous random variable X , if Y = aX + b is a linear
transformation of X with a 6= 0, then
1 y −b
fY (y ) = fX
|a| a
Result
If U and V are functions of continuous random variables X and Y , then
where
∂x ∂x
∂u ∂v
J= ∂y ∂y
∂u ∂v
Result
If X1 , X2 , ..., Xn are independent with Xi ∼ Gamma(αi , β), then
n
X Xn
Xi ∼ Gamma( αi , β).
i=1 i=1
Result
Suppose that X and Y are independent random variables with moment
generating functions mX and mY . Then
Result
If X ∼ N(µX , σX2 ) and Y ∼ N(σY , σY2 ) are independent then
4 Inequalities
1 Jensen’s Inequality
2 Markov’s inequality
3 Chebyshev’s inequality
E[X ]
P(X ≥ a) ≤ .
a
That is, the probability that X is at least a is at most the expectation of X
divided by a.
1
P(|X − µ| > kσ) ≤ .
k2
That is, the probability that X is more than k standard deviations from its
mean is less than k12 .
4 Inequalities
d
for all x where FX is continuous. A common shorthand is Xn → X .
We say that FX is the limiting distribution of Xn .
The Weak Law of Large Numbers describes how the sample average
converges to the distributional average as the sample size increases.
X̄n − µ d
√ →Z
σ/ n
X̄n − µ d
√ → N(0, 1).
σ/ n
2
Note that E(X̄n = µ and Var(X̄n ) = σn so that Central Limit Theorem
states that the limiting distribution of any standardised average of
independent random variables is the standard Normal distribution.
Jeffrey Yang (UNSW Society of Statistics/UNSW Mathematics
MATH2901Society)
Revision Seminar August 2020 81 / 116
Alternative Forms of the Central Limit Theorem
Sometimes,
Pn probabilities involving
Pn related quantities such as the sum
i=1 Xi are required. Since i=1 Xi = nX̄ , the Central Limit Theorem
also applies to the sum of a sequence of random variables.
Results
√ d
1 n(X̄ − µ) → N(0, σ 2 )
P
Xi −nµ d
2 i√
σ n
→ N(0, 1)
P
X −nµ d
3 i √i
n
→ N(0, σ 2 )
X − np d
p → N(0, 1)
np(1 − p)
Alternatively,
g (Yn ) − g (θ) d
√ → N(0, 1).
g 0 (θ)/ n
4 Inequalities
P(T | D)P(D)
We require P(D | T ) = .
P(T )
0.99 × 0.02
∴ P(D | T ) = ≈ 0.5025
0.0394
A lot of people get stuck with Bayes’ law, especially when used with other
results. Use a tree diagram!
Example
Given the distribution of X below, compute its expectation and standard
deviation.
x 0 3 9 27
P(X = x) 0.3 0.1 0.5 0.1
Example
Given the distribution of X below, compute its expectation and standard
deviation.
x 0 3 9 27
P(X = x) 0.3 0.1 0.5 0.1
X
E[X ] = x P(X = x)
all x
= 0 × 0.3 + 3 × 0.1 + 9 × 0.5 + 27 × 0.1
= 7.5
Example
Given the distribution of X below, compute its expectation and standard
deviation.
x 0 3 9 27
P(X = x) 0.3 0.1 0.5 0.1
E[X ] = 7.5
Example
Given the distribution of X below, compute its expectation and standard
deviation.
x 0 3 9 27
P(X = x) 0.3 0.1 0.5 0.1
E[X ] = 7.5
E[X 2 ] = 114.3
q p √
σX = E[X 2 ] − (E[X ])2 = 114.3 − 7.52 = 58.05 ≈ 7.619
X ∞
X
E[X ] = x P(X = x) = xp(1 − p)x−1
all x x=1
∞
X
E[X ] = xp(1 − p)x−1
x=1
∞
X
= (y + 1)p(1 − p)y (y = x − 1)
y =0
X∞
= (1 − p) (y + 1)p(1 − p)y −1
y =0
∞
X
E[X ] = xp(1 − p)x−1
x=1
∞
X
= (y + 1)p(1 − p)y (y = x − 1)
y =0
X∞
= (1 − p) (y + 1)p(1 − p)y −1
y =0
∞
X ∞
X
y −1
= (1 − p) yp(1 − p) + (1 − p) p(1 − p)y −1
y =0 y =0
∞
X
E[X ] = xp(1 − p)x−1
x=1
∞
X ∞
X
= (1 − p) yp(1 − p)y −1 + (1 − p) p(1 − p)y −1
y =0 y =0
∞
X ∞
X
= (1 − p) yp(1 − p)y −1 + (1 − p) p(1 − p)y −1
y =1 y =1
−1
+ p(1 − p) (evaluating at y = 0)
−1
= (1 − p)E[X ] + (1 − p) 1 + p(1 − p)
∴ pE[X ] = (1 − p) + p
1
E[X ] =
p
In general, can be done with the aid of Taylor series or binomial theorem.
But preferably just do this:
Example
2
Let fX (x) = θ2
x for 0 < x < θ. Compute the MGF and (2901) assert its
existence.
Example
2
Let fX (x) = θ2
x for 0 < x < θ. Compute the MGF and (2901) assert its
existence.
Integrate by parts
Z θ
uX 2
mX (u) = E[e ]= 2 xe ux dx
θ 0
!
θ θ
xe ux e ux
Z
2
= 2 − dx
θ u 0 0 u
Example
2
Let fX (x) = θ2
x for 0 < x < θ. Compute the MGF and (2901) assert its
existence.
Slowly tidy everything up
Z θ
uX 2
mX (u) = E[e ]= 2 xe ux dx
θ 0
!
θ θ
xe ux e ux
Z
2
= 2 − dx
θ u 0 0 u
!
θ
2θe uθ 2 e ux
= 2
− 2
uθ θ u2 0
Example
2
Let fX (x) = θ2
x for 0 < x < θ. Compute the MGF and (2901) assert its
existence.
Slowly tidy everything up
Z θ
uX 2
mX (u) = E[e ]= 2 xe ux dx
θ 0
!
θ θ
xe ux e ux
Z
2
= 2 − dx
θ u 0 0 u
!
θ
2θe uθ 2 e ux
= 2
− 2
uθ θ u2 0
2(uθe uθ − e uθ + 1)
=
u 2 θ2
Example
2
Let fX (x) = θ2
x for 0 < x < θ. Compute the MGF and (2901) assert its
existence.
2(uθe uθ − e uθ + 1)
mX (u) =
u 2 θ2
GeoGebra simulation
Example
2
Let fX (x) = θ2
x for 0 < x < θ. Compute the MGF and (2901) assert its
existence.
Idea: Can check that the limit as u → 0 is finite. The finiteness of the limit
implies the required result.
2 θe uθ + uθ2 e uθ − θe uθ
2(uθe uθ − e uθ + 1) LH
lim = lim
u→0 u 2 θ2 u→0 2uθ2
uθ
= lim e
u→0
=1
Example
Use the MGF of X ∼ Bin(n, p) to prove that E[X ] = np.
d
E[X ] = lim (1 − p + pe u )n
u→0 du
Example
Use the MGF of X ∼ Bin(n, p) to prove that E[X ] = np.
d
E[X ] = lim (1 − p + pe u )n
du
u→0
= lim n(1 − p + pe u )n−1 · pe u
u→0
= n(1 − p + p)n−1 · p
= np
Example
A busy switchboard receives 150 calls an hour on average. Assume that
calls are independent from each other and can be modelled with a Poisson
distribution. Find the probability of
1 Exactly 3 calls in a given minute
2 At least 10 calls in a given 5 minute period.
Naive:
X ∼ Poisson(150).
Example
A busy switchboard receives 150 calls an hour on average. Assume that
calls are independent from each other and can be modelled with a Poisson
distribution. Find the probability of
1 Exactly 3 calls in a given minute
2 At least 10 calls in a given 5 minute period.
2.53
P(X = 3) = e −2.5 ≈ 0.2138
3!
Example
A busy switchboard receives 150 calls an hour on average. Assume that
calls are independent from each other and can be modelled with a Poisson
distribution. Find the probability of
1 Exactly 3 calls in a given minute
2 At least 10 calls in a given 5 minute period.
Example
A busy switchboard receives 150 calls an hour on average. Assume that
calls are independent from each other and can be modelled with a Poisson
distribution. Find the probability of
1 Exactly 3 calls in a given minute
2 At least 10 calls in a given 5 minute period.
= e −5/24
Um, ye wat?
Example
A random variable has the following distribution:
x -1 0 1 2
P(X = x) 0.38 0.21 0.14 0.27
Example
A random variable has the following distribution:
x -1 0 1 2
P(X = x) 0.38 0.21 0.14 0.27
Example
A random variable has the following distribution:
x -1 0 1 2
P(X = x) 0.38 0.21 0.14 0.27
Example
A random variable has the following distribution:
x -1 0 1 2
P(X = x) 0.38 0.21 0.14 0.27
√
y
(
e −λ (λ√y )! if y = 0, 1, 4, 9, . . .
P(Y = y ) =
0 otherwise
Example
Let X ∼ Exp(λ). What is the density of Y = X 2 ?
Example
Let X ∼ Exp(λ). What is the density of Y = X 2 ?
√ 1
∴ fY (y ) = fX ( y ) √
2 y
Example
Let X ∼ Exp(λ). What is the density of Y = X 2 ?
√ 1
∴ fY (y ) = fX ( y ) √
2 y
1 √ 1
= e − y /λ √
λ 2 y
1 √
= √ e − y /λ
2λ y
Example
Let X ∼ Exp(λ). What is the density of Y = X 2 ?
1 √
fY (y ) = √ e − y /λ
2λ y
Since x > 0 and y = x 2 , y > 0 as well.
Example
Let X ∼ Unif(−10, 10). What is the density of Y = X 2 ?
1
fY (y ) = √
20 y
Since −10 < x < 10 and y = x 2 , we must have 0 < y < 100.
Example
The joint probability distribution of X and Y is
y
0 1 2
0 1/16 1/8 1/8
x 1 1/8 1/16 0
2 3/16 1/4 1/16
1
P(X = 0, Y = 1) =
8
Example
The joint probability distribution of X and Y is
y
0 1 2
0 1/16 1/8 1/8
x 1 1/8 1/16 0
2 3/16 1/4 1/16
Example
The joint probability distribution of X and Y is
y
0 1 2
0 1/16 1/8 1/8
x 1 1/8 1/16 0
2 3/16 1/4 1/16
Example
1
fX ,Y (x, y ) = x ≥ 1, y ≥ 1
x 2y 2
is the joint density of the continuous r.v.s X and Y . Find P(X < 2, Y ≥ 4)
and P(X ≤ Y 2 ).
Example
1
fX ,Y (x, y ) = x ≥ 1, y ≥ 1
x 2y 2
is the joint density of the continuous r.v.s X and Y . Find P(X < 2, Y ≥ 4)
and P(X ≤ Y 2 ).
Z 2Z ∞
1
P(X < 2, Y ≥ 4) = dy dx
1 4 x 2y 2
Z 2
1
= dx
1 4x 2
1
=
8
Example
1
fX ,Y (x, y ) = x ≥ 1, y ≥ 1
x 2y 2
is the joint density of the continuous r.v.s X and Y . Find P(X < 2, Y ≥ 4)
and P(X ≤ Y 2 ).
Z ∞ Z x2
2 1
P(X ≤ Y ) = dy dx
x 2y 2
Z1 ∞ 1
1 1
= − dx
1 x2 x4
2
=
3
Example
Find E[Y 2 ln X ] for the following distribution
y
1 2
x 1 1/10 1/5
2 3/10 2/5
Example
Find E[Y 2 ln X ] for the following distribution
y
1 2
x 1 1/10 1/5
2 3/10 2/5
Example
Find E[Y 2 ln X ] for the following distribution
y
1 2
x 1 1/10 1/5
2 3/10 2/5
Problem
Examine the existence of E[XY ] for the earlier example:
1
fX ,Y (x, y ) = for x, y ≥ 1.
x 2y 2
FX ,Y (x, y ) = P(X ≤ x, Y ≤ y )
Example
For the earlier example, FX ,Y (x, y ) = 0 if x < 1 or y < 1. Else:
Z xZ y
1 1 1
FX ,Y (x, y ) = 2 2
du dv = 1 − 1−
1 1 u v x y
Example
Test if X and Y are independent, for
1
fX ,Y (x, y ) = x, y ≥ 1.
x 2y 2
Example
Test if X and Y are independent, for
1
fX ,Y (x, y ) = x, y ≥ 1.
x 2y 2
Z ∞
1
fX (x) = dy
1 x 2y 2
1
= x ≥1
x2
1
Similarly fY (y ) = y ≥ 1.
y2
Therefore since fX ,Y (x, y ) = fX (x)fY (y ), X and Y are independent.
Example
Determine P(X = x | Y = 2), i.e. fX |Y (x | 2), for
y
1 2
x 1 1/10 1/5
2 3/10 2/5
Example
Determine P(X = x | Y = 2), i.e. fX |Y (x | 2), for
y
1 2
x 1 1/10 1/5
2 3/10 2/5
3
P(Y = 2) =
5
P(X = 1, Y = 2) 1
P(X = 1 | Y = 2) = =
P(Y = 2) 3
P(X = 2, Y = 2) 2
P(X = 2 | Y = 2) = =
P(Y = 2) 3
fY |X (y | x) = fY (y )
or
fX |Y (x | y ) = fX (x)
Investigation
For the earlier example with fX ,Y (x, y ) = x −2 y −2 for x ≥ 1, y ≥ 1, prove
the independence of X and Y using this lemma instead.
(And similarly for Y . Basically, just add the condition to the original
formula.)
Example
Find E[X | Y = 2] and Var(X | Y = 2) for
y
1 2
x 1 1/10 1/5
2 3/10 2/5
Example
Find E[X | Y = 2] and Var(X | Y = 2) for
y
1 2
x 1 1/10 1/5
2 3/10 2/5
Example
Find E[X | Y = 2] and Var(X | Y = 2) for
y
1 2
x 1 1/10 1/5
2 3/10 2/5
2
2 5 2
Var(X | Y = 2) = 3 − =
3 9
Example
Let fX ,Y (x, y ) = xy for x ∈ [0, 1], y ∈ [0, 2]. Determine their covariance in
the old fashioned way.
Example
Let fX ,Y (x, y ) = xy for x ∈ [0, 1], y ∈ [0, 2]. Determine their covariance in
the old fashioned way.
Example
Let fX ,Y (x, y ) = xy for x ∈ [0, 1], y ∈ [0, 2]. Determine their covariance in
the old fashioned way.
Example
Let fX ,Y (x, y ) = xy for x ∈ [0, 1], y ∈ [0, 2].Determine their covariance in
the old fashioned way.
Example (2901)
Let Z ∼ N (0, 1) and W satisfy P(X = 1) = P(X = −1) = 12 . Suppose
that W and Z are independent and define X := WZ .
Example (2901)
Let Z ∼ N (0, 1) and W satisfy P(X = 1) = P(X = −1) = 12 . Suppose
that W and Z are independent and define X := WZ .
Example (2901)
Let Z ∼ N (0, 1) and W satisfy P(X = 1) = P(X = −1) = 12 . Suppose
that W and Z are independent and define X := WZ .
Observe that
U = 21 (X − Y ) and V = Y .
1
U = (X − Y ) and V = Y .
2
We have y = v and
1
u = (x − v ) =⇒ x = 2u + v .
2
2 1
∴J= and det(J) = 2.
0 1
1
U = (X − Y ) and V = Y .
2
1 −(x+y )/4
fX ,Y (x, y ) =
e
16
Since y = v and x = 2u + v , we get x + y = 2u + 2v . Therefore
1
fU,V (u, v ) = e −(u+v )/2 .
8
1
U = (X − Y ) and V = Y .
2
We know that y > 0. Since v = y , it immediately follows that v > 0.
1
U = (X − Y ) and V = Y .
2
We know that y > 0. Since v = y , it immediately follows that v > 0.
However, x > 0 and x = 2u + v . Therefore:
2u + v > 0
v
u>−
2
Example
Let X and Y be i.i.d. Geom(p). Use convolutions to find the probability
function of Z := X + Y .
for z − y = 1, 2, 3, . . .,
Example
Let X and Y be i.i.d. Geom(p). Use convolutions to find the probability
function of Z := X + Y .
for z − y = 1, 2, 3, . . ., i.e.
y − z = . . . , −3, −2, −1 ⇐⇒ y = . . . , z − 3, z − 2, z − 1
Example
Let X and Y be i.i.d. Geom(p). Use convolutions to find the probability
function of Z := X + Y .
y = 0, 1, 2, . . .
and y = . . . , z − 3, z − 2, z − 1.
Therefore, y = 0, 1, 2, . . . , z − 3, z − 2, z − 1.
Example
Let X and Y be i.i.d. Geom(p). Use convolutions to find the probability
function of Z := X + Y .
z−1
X
∴ P(Z = z) = p 2 (1 − p)z
y =0
= zp (1 − p)z
2
(sum only depends on y !)
Example
Let X and Y be i.i.d. Geom(p). Use convolutions to find the probability
function of Z := X + Y .
z−1
X
∴ P(Z = z) = p 2 (1 − p)z
y =0
= zp (1 − p)z
2
(sum only depends on y !)
Example
Let X and Y be i.i.d. Exp(1). Prove that Z := X + Y follows a
Gamma(2, 1) distribution using a convolution.
Example
Let X and Y be i.i.d. Exp(1). Prove that Z := X + Y follows a
Gamma(2, 1) distribution using a convolution.
Example
Let X and Y be i.i.d. Exp(1). Prove that Z := X + Y follows a
Gamma(2, 1) distribution using a convolution.
Z z
∴ fZ (z) = e −z dy
0
−z
=e z
Example
Let X and Y be i.i.d. Exp(1). Prove that Z := X + Y follows a
Gamma(2, 1) distribution using a convolution.
Z z
∴ fZ (z) = e −z dy
0
−z
=e z
e −z/1 z 2−1
=
Γ(2)12
Example
Let X and Y be i.i.d. Exp(1). Prove that Z := X + Y follows a
Gamma(2, 1) distribution from quoting MGFs.
Example
Let X and Y be i.i.d. Exp(1). Prove that Z := X + Y follows a
Gamma(2, 1) distribution from quoting MGFs.
1 1
mX (u) = 1−u and mY (u) = 1−u . So clearly
2
1
mZ (u) = mX (u)mY (u) = ,
1−u
Example
Let X1 , . . . , Xn be a sequence of i.i.d. Unif(0, 1) random variables. Define
d
Yn = n min{U1 , . . . , Un }. Prove that Yn → Y , where Y ∼ Exp(1).
Example
Let X1 , . . . , Xn be a sequence of i.i.d. Unif(0, 1) random variables. Define
d
Yn = n min{U1 , . . . , Un }. Prove that Yn → Y , where Y ∼ Exp(1).
Example
Let X1 , . . . , Xn be a sequence of i.i.d. Unif(0, 1) random variables. Define
d
Yn = n min{U1 , . . . , Un }. Prove that Yn → Y , where Y ∼ Exp(1).
Example
Let X1 , . . . , Xn be a sequence of i.i.d. Unif(0, 1) random variables. Define
d
Yn = n min{U1 , . . . , Un }. Prove that Yn → Y , where Y ∼ Exp(1).
y y
FYn (y ) = 1 − P U1 > , . . . , Un >
n n y
y
= 1 − P U1 > . . . P Un > (independence)
n i n
h y n
= 1 − P U1 > (id. distributed)
n
Example
Let X1 , . . . , Xn be a sequence of i.i.d. Unif(0, 1) random variables. Define
d
Yn = n min{U1 , . . . , Un }. Prove that Yn → Y , where Y ∼ Exp(1).
y y
FYn (y ) = 1 − P U1 > , . . . , Un >
n n y
y
= 1 − P U1 > . . . P Un > (independence)
n i n
h y n
= 1 − P U1 > (id. distributed)
"Z #nn
1 y n
=1− 1 dt = 1 − 1 −
y /n n
Example
Let X1 , . . . , Xn be a sequence of i.i.d. Unif(0, 1) random variables. Define
d
Yn = n min{U1 , . . . , Un }. Prove that Yn → Y , where Y ∼ Exp(1).
∴ lim FYn (y ) = 1 − e −y = FY (y )
n→∞
d
Hence Yn → Y .
X − 68 d
√ →Z
4/ 40
X40 − 68 80 − 68
∴ P(X40 > 80) = P √ > √
4/ 40 4/ 40
80 − 68
≈P Z > √
4/ 40
X40 − 68 80 − 68
∴ P(X40 > 80) = P √ > √
4/ 40 4/ 40
80 − 68
≈P Z > √
4/ 40
√
= P(Z > 3 40)
= 1-pnorm(3*sqrt(40))
or pnorm(3*sqrt(40), lower.tail=FALSE)
X − np d
p → N (0, 1)
np(1 − p)
Simplifying, we have
√ d
n[(¯(X )3n − 8] → N(0, 1008).
Prove that the real part of every non-trivial zero of the Riemann zeta
function is 12 .