T 4 Sampling Distributions
T 4 Sampling Distributions
Rohini Somanathan
Since estimators are statistics, they are random variables with their own probability distributions
We call these sampling distributions because they are induced by the sample and determined by the joint
distribution of the sample
The sampling distribution of estimator based on a random sample is determined by F (the population from
which the sample is drawn) and n, the sample size
& %
Page 1 Rohini Somanathan
' $
Examples:
1. What if Xi ∼ N(θ, 4), and we want E(X̄n − θ)2 ≤ .1? This is simply the variance of X̄n , and we know
X̄n ∼ N(θ, 4/n).
4
≤ .1 if n ≥ 40
n
2. Consider a random sample of size n from a Uniform distribution on [0, θ], and the statistic
U = max{X1 , . . . , Xn }. The CDF of U is given by:
0 if u ≤ 0
n
F(u) = u
if 0 < u < θ
θ
1 if u ≥ θ
We can now use this to see how large our sample must be if we want a certain level of precision in our estimate
for θ. Suppose we want the probability that our estimate lies within .1θ for any level of θ to be bigger than 0.95:
We want this to be bigger than 0.95, or 0.9n ≤ 0.05. With the LHS decreasing in n, we choose
log(.05)
n ≥ log(.9) = 28.43. Our minimum sample size is therefore 29.
& %
Page 2 Rohini Somanathan
' $
Chi-square: The Chi-square with n degrees of freedom (χ2n ) is in the family of Gamma distributions.
We will now define these and connect them with each other.
& %
Page 3 Rohini Somanathan
' $
1 P
n
For a random sample from a Normal distribution, the MLEs are µ̂ = X̄n and σ̂2 = n (Xi − X̄n )2 .
i=1
Theorem: If X1 , . . . Xn form a random sample from a normal distribution with mean µ and variance σ2 , then
1 P
n
the sample mean X̄n and the sample variance n (Xi − X̄n )2 are independent random variables and
i=1
σ2
X̄n ∼ N(µ, )
n
P
n
(Xi − X̄n )2
i=1
∼ χ2n−1
σ2
& %
Page 4 Rohini Somanathan
' $
The t-distribution
Let Z ∼ N(0, 1), let Y ∼ χ2v , and let Z and Y be independent random variables. Then
Z
X = q ∼ tv
Y
v
The shape of the density is similar to that of the standard normal (bell-shaped) but with fatter tails.
& %
Page 5 Rohini Somanathan
' $
√
n(X −µ) S2
Proof: We know that σ
n
∼ N(0, 1) and that σn2 ∼ χ2n−1 . Dividing the first random variable by the square
root of the second, divided by its degrees of freedom, the σ in the numerator and denominator cancels to obtain U.
Implication: We cannot make statements about |X̄n − µ| using the normal distribution if σ2 is unknown. This result
P
n
(Xn −µ)
allows us to use its estimate σ̂ = (Xi − X̄n )2 /n since σ̂/√
2
n−1
∼ tn−1
i=1
RESULT 2 As n → ∞, U −→ Z ∼ N(0, 1)
q √
n(Xn −µ)
To see why: U can be written as n−1
n σ̂ ∼ tn−1 . As n gets large σ̂ gets very close to σ and n−1
n is close
to 1.
F−1 (.55) = .129 for t10 , .127 for t20 and .126 for the standard normal distribution. The differences between these
values increases for higher values of their distribution functions (why?)
& %
Page 6 Rohini Somanathan
' $
Given σ2 , let us see how we can obtain an interval estimate for µ, i.e. an interval which is likely to contain µ with a
pre-specified probability.
(Xn −µ) (Xn −µ)
Since σ/√n ∼ N(0, 1), Pr −2 < σ/√n < 2 = .955
2σ 2σ 2σ 2σ
But this event is equivalent to the events − √n
< Xn − µ < √
n
and Xn − √
n
< µ < Xn + √
n
2σ 2σ
With known σ, each of the random variables Xn − √ n
and Xn + √ n
are statistics. Therefore, we have
derived a random interval within which the population parameter lies with probability .955, i.e.
2σ 2σ
Pr Xn − √ < µ < Xn + √ = .955 = γ
n n
Notice that there are many intervals for the same γ, this is the shortest one.
Now, given our sample, our statistics take particular values and the resulting interval either contains or does not
contain µ. We can therefore no longer talk about the probability that it contains µ because the experiment has
already been performed.
2σ 2σ
We say that (xn − √ n
< µ < xn + √ n
) is a 95.5% confidence interval for µ. Alternatively, we may say that µ
lies in the above interval with confidence γ or that the above interval is a confidence interval for µ with
confidence coefficient γ
& %
Page 7 Rohini Somanathan
' $
& %
Page 8 Rohini Somanathan
' $
Let X denote the sample mean of a random sample of size 25 from a distribution
with variance 100 and mean µ. In this case, √σn = 2 and, making use of the central
limit theorem the following statement is approximately true:
(Xn − µ)
Pr −1.96 < < 1.96 = .95 or Pr Xn − 3.92 < µ < Xn + 3.92 = .95
2
If the sample mean is given by xn = 67.53, an approximate 95% confidence interval
for the sample mean is given by (63.61, 71.45).
& %
Page 9 Rohini Somanathan
' $
We know that
(Xn − µ)
√ ∼ tn−1
σ̂/ n − 1
so can use the t-distribution with (n − 1) degrees of freedom to construct our
interval estimate.
& %
Page 10 Rohini Somanathan
' $
Suppose we want a 95% confidence interval for the difference in the means:
Using the above t-distribution, we find a number b for which Pr −b < X < b = .95
The random interval (X̄ − Ȳ) − bR, (X̄ − Ȳ) + bR contains the true difference in means with 95% probability.
A confidence interval is now based on sample values, (x̄n − ȳm ) and corresponding sample variances.
Based on the CLT, we can use the same procedure even when our samples are not normal.
& %
Page 11 Rohini Somanathan
' $
The F-distribution
RESULT : Let Y ∼ χ2m , Z ∼ χ2n , and let Y and Z be independent random variables. Then
Y/m nY
F= =
Z/n mZ
Γ( m+n
2 )m
m/2 nn/2
xm/2−1
f(x) = I
(m+n)/2 (0,∞)
(x)
Γ( m
2 )Γ( n
2 ) (mx + n)
– m andn are referred to as the numerator and denominator degrees of freedom respectively.
– the F density is defined on the positive real numbers and is skewed to the right.
– from the definition of the F distribution (the ratio of two χ2 distributions ), it follows that
1
= Fv2 ,v1
Fv1 ,v2
So if X ∼ F(m, n), 1
X ∼ F(n, m).
– The square of a random variable with a t distribution with n degrees of freedom has an F distribution with
(1, n) degrees of freedom.
The F test is used in many hypothesis testing problems (e.g. testing for equality of variances of two distributions)
& %
Page 12 Rohini Somanathan