Lecture 03 & 04 EC5002 CLT & Normal Distribution 2015-16
Lecture 03 & 04 EC5002 CLT & Normal Distribution 2015-16
(EC5002)
Lecture 3 and 4
The Central Limit Theorem, the normal distribution,
confidence intervals and hypothesis testing
JUS:EC5002:2013-14:L3
Learning outcomes:
By the end of this class and completion of the
reading you should be able to know:
The Central Limit Theorem
The sample distribution of X-bar
Confidence Intervals
Hypothesis testing
JUS:EC5002:2013-14:L3
and variance:
2x
That is: X ~ N( , x )
x n
JUS:EC5002:2013-14:L3
X-
2
Z~
where
Standard Normal:
0.4
n
2 25
21 - 20
a) P( X 21) P(Z
) P(Z 2.5) 0.0062 (very unlikely)
0.4
b) P( X 18) P(Z
JUS:EC5002:2013-14:L3
18 - 20
) P(Z 5) 0 (so unlikely to be virtually zero)
0.4
Confidence intervals
Use sample information and find a confidence interval which is
very likely to contain the true population parameter X.
Statisticians often choose a probability of 0.95 (or 95%), i.e. 1 in
20 chance of being wrong.
The sample mean has a normal distribution with a large enough
sample size, use standardised normal (Z) distribution.
From Z table: Z values of -1.96 to + 1.96 include 95% of all Z
values, i.e. P(-1.96 < Z < + 1.96) = 0.95
JUS:EC5002:2013-14:L3
Confidence intervals
X X
P(-1.96
1.96) 0.95
n
X X X X
note that
P(-1.96
(X ) n
X
1.96) 0.95
Multiply by
P(-1.96 ( X ) n 1.96 ) 0.95 divide by n
X
divide by n
P -1.96
X X 1.96
0.95
n
n
JUS:EC5002:2013-14:L3
Substract X
X
X
-1.96
1
.
96
X
n
n
0.95
Multiply by (-1)
P X 1.96
n
P X 1.96
n
JUS:EC5002:2013-14:L3
X 1.96
n
X 1.96
n
0.95
0.95
P X -1.96
X X 1.96
n
n
P( 19.5-1.96 0.4)
P( 18.716
JUS:EC5002:2013-14:L3
0.95
20.284) 0.95
8
JUS:EC5002:2013-14:L3
Hypothesis Testing
Hypothesize, X = 13, i.e. test whether sample information we
have leads us to suppose X = 13. This is called
NULL of maintained hypothesis: H0: X = 13
The alternative hypothesis, HA (or H1): X 13 (two tailed test)
or
HA: X > 13 or (< 13) (one tailed test).
If X were long way from 13 then we would not maintain
H0: X = 13. Use standardise distance (to take out scaling factors)
by dividing the s.e. This gives us a t value, that is the distance
expressed in numbers of s.e. from 0. This is known as,
X
X
Test statistics
~ t with degrees of freedom (d.f.) n - 1
s n
JUS:EC5002:2013-14:L3
10
The test statistic lies in the tails (rejection region) reject the
null hypothesis.
In our example, test statistic is -3.4826, inside the rejection
reject the null that true population mean is 13.
The degree of certainty: 0.99 (99%) with about 50 d.f.: CV=-2.68
to + 2.68
One tail test
JUS:EC5002:2013-14:L3
11
Error
The degree of certainty: 0.95 (95%) chances of being wrong:
0.05 (5%). This is known as the level of significance or size of the
test. We choose this level so we are choosing the chance of being
wrong. The ways of being wrong are:
TYPE I: rejecting the null (H0) given that it is true.
TYPE II: accepting the null (H0) given that is false.
When we choose the size of a test we are choosing the probability
of committing a type I error.
In general, when testing we want to minimize the probability of
committing type II error given the chosen size of the test.
JUS:EC5002:2013-14:L3
12
X
s
Make a decision.
JUS:EC5002:2013-14:L3
13