0% found this document useful (0 votes)
23 views96 pages

Statistics Training

The document discusses statistical inference and covers topics like the weak law of large numbers, central limit theorem, estimation theory, Markov's inequality, Chebyshev's inequality, and their applications. It also provides an example explaining how to determine the minimum sample size needed for a pre-election survey to estimate the population proportion within 1% error at 95% confidence level.

Uploaded by

tgnr890
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
23 views96 pages

Statistics Training

The document discusses statistical inference and covers topics like the weak law of large numbers, central limit theorem, estimation theory, Markov's inequality, Chebyshev's inequality, and their applications. It also provides an example explaining how to determine the minimum sample size needed for a pre-election survey to estimate the population proportion within 1% error at 95% confidence level.

Uploaded by

tgnr890
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 96

STATISTICAL INFERENCE LECTURE

- Dripto Bakshi
Inferential Statistics

 Inferential statistics: the part of statistics that allows


researchers to generalize their findings beyond data
collected.

 Statistical inference: a procedure for making inferences or


generalizations about a larger population from a sample of
that population
Lecture Parts

 Weak Law of Large Numbers & Central Limit Theorem

 Estimation Theory
Weak Law of Large Numbers & Central Limit
Theorem
Pre - Poll / Election Survey
Markov’ Inequality

 Consider any positive random variable X∈ (0, ∞) with density 𝑓𝑋 .


Markov’ Inequality

 Consider any positive random variable X∈ (0, ∞) with density 𝑓𝑋 .

+∞
 E(X) = ‫׬‬0 𝑥 𝑓𝑋 (x) dx
Markov’ Inequality

 Consider any positive random variable X∈ (0, ∞) with density 𝑓𝑋 .

+∞ +∞
 E(X) = ‫׬‬0 𝑥 𝑓𝑋 (x) dx ≥ ‫𝑋𝑓 𝑥 𝑎׬‬ (x) dx , where a > 0
Markov’ Inequality

 Consider any positive random variable X∈ (0, ∞) with density 𝑓𝑋 .

+∞ +∞
 E(X) = ‫׬‬0 𝑥 𝑓𝑋 (x) dx ≥ ‫𝑋𝑓 𝑥 𝑎׬‬ (x) dx , where a > 0

+∞ +∞
 ‫𝑋𝑓 𝑥 𝑎׬‬ (x) dx ≥ ‫𝑎 𝑎׬‬ . 𝑓𝑋 (x) dx = a . P(X ≥ a)
Markov’ Inequality

 Consider any positive random variable X∈ (0, ∞) with density 𝑓𝑋 .

+∞ +∞
 E(X) = ‫׬‬0 𝑥 𝑓𝑋 (x) dx ≥ ‫𝑋𝑓 𝑥 𝑎׬‬ (x) dx , where a > 0

+∞ +∞
 ‫𝑋𝑓 𝑥 𝑎׬‬ (x) dx ≥ a ‫׬‬0 1. 𝑓𝑋 (x) dx = a . P(X ≥ a)

𝐸(𝑋)
 P(X ≥ a) ≤
𝑎
Chebyshev’s Inequality
 Consider any random variable X ~ 𝐹𝑋 (𝜇, 𝜎 2 )
Chebyshev’s Inequality
 Consider any random variable X ~ 𝐹𝑋 (𝜇, 𝜎 2 )

 Define Q = (𝑋 − 𝜇)2 . Q is definitely a positive random variable.


Chebyshev’s Inequality
 Consider any random variable X ~ 𝐹𝑋 (𝜇, 𝜎 2 )

 Define Q = (𝑋 − 𝜇)2 . Q is definitely a positive random variable.

 By Markov’s inequality, for any 𝜀 > 0

𝐸((𝑋 −𝜇)2 ) 𝜎2
P ((𝑋 − 𝜇)2 > 𝜀 2 ) ≤ =
𝜀2 𝜀2
Chebyshev’s Inequality
 Consider any random variable X ~ 𝐹𝑋 (𝜇, 𝜎 2 )

 Define Q = (𝑋 − 𝜇)2 . Q is definitely a positive random variable.

 By Markov’s inequality, for any 𝜀 > 0

2 2 𝐸((𝑋 −𝜇)2 ) 𝜎 2
P ((𝑋 − 𝜇) > 𝜀 ) ≤ = 2
𝜀2 𝜀

𝜎2
 P(|X − µ| ≥ 𝜀) ≤ 𝜀2
Weak Law of Large Numbers

 𝑋1 , 𝑋2 , 𝑋3 ,…… 𝑋𝑛 are iid, where 𝑋𝑗 ~ F (𝜇, 𝜎 2 ) ∀ 𝑗 = 1,2, … . . 𝑛


Weak Law of Large Numbers

 𝑋1 , 𝑋2 , 𝑋3 ,…… 𝑋𝑛 are iid, where 𝑋𝑗 ~ F (𝜇, 𝜎 2 ) ∀ 𝑗 = 1,2, … . . 𝑛

𝑋1 + 𝑋2 + 𝑋3 ….𝑋𝑛
 𝑀𝑛 =
𝑛
Weak Law of Large Numbers

 𝑋1 , 𝑋2 , 𝑋3 ,…… 𝑋𝑛 are iid, where 𝑋𝑗 ~ F (𝜇, 𝜎 2 ) ∀ 𝑗 = 1,2, … . . 𝑛

𝑋1 + 𝑋2 + 𝑋3 ….𝑋𝑛
 𝑀𝑛 =
𝑛

 What happens if n -> ∞ ?


Convergence in Probability
𝜎2
 E(𝑀𝑛 ) = 𝜇 & V(𝑀𝑛 ) =
𝑛
Convergence in Probability
𝜎2
 E(𝑀𝑛 ) = 𝜇 & V(𝑀𝑛 ) =
𝑛

 According to Chebyshev’s Inequality: For any finite 𝜀 > 0


𝜎2
𝑛 𝜎2
P(| 𝑀𝑛 − µ| ≥ 𝜀) ≤ =
𝜀2 𝑛𝜀 2
Convergence in Probability
𝜎2
 E(𝑀𝑛 ) = 𝜇 & V(𝑀𝑛 ) =
𝑛

 According to Chebyshev’s Inequality: For any finite 𝜀 > 0


𝜎2
𝑛 𝜎2
P(| 𝑀𝑛 − µ| ≥ 𝜀) ≤ =
𝜀2 𝑛𝜀 2

 Hence as n -> ∞ , P(| 𝑀𝑛 − µ| ≥ 𝜀) -> 0 for any 𝜀 > 0


𝑀𝑛 converges in probability to µ
Pre - Poll / Election Survey
The Methodology
 f: The proportion of people respond “voting for Republicans”
The Methodology
 f: The proportion of people respond “voting for Republicans”

 Define a random variable 𝑋𝑗 for any individual j where


𝑋𝑗 = 1 ; “vote for Republican”
= 0 ; “vote for Democrat”
The Methodology
 f: The proportion of people respond “voting for Republicans”

 Define a random variable 𝑋𝑗 for any individual j where


𝑋𝑗 = 1 ; “vote for Republican”
= 0 ; “vote for Democrat”

𝑋1 + 𝑋2 + 𝑋3 ….𝑋𝑛
 𝑀𝑛 =
𝑛

 𝑀𝑛 denotes the fraction of people responding “voting for Republicans”


The Methodology
 f: The proportion of people respond “voting for Republicans”

 Define a random variable 𝑋𝑗 for any individual j where


𝑋𝑗 = 1 ; “vote for Republican”
= 0 ; “vote for Democrat”

𝑋1 + 𝑋2 + 𝑋3 ….𝑋𝑛
 𝑀𝑛 = 𝑛

 𝑀𝑛 denotes the fraction of people responding “voting for Republicans”

 Goal: 95% confidence of ≤1% error


P(|𝑀𝑛 − f| ≥ .01) ≤ .05
 𝑋𝑗 ~ Ber (f)
 𝑋𝑗 ~ Ber (f)

 E(𝑋𝑗 ) = f & V(𝑋𝑗 ) = f(1-f)


 𝑋𝑗 ~ Ber (f)

 E(𝑋𝑗 ) = f & V(𝑋𝑗 ) = f(1-f)

𝑋1 + 𝑋2 + 𝑋3 ….𝑋𝑛
 𝑀𝑛 =
𝑛
 𝑋𝑗 ~ Ber (f)

 E(𝑋𝑗 ) = f & V(𝑋𝑗 ) = f(1-f)

𝑋1 + 𝑋2 + 𝑋3 ….𝑋𝑛
 𝑀𝑛 =
𝑛

𝑓(1−𝑓)
 E(𝑀𝑛 ) = f & V(𝑀𝑛 ) =
𝑛
 Using Chebyshev’s Inequality
V(𝑀𝑛 )
P(|𝑀𝑛 − f| ≥ .01) ≤ 2
(.01)
 Using Chebyshev’s Inequality
𝑓(1−𝑓)
V(𝑀𝑛 ) 𝑛
P(|𝑀𝑛 − f| ≥ .01) ≤ =
(.01)2 (.01)2
 Using Chebyshev’s Inequality
𝑓(1−𝑓)
V(𝑀𝑛 ) 𝑛
P(|𝑀𝑛 − f| ≥ .01) ≤ =
(.01)2 (.01)2

𝑓(1−𝑓)
 We want 𝑛
< 0.05
(.01)2
𝑓(1−𝑓)
 We want 𝑛
< 0.05
(.01)2

𝑓(1−𝑓)
 Choose the sample size (n) such that 𝑛
< 0.05
(.01)2
𝑓(1−𝑓)
 We want 𝑛
(.01)2
< 0.05

𝑓(1−𝑓)
 Choose the sample size (n) such that 𝑛
(.01)2
< 0.05

 But we don’t know f

 In fact we are trying to find / estimate f.


 f ∈ 0,1 & 𝑓 1 − 𝑓 ≤ ¼
 f ∈ 0,1 & 𝑓 1 − 𝑓 ≤ ¼

𝑓(1−𝑓)
1
 Therefore 𝑛

(.01)2 4𝑛(.01)2
 f ∈ 0,1 & 𝑓 1 − 𝑓 ≤ ¼

𝑓(1−𝑓)
1
 Therefore 𝑛

(.01)2 4𝑛(.01)2

1
 Now < 0.05 if n ≥ 50000
4𝑛(.01)2
 f ∈ 0,1 & 𝑓 1 − 𝑓 ≤ ¼

𝑓(1−𝑓)
1
 Therefore 𝑛

(.01)2 4𝑛(.01)2

1
 Now < 0.05 if n ≥ 50000
4𝑛(.01)2

 So the minimum sample size (n) = 50000 !!!!


Central Limit Theorem
 𝑋1 , 𝑋2 , 𝑋3 ,…… 𝑋𝑛 are iid, where 𝑋𝑗 ~ F (𝜇, 𝜎 2 ) ∀ 𝑗 = 1,2, … . . 𝑛

 E(𝑋𝑗 ) = 𝜇 & V(𝑋𝑗 ) =𝜎 2 ∀ 𝑗 = 1,2, … . . 𝑛


Central Limit Theorem
 𝑋1 , 𝑋2 , 𝑋3 ,…… 𝑋𝑛 are iid, where 𝑋𝑗 ~ F (𝜇, 𝜎 2 ) ∀ 𝑗 = 1,2, … . . 𝑛

 E(𝑋𝑗 ) = 𝜇 & V(𝑋𝑗 ) =𝜎 2 ∀ 𝑗 = 1,2, … . . 𝑛

 𝑆𝑛 = 𝑋1 + 𝑋2 + 𝑋3 … . 𝑋𝑛
Central Limit Theorem
 𝑋1 , 𝑋2 , 𝑋3 ,…… 𝑋𝑛 are iid, where 𝑋𝑗 ~ F (𝜇, 𝜎 2 ) ∀ 𝑗 = 1,2, … . . 𝑛

 E(𝑋𝑗 ) = 𝜇 & V(𝑋𝑗 ) =𝜎 2 ∀ 𝑗 = 1,2, … . . 𝑛

 𝑆𝑛 = 𝑋1 + 𝑋2 + 𝑋3 … . 𝑋𝑛

 E(𝑆𝑛 ) = n𝜇 & V(𝑆𝑛 ) = n 𝜎 2


Central Limit Theorem
 𝑋1 , 𝑋2 , 𝑋3 ,…… 𝑋𝑛 are iid, where 𝑋𝑗 ~ F (𝜇, 𝜎 2 ) ∀ 𝑗 = 1,2, … . . 𝑛

 E(𝑋𝑗 ) = 𝜇 & V(𝑋𝑗 ) =𝜎 2 ∀ 𝑗 = 1,2, … . . 𝑛

 𝑆𝑛 = 𝑋1 + 𝑋2 + 𝑋3 … . 𝑋𝑛

 E(𝑆𝑛 ) = n𝜇 & V(𝑆𝑛 ) = n 𝜎 2

𝑆𝑛 −E(𝑆𝑛 ) 𝑆𝑛 −n𝜇 𝑀𝑛 −𝜇
 Standardize: 𝑍𝑛 = = = 𝜎
𝜎 n
V(𝑆𝑛) n
 E(𝑍𝑛 ) = 0 & V(𝑍𝑛 ) = 1
 E(𝑍𝑛 ) = 0 & V(𝑍𝑛 ) = 1

 Let Z be a standard Normal random variable. i.e Z ~ 𝑁 0,1


 E(𝑍𝑛 ) = 0 & V(𝑍𝑛 ) = 1

 Let Z be a standard Normal random variable. i.e Z ~ 𝑁 0,1

 The Central Limit Theorem states that:

For every c, P(Zn ≤ c) → P(Z ≤ c) as n becomes large enough. (Convergence in


distribution)

P(Z ≤ c) is the standard normal CDF, Φ(c), available from the normal tables
Election problem – (with CLT)
𝑓(1−𝑓)
 E(𝑀𝑛 ) = f & V(𝑀𝑛 ) =
𝑛
Election problem – (with CLT)
𝑓(1−𝑓)
 E(𝑀𝑛 ) = f & V(𝑀𝑛 ) =
𝑛

 Now we want P(|𝑀𝑛 − f| ≥ .01) ≤ .05


Election problem – (with CLT)

𝑓(1−𝑓)
 E(𝑀𝑛 ) = f & V(𝑀𝑛 ) =
𝑛

 Now we want P(|𝑀𝑛 − f| ≥ .01) ≤ .05

𝑀𝑛 −f .01
 P(| |≥ ) ≤ .05
𝑓(1−𝑓) 𝑓(1−𝑓)
𝑛 𝑛
Election problem – (with CLT)

𝑓(1−𝑓)
 E(𝑀𝑛 ) = f & V(𝑀𝑛 ) =
𝑛

 Now we want P(|𝑀𝑛 − f| ≥ .01) ≤ .05

𝑀𝑛 −f .01
 P(| |≥ ) ≤ .05
𝑓(1−𝑓) 𝑓(1−𝑓)
𝑛 𝑛

.01
 P(|Z| ≥ ) ≤ .05
𝑓(1−𝑓)
𝑛
 f ∈ 0,1 & 𝑓 1 − 𝑓 ≤ ¼
 f ∈ 0,1 & 𝑓 1 − 𝑓 ≤ ¼ => 𝑓 1 − 𝑓 ≤ ½
 f ∈ 0,1 & 𝑓 1 − 𝑓 ≤ ¼ => 𝑓 1 − 𝑓 ≤ ½

.01 .01 𝑛
 ≥ 1 = .02 𝑛
𝑓(1−𝑓)
2
𝑛
 f ∈ 0,1 & 𝑓 1 − 𝑓 ≤ ¼ => 𝑓 1 − 𝑓 ≤ ½

.01 .01 𝑛
 ≥ 1 = .02 𝑛
𝑓(1−𝑓)
2
𝑛

.01
 P(|Z| ≥ ) ≤ P(|Z| ≥ .02 𝑛)
𝑓(1−𝑓)
𝑛
 If n = 10000, P(|Z| ≥ .02 𝑛) = P(|Z| ≥ 2) < .05
 If n = 10000, P(|Z| ≥ .02 𝑛) = P(|Z| ≥ 2) < .05

 In fact P(|Z| ≥ .02 𝑛) = .05 implies .02 𝑛 = 1.96 i.e n = 9604.


 If n = 10000, P(|Z| ≥ .02 𝑛) = P(|Z| ≥ 2) < .05

 In fact P(|Z| ≥ .02 𝑛) = .05 implies .02 𝑛 = 1.96 i.e n = 9604.

 And 9604 << 50000 !!!!!


ESTIMATION
Estimation - categories

 Point Estimation

 Interval estimation
Point Estimation
 The Statistic

 The desirable properties of an estimator:


i. Unbiasedness
ii. Consistency
Interval Estimation

• Interval Estimation: an inferential statistical procedure used to


estimate population parameters from sample data through the
building of confidence intervals

• Confidence Intervals: a range of values computed from sample data


that has a known probability of capturing some population parameter
of interest
The Problem Statement

 Find the average height of Indian males.

 But India has 72 Cr males ! It’s impossible to measure everyone.

 Remedy: Collect a sample, say S = { 𝑋1 , 𝑋2 , 𝑋3 … . 𝑋𝑛 }

 Each sample observation is a realisation of a random variable. Why??

 Since the sample is drawn from the populations 𝑋𝑖 ’s are i.i.d.


The Assumption

 Let us assume that height of Indian males are normally distributed


with mean 𝜇 and variance 𝜎 2 .

 Our aim is to find (estimate) 𝜇 and 𝜎 2

 These are called “population parameters”.


Sample Statistics

1
 Sample mean 𝑋 = σ𝑛𝑖=1 𝑋𝑖

𝑛

1
2
 Sample variance 𝑠 = σ𝑛𝑖=1(𝑋𝑖 − 𝑋)
ത 2
𝑛−1
Sample Statistics

1
 Sample mean 𝑋 = σ𝑛𝑖=1 𝑋𝑖

𝑛

1
2
 Sample variance 𝑠 = σ𝑛𝑖=1(𝑋𝑖 − 𝑋)
ത 2
𝑛−1

 I would want to use the sample statistics to “estimate” the


population parameters.

 Let’s see how.


Point Estimation

ത and sample variance (𝑠 2 ) as proxies


 Just take the sample mean (𝑋)
for the population mean (𝜇) and population variance (𝜎 2 ),
respectively.

 𝑋ത and 𝑠 2 are called point estimators of 𝜇 and 𝜎 2 respectively.

 But are they “good” proxies?


The “Good” Properties

The good / desirable properties of a point estimator are

i. Unbiasedness
ii. Consistency
The “Good” Properties
The good / desirable properties of a point estimator are

i. Unbiasedness:
ത =𝜇
 E(𝑋)
 E(𝑠 2 ) = 𝜎 2

ii. Consistency
The “Good” Properties
The good / desirable properties of a point estimator are

i. Unbiasedness:

ത =𝜇
 E(𝑋)
 E(𝑠 2 ) = 𝜎 2

ii. Consistency:

ത =0
 lim 𝑉(𝑋)
𝑛→ ∞
 lim 𝑉(𝑠 2 ) = 0
𝑛→ ∞
Interval estimation of mean
 Sample observations S = { 𝑋1 , 𝑋2 , 𝑋3 … . 𝑋𝑛 }
Interval estimation of mean
 Sample observations S = { 𝑋1 , 𝑋2 , 𝑋3 … . 𝑋𝑛 }

 𝑋𝑖 ~ N (𝜇, 𝜎 2 )
Interval estimation of mean
 Sample observations S = { 𝑋1 , 𝑋2 , 𝑋3 … . 𝑋𝑛 }

 𝑋𝑖 ~ N (𝜇, 𝜎 2 )

1
 𝑋 = σ𝑛𝑖=1 𝑋𝑖

𝑛
Interval estimation of mean
 Sample observations S = { 𝑋1 , 𝑋2 , 𝑋3 … . 𝑋𝑛 }

 𝑋𝑖 ~ N (𝜇, 𝜎 2 )

1
 𝑋 = σ𝑛𝑖=1 𝑋𝑖

𝑛

𝜎2
 𝑋ത ~ N (𝜇, )
𝑛
Interval estimation of mean
 Sample observations S = { 𝑋1 , 𝑋2 , 𝑋3 … . 𝑋𝑛 }

 𝑋𝑖 ~ N (𝜇, 𝜎 2 )

1
 𝑋ത = σ𝑛𝑖=1 𝑋𝑖
𝑛

𝜎2
 𝑋ത ~ N (𝜇, )
𝑛

𝑋ത −𝜇
 Z= 𝜎 ~ N (0,1)
𝑛
The Standard Normal Distribution
The Interval
 P( 𝑍 < 𝑍𝛼/2 ) = 1 - 𝛼
The Interval
 P( 𝑍 < 𝑍𝛼/2 ) = 1 - 𝛼

 P(- 𝑍𝛼/2 < Z < 𝑍𝛼/2 ) = 1 - 𝛼


The Interval
 P( 𝑍 < 𝑍𝛼/2 ) = 1 - 𝛼

 P(- 𝑍𝛼/2 < Z < 𝑍𝛼/2 ) = 1 - 𝛼

𝑋ത −𝜇
 P(- 𝑍𝛼/2 < 𝜎 < 𝑍𝛼/2 ) = 1 - 𝛼
𝑛
The Interval
 P( 𝑍 < 𝑍𝛼/2 ) = 1 - 𝛼

 P(- 𝑍𝛼/2 < Z < 𝑍𝛼/2 ) = 1 - 𝛼

𝑋ത −𝜇
 P(- 𝑍𝛼/2 < 𝜎 < 𝑍𝛼/2 ) = 1 - 𝛼
𝑛

𝜎 𝜎
 P(𝜇 ∈ (𝑋ത − . 𝑍𝛼 , 𝑋ത + . 𝑍𝛼 ) = 1 - 𝛼
𝑛 2 𝑛 2
The Interval
 P( 𝑍 < 𝑍𝛼/2 ) = 1 - 𝛼

 P(- 𝑍𝛼/2 < Z < 𝑍𝛼/2 ) = 1 - 𝛼

𝑋ത −𝜇
 P(- 𝑍𝛼/2 < 𝜎 < 𝑍𝛼/2 ) = 1 - 𝛼
𝑛

𝜎 𝜎
 P(𝜇 ∈ (𝑋ത − . 𝑍𝛼 , 𝑋ത + . 𝑍𝛼 ) = 1 - 𝛼
𝑛 2 𝑛 2

 This is the interval within which 𝜇 lies with probability = 1 - 𝛼


 We can compute 𝑋ത from the sample. n is the sample size. 𝑍𝛼/2 can be computed from
the standard normal table. So if we know 𝜎, we are done !!
We may not know 𝝈 !!
The t - statistic

𝑛−1 .𝑠 2
 ~ χ2(𝑛−1)
𝜎2
The t - statistic

𝑛−1 .𝑠 2
 ~ χ2(𝑛−1)
𝜎2

𝑋ത −𝜇
 Define a statistic t = 𝑠
𝑛
The t - statistic

𝑛−1 .𝑠 2
 ~ χ2(𝑛−1)
𝜎2

ഥ −𝜇
𝑋
𝜎
𝑋ത −𝜇 𝑛
 Define a statistic t = 𝑠 =
𝑠/𝜎
𝑛
The t - statistic

𝑛−1 .𝑠 2
 ~ χ2(𝑛−1)
𝜎2

ഥ −𝜇
𝑋 ഥ −𝜇
𝑋
𝜎 𝜎
𝑋ത −𝜇 𝑛 𝑛 𝑍
 Define a statistic t = 𝑠 = = = ~ 𝑡𝑛−1
𝑠/𝜎 𝑛−1 .𝑠2 1
𝑛 .𝑛−1 χ2(𝑛−1)
𝜎2 .
𝑛−1
The t - statistic

𝑛−1 .𝑠 2
 ~ χ2(𝑛−1)
𝜎2

ഥ −𝜇
𝑋 ഥ −𝜇
𝑋
𝜎 𝜎
𝑋ത −𝜇 𝑛 𝑛 𝑍
 Define a statistic t = 𝑠 = = = ~ 𝑡𝑛−1
𝑠/𝜎 𝑛−1 .𝑠2 1
𝑛 .𝑛−1 χ2(𝑛−1)
𝜎2 .
𝑛−1

i.e the student’s t – distribution with n-1 degrees of freedom.


Student’s t - distribution
The Interval
(𝛼/2)
 P( 𝑡 < 𝑡𝑛−1 )=1-𝛼
The Interval
(𝛼/2)
 P( 𝑡 < 𝑡𝑛−1 )=1-𝛼

(𝛼/2) (𝛼/2)
 P(-𝑡𝑛−1 < t< 𝑡𝑛−1 ) =1-𝛼

(𝛼/2) 𝑋ത −𝜇 (𝛼/2)
 P(-𝑡𝑛−1 < 𝑠 <𝑡𝑛−1 ) =1-𝛼
𝑛
The Interval
(𝛼/2)
 P( 𝑡 < 𝑡𝑛−1 )=1-𝛼

(𝛼/2) (𝛼/2)
 P(-𝑡𝑛−1 < t < 𝑡𝑛−1 ) = 1 - 𝛼

(𝛼/2) 𝑋ത −𝜇 (𝛼/2)
 P(-𝑡𝑛−1 < 𝑠 <𝑡𝑛−1 ) =1-𝛼
𝑛

𝑠 (𝛼/2) ത 𝑠 (𝛼/2)
 P(𝜇 ∈ (𝑋ത − . 𝑡𝑛−1 , 𝑋 + . 𝑡𝑛−1 ) =1-𝛼
𝑛 𝑛
The Interval
(𝛼/2)
 P( 𝑡 < 𝑡𝑛−1 ) = 1 - 𝛼

(𝛼/2) (𝛼/2)
 P(-𝑡𝑛−1 < t < 𝑡𝑛−1 ) = 1 - 𝛼

(𝛼/2) 𝑋ത −𝜇 (𝛼/2)
 P(-𝑡𝑛−1 < 𝑠 <𝑡𝑛−1 ) = 1 - 𝛼
𝑛

𝑠 (𝛼/2) 𝑠 (𝛼/2)
 P(𝜇 ∈ (𝑋ത − . 𝑡𝑛−1 , 𝑋ത + . 𝑡𝑛−1 ) = 1 - 𝛼
𝑛 𝑛

 This is the interval within which 𝜇 lies with probability = (1 - 𝛼)

(𝛼/2)
 We can compute 𝑋ത from the sample. n is the sample size.𝑡𝑛−1 can be computed from the t - distribution
table.
Interval estimation of Variance

𝑛−1 .𝑠 2
 ~ χ2(𝑛−1)
𝜎2
𝛼 𝛼
2
(1− ) 𝑛−1 .𝑠2 ( )
2 2
 P( χ 𝑛−12 < < χ 𝑛−1 ) = 1-𝛼
𝜎2

2 𝑛−1 .𝑠2 𝑛−1 .𝑠2


 P (𝜎 ∈ ( 𝛼
(1− 2 )
, 𝛼 )=1-𝛼
χ 2
𝑛−1 χ 2 2
𝑛−1

 Given a sample we can compute 𝑠 2 and we know n (sample size). We can


𝛼 𝛼
( ) (1− )
2 2 2
compute χ 𝑛−1 & χ 𝑛−12 from the chi – square distribution table.

 Thus we have found the interval such that 𝜎 2 lies in that interval with probability
(1- 𝛼 )

You might also like