0% found this document useful (0 votes)
78 views4 pages

18 05 Lec37

1) The document provides a summary of solutions to practice problems for a statistics final exam. It includes the solutions to 8 practice problems covering topics like maximum likelihood estimation, Bayesian inference, hypothesis testing, and chi-square tests. 2) The problems cover a range of statistical techniques including finding MLEs, calculating posterior distributions, comparing hypothesis tests, and determining appropriate decision rules. 3) Step-by-step solutions are shown for each problem to demonstrate how to apply statistical methods to different scenarios like estimating parameters of distributions or comparing means of populations.

Uploaded by

anshul77
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
78 views4 pages

18 05 Lec37

1) The document provides a summary of solutions to practice problems for a statistics final exam. It includes the solutions to 8 practice problems covering topics like maximum likelihood estimation, Bayesian inference, hypothesis testing, and chi-square tests. 2) The problems cover a range of statistical techniques including finding MLEs, calculating posterior distributions, comparing hypothesis tests, and determining appropriate decision rules. 3) Step-by-step solutions are shown for each problem to demonstrate how to apply statistical methods to different scenarios like estimating parameters of distributions or comparing means of populations.

Uploaded by

anshul77
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

18.

05 Lecture 37
May 17, 2005

Final Exam Review - solutions to practice final.

1. f (x|v) = {uv −u xu−1 e− ( xv )u for x ∼ 0; 0 otherwise.}


Find the MLE of v.
↔ Maximize v in the likelihood function = joint p.d.f.
 u
ξ(v) = un v −nu ( xi )u−1 e− (xi /v)
P

� �
log ξ(v) = n log u − nu log v + (u − 1) log( xi ) − ( xvi )u
Maximize with respect to v.

�(log ξ(v)) −nu � u � −u −nu �


= − xi v = +u xui v −(u+1) = 0
�v v �v v
� � u
v u xui xi
n = ( u+1 ) =
u v vu
1� u
vu = xi
n
1� u 1
v=( xi ) u ↔ MLE
n
2. Xi , ..., Xn ≈ U [0, χ]
f (x|χ) = 1β I(0 ← x ← χ)
Prior:
192
f (χ) = I(χ ∼ 4)
χ4
Data: X1 = 5, X2 = 3, X3 = 8
Posterior: f (χ|x1 , ..., xn ) ≈ f (x1 , ..., xn |χ)f (χ)
f (x1 , ..., xn |χ) = β1n I(0 ← all x’s ← χ) = β1n I(max(X1 , ..., Xn ) ← χ)
1 1
f (χ|x1 , ..., xn ) ≈ βn+4 I(χ ∼ 4)I(max(x1 , ..., xn ) ← χ) ≈ βn+4 I(χ ∼ 8)

Find constant so it integrates to 1.


� ∗ � ∗
1 −−−↔ cχ−6 ∗ c
1= n+4
dχ n = 3 1 = cχ−7 dχ ↔ |8 = 8−6 = 1
8 χ 8 −6 6

c = 6 × 86

3. Two observations (X1 , X2 ) from f(x)

H1 : f (x) = 1/2, I(0 ← x ← 2)

H2 : f (x) = {1/2, 0 ← x ← 1, 2/3, 1 < x


← 2}
H3 : f (x) = {3/4, 0 ← x ← 1, 1/4, 1 < x ← 2}
β minimizes ∂1 (β) + 2∂2 (β) + 2∂3 (β)
= P(β ∈= Hi |Hi )
∂i (β) �
Find �(i)∂i , Decision rule picks �(i)fi (x1 , ..., xn ) ↔ max for each region.

115
�(i)fi (x1 )fi (x2 ) H1 H2 H3
both x1 , x2 ⊂ [0, 1] (1)(1/2)(1/2) = 1/4 (2)(1/3)(1/3) = 1/3 (2)(3/4)(3/4) = 9/8
point in [0, 1], [1, 2] (1)(1/2)(1/2) = 1/4 (2)(1/3)(2/3) = 4/9 (2)(3/4)(1/4) = 3/8
both in [1, 2] (1)(1/2)(1/2) = 1/4 (2)(1/3)(2/3) = 8/9 (2)(1/4)(1/4) = 1/8

Decision Rule:

β = {H1 : never pick , H2 : both in [1, 2], one in [0, 1], [1, 2] , H3 : both in [0, 1]}

If two hypotheses:

f1 (x) �(2)
>
f2 (x) �(1)
Choose H1 , ← H2

4.
1 −1 2
f (x|µ) = { ∩ e 2 (ln x−µ) for x ∼ 0, 0 for x < 0}
x 2ψ
If X has this distribution, find distribution of ln X.
Y = ln X � ey
c.d.f. of Y: P(Y ← y) = (ln x ← y) = P(x ← ey ) = 0 f (x)dx
However, you don’t need to integrate.

p.d.f. of Y, f (y) = �y P(Y ← y) = f (ey ) × ey
1 −1 y
2 (ln e −µ)
2 1 −1 2
= ∩ e × ey = ∩ e 2 (y−µ) ≈ N (µ, 1)
ey 2ψ 2ψ

5. n = 10, H1 : µ = −1, H2 : µ = 1
∂ = 0.05 ↔ ∂1 (β) = P1 ( ff21 < c)

f1 (↔
−x)
β = {H1 : ↔
− ∼ c, H2 if less than}
f2 ( x )
1
f1 (↔
− 1 2
e− 2 (ln xi +1)
P
x)= � ∩
xi ( 2ψ) n

1
f2 (↔
− 1 2
e− 2 (ln xi −1)
P
x)= � ∩
xi ( 2ψ) n

f1 (x) �
= e−2 ln xi ∼ c ∝ ln xi ← c∅
P

f2 (x)
� �
β = {H1 : ln xi ← c = −4.81, H2 : ln xi > c = −4.81}
� �
0.05 = P1 ( ln xi ∼ c) = P1 ( N (−1, 1) ∼ c) = P1 ( x≥i −nµ
P
n
∼ c−nµ


) = P1 (Z ∼ c−nµ


)
c − nµ
∩ = 1.64, c = −4.81

� �
Power = 1 - type 2 error = 1 − P2 (β ∈= H2 ) = 1 − P2 ( ln xi < c) = 1 − P2 ( N (1, 1) < c)

xi − n(1) −4.81 − 10
= 1 − P2 ( ∩ < ∩ )�1
n 10

116
6. H1 : p1 = 2β , p2 = β3 , p3 = 1 − 56β , χ ⊂ [0, 1]

Step 1) Find MLE χ�

� �
5β �
Step 2) p�1 = β2 , p2� = β3 , p�

3 =1− 6
Step 3) Calculate T statistic.
r
� (Ni − np� )2 i
T = ≈ α2r−s−1=3−1−1=1
i=1
npi

χ χ 5χ
ξ(χ) = ( )N1 ( )N2 (1 − )N3
2 3 6

log ξ(χ) = (N1 + N2 ) log(χ) + N3 log(1 − ) − N1 log(2)N2 log(3) ↔ max χ
6
� N1 + N2 −5/6
= + N3
�χ χ 1 − 56β
5 5
N1 + N2 − (N1 + N2 )χ − N3 χ = 0
6 6
6 N1 + N 2 23
solve for χ ↔ χ � = ( )=
5 n 25
Compute statistic, T = 0.586
β = {H1 : T ← 3.841, H2 : T > 3.841}
Accept H1

7. n = 17, x = 3.2, x = 0.09


From N (µ, θ 2 )
H1 : µ ← 3
H2 : µ > 3
at ∂ = 0.05
x − µ0 3.2 − 3
T =� ≈ tn−1 = � = 2.67
1 2 − (x)2 ) 1
n−1 (x 16 (0.09)
Choose decision rule from the chi-square table with 17-1 degrees of freedom:

β : {H1 : T < 1.746, H2 : T > 1.746}


H1 is rejected.

8. Calculate t statistic:

� (Nij − N+j Ni+ 2


n )
T = N+j Ni+
= 12.1
i,j n

117
α2(a−1)(b−1) = α23×2 = α62
at 0.05, c = 12.59
β : {H1 : T ← 12.59, H2 : T > 12.59}

Accept H1 . But note if confidence level changes, ∂ ↔ 0.10, c = 10.64, would reject H1

9.

� x I(0 ← x ← 2)

f (x) = 1/2,
F (x) = −∗ f (t)dt = x/2, x ← 2

x:
0.02 0.18 0.20 ...
F(x):
0.01 0.09 0.10 ...
F(x) before:
0 0.1 0.2 ...
F(x) after:
0.1 0.2 0.3 ...
diff F(x) before: 0.01 0.01 0.1 ...
diff F(x) after: 0.09 0.11 0.2 ...

�n = |F (x) − Fn (x)|
max �n = 0.295
c for ∂∩= 0.05 is 1.35
Dn = 10(0.295) = 0.932872
β = {H1 : 0.932872 ← 1.35, H2 : 0.932872 > 1.35}
Accept H1

** End of Lecture 37

*** End of 18.05 Spring 2005 Lecture Notes.

118

You might also like