Lecture 5
Lecture 5
Z τ Z ∞
P [error] = 0.5 f1 (X) dX + 0.5 f0 (X) dX
−∞ τ
τ − µ1 τ − µ0
µ ¶ µ ¶
= 0.5Φ + 0.5(1 − Φ )
σ σ
• As we vary τ we trade one kind of error with another.
In Bayes classifier, the loss function determines the
‘exchange rate’.
Then we have
e0 = P [X ≤τ |X ∈ c-1] (a miss)
e1 = P [X >τ |X ∈ c-0] (false alarm)
1 − e0 = P [X >τ |X ∈ c-1] (correct detection)
1 − e1 = P [X ≤τ |X ∈ c-0] (correct rejection)
Then we have
e0 = P [X ≤τ |X ∈ c-1] (a miss)
e1 = P [X >τ |X ∈ c-0] (false alarm)
1 − e0 = P [X >τ |X ∈ c-1] (correct detection)
1 − e1 = P [X ≤τ |X ∈ c-0] (correct rejection)
• For fixed class conditional densities, if we vary τ the
point (e1 , 1 − e0 ) moves on a smooth curve in ℜ2 .
• This is traditionally called the ROC curve. (Choice of
coordinates is arbitrary)
Eθ [θ̂] = θ
Eθ [θ̂] = θ
• θ̂ is an unbiased estimator, if for every density in the
class of densities we are interested in (i.e., every
value of the parameter in the parameter space),
expected value of the estimator is the true parameter
value.
Bθ (θ̂) = Eθ [θ̂] − θ
Bθ (θ̂) = Eθ [θ̂] − θ
• For unbiased estimators the variance is the mean
square error (because bias is zero).
PR NPTEL course – p.60/127
• Proof:
σ2
Vθ (θ̂n ) =
n
σ2
Vθ (θ̂n ) =
n
For θ̂n′ = 0.5(x1 + x2 ),
2
σ
Vθ (θ̂n′ ) =
2
σ2
Vθ (θ̂n ) =
n
For θ̂n′ = 0.5(x1 + x2 ),
2
σ
Vθ (θ̂n′ ) =
2
• Hence θ̂ is better than θ̂′ PR NPTEL course – p.69/127
• So, unbiased estimators with low mean square error
are good.
2θ X
E[ (xi − θ)]
(n + 1) 2
2θ X
E[ (xi − θ)]
(n + 1) 2
n 2 1 2
= σ + θ
(n + 1)2 (n + 1) 2
1
Pn
where x̄ = n j=1 xj is the sample mean.
1
Pn
where x̄ = n j=1 xj is the sample mean.
• The loglikelihood is given by