I2ml3e Chap19
I2ml3e Chap19
INTRODUCTION
TO
MACHINE
LEARNING
3RD EDITION
ETHEM ALPAYDIN
© The MIT Press, 2014
[email protected]
http://www.cmpe.boun.edu.tr/~ethem/i2ml3e
CHAPTER 19:
Questions:
Assessment of the expected error of a learning algorithm: Is
the error rate of 1-NN less than 2%?
Comparing the expected errors of two algorithms: Is k-NN
more accurate than MLP ?
Training/validation/test sets
Resampling methods: K-fold cross-validation
Algorithm Preference
4
Criteria (Application-dependent):
Misclassification error, or risk (loss functions)
Training time/space complexity
Interpretability
Easy programmability
Cost-sensitive learning
Factors and Response
5
X = { xt }t where xt ~ N ( μ, σ2)
m ~ N ( μ, σ2/N)
m
N ~Z
m
P 1.96 N 1.96 0.95
P m 1.96 m 1.96 0.95
N N
P m z / 2 m z / 2 1 100(1- α) percent
N N confidence interval
100(1- α) percent one-sided
m confidence interval
P N 1.64 0.95
P m 1.64 0.95
N
P m z 1
N
j 1 j
N=100, e=20
1- α
Normal Approximation to the Binomial
20
1- α
Paired t Test
21
N
With m and s2 average and var of pi , we accept p0 or
less error if
K m p0
~ tK 1
S
is less than tα,K-1
Comparing Classifiers: H0:μ0=μ1 vs.
22
H1:μ0≠μ1
Single training/validation set: McNemar’s Test
~ X12
e01 e10
i 1 pi
K K
i 1 ip m 2
m s2
K K 1
K m 0 K m
~ t K 1 Accept if in t / 2 ,K 1 ,t / 2 ,K 1
s s
5×2 cv Paired t Test
24
p11
~ t5
5 2
s /5
i 1 i
p j 2
5 2
i 1 j 1 i
~ F10,5
2 s
5 2
i 1 i
H0 : 1 2 L
Errors of L algorithms on K folds
X ij ~ N j , 2 , j 1,..., L, i 1,..., K
~ N , 2 / K
K X ij
mj
i 1 K
j 1 m j m
L
j m 2
m S2 j
L L 1
Thus an estimator of 2 is K S 2 , namely,
ˆ 2 K
L m j m 2
j 1 L 1
m m 2
~ X L21 SSb K m j m 2
j
j
2 /K j
X
K
m 2 L S 2j X m 2
ˆ 2
i 1 ij j ij j
S 2j
K 1 j 1 L j i LK 1
SSw X ij m j 2
j i
S 2j SSw
K 1 ~X 2
K 1 ~ X L2K 1
2 2
SSb / 2 SSw / 2 SSb /L 1
/ ~ FL1,L K 1
L 1 LK 1 SSw /LK 1
H0 : 1 2 L if F ,L1,L K 1
28
ANOVA table
29
Paired differences: