Lecture 14
Lecture 14
HHTHT
HHHHH
What process produced these sequences?
What are the plausible alternative hypotheses,
and their priors?
Comparing two simple hypotheses
P( H ) P( D | H )
P( H | D) =
P( D)
• With two hypotheses, use odds form
Comparing two simple hypotheses
P( H1 | D) P( D | H1 ) P( H1 )
= ´
P( H 2 | D) P( D | H 2 ) P( H 2 )
D: HHTHT
H1, H2: “fair coin”, “always heads”
P(D|H1) = 1/25 P(H1) = ?
P(D|H2) = 0 P(H2) = ?
Comparing two simple hypotheses
P( H1 | D) P( D | H1 ) P( H1 )
= ´
P( H 2 | D) P( D | H 2 ) P( H 2 )
D: HHTHT
H1, H2: “fair coin”, “always heads”
P(D|H1) = 1/25 P(H1) = 999/1000
P(D|H2) = 0 P(H2) = 1/1000
P( H1 | D) 1 32 999
= ´ = infinity
P( H 2 | D) 0 1
Comparing two simple hypotheses
P( H1 | D) P( D | H1 ) P( H1 )
= ´
P( H 2 | D) P( D | H 2 ) P( H 2 )
D: HHHHH
H1, H2: “fair coin”, “always heads”
P(D|H1) = 1/25 P(H1) = 999/1000
P(D|H2) = 1 P(H2) = 1/1000
P( H1 | D) 1 999
= ´ » 30
P( H 2 | D) 32 1
Comparing two simple hypotheses
P( H1 | D) P( D | H1 ) P( H1 )
= ´
P( H 2 | D) P( D | H 2 ) P( H 2 )
D: HHHHHHHHHH
H1, H2: “fair coin”, “always heads”
P(D|H1) = 1/210 P(H1) = 999/1000
P(D|H2) = 1 P(H2) = 1/1000
P ( H1 | D ) 1 999
= ´ »1
P( H 2 | D) 1024 1
Comparing two simple hypotheses
P( H1 | D) P( D | H1 ) P( H1 )
= ´
P( H 2 | D) P( D | H 2 ) P( H 2 )
D: HHHHHHHHHHHHHHH
H1, H2: “fair coin”, “always heads”
P(D|H1) = 1/215 P(H1) = 999/1000
P(D|H2) = 1 P(H2) = 1/1000
P ( H1 | D ) 1 999
= ´ » 1~0.03
P( H 2 | D) 1024 1
~32000
An alternative analysis
P( H1 | D) P( D | H1 ) P( H1 )
= ´
P( H 2 | D) P( D | H 2 ) P( H 2 )
D: HHTHT
H1, H2: “fair coin”, “coin that always comes up HHTHT”
P(D|H1) = 1/25 P(H1) = 999/1000
P(D|H2) = 1 P(H2) = 1/1000
P( H1 | D) 1 999
= ´ » 30
P( H 2 | D) 32 1
An alternative analysis
P( H1 | D) P( D | H1 ) P( H1 )
= ´
P( H 2 | D) P( D | H 2 ) P( H 2 )
D: HHHHH
H1, H2: “fair coin”, “coin that always comes up HHTHT”
P(D|H1) = 1/25 P(H1) = 999/1000
P(D|H2) = 0 P(H2) = 1/1000
P( H1 | D) 1 32 999
= ´ = infinity
P( H 2 | D) 0 1
The role of priors
HHTHT
How about ...
HHTHTHHTHT
How about ...
HHTHTHHTHTHHTHT
How about ...
HHTHTHHTHTHHTHTHHTHT
How about ...
HHTHTHHTHTHHTHTHHTHTHHTHT
How about ...
HHTHTHHTHTHHTHTHHTHTHHTHT
• Your task:
– Observe one or more positive (“yes”) examples.
– Judge whether other numbers are “yes” or “no”.
Bayesian concept learning
“horse” “horse” “horse”
“tufa”
“tufa”
“tufa”
The number game
Examples of Generalization
“yes” numbers judgments (N = 20)
60
Diffuse similarity
The number game
Examples of Generalization
“yes” numbers judgments (N = 20)
60
Diffuse similarity
60 80 10 30 Rule:
“multiples of 10”
The number game
Examples of Generalization
“yes” numbers judgments (N = 20)
60
Diffuse similarity
60 80 10 30 Rule:
“multiples of 10”
60 52 57 55 Focused similarity:
numbers near 50-60
The number game
Examples of Generalization
“yes” numbers judgments (N = 20)
16
Diffuse similarity
16 8 2 64 Rule:
“powers of 2”
16 23 19 20 Focused similarity:
numbers near 20
The number game
60
Diffuse similarity
60 80 10 30 Rule:
“multiples of 10”
60 52 57 55 Focused similarity:
numbers near 50-60
åx
p( X = x | Y = y ) = 1
Prior: p(h)
• Choice of hypothesis space embodies a strong prior:
effectively, p(h) ~ 0 for many logically possible but
conceptually unnatural hypotheses.
• Do we need this? Why not allow all logically possible
hypotheses, with uniform priors, and let the data sort
them out (via the likelihood)?
Prior: p(h)
• Choice of hypothesis space embodies a strong prior:
effectively, p(h) ~ 0 for many logically possible but
conceptually unnatural hypotheses.
• Prevents overfitting by highly specific but unnatural
hypotheses, e.g. “multiples of 10 except 50 and 70”.