Chapter 6 Bayesianlearning
Chapter 6 Bayesianlearning
Bayesian Learning
Introduction
– Bayesian learning algorithms calculate explicit
probabilities for hypotheses
– Practical approach to certain learning problems
– Provide useful perspective for understanding
learning algorithms
Real Life applications
– If h is consistent with D:
P(h|D) = 1. (1/|H|) / (|VSH,D| / |H|) = 1/ |VSH,D|
Consistent Learners
– Learning algorithms whose outputs are
hypotheses that commit zero errors over the
training examples (consistent hypotheses)
Then
Cross Entropy
- i=1,m di log[h(xi)] + [1-di] log[1-h(xi)]
MDL Principle:
Choose hMDL = argminhH {LC1(h) + LC2(D|h)}
Representation
• A Bayesian network represents the joint probability
distribution of a set of variables
• Each variable is represented by a node
Example:
P(Campfire=True|Storm=True,BusTourGroup=True)=0.4
Inference
• We wish to infer the probability distribution for
some variable given observed values for (a subset
of) the other variables
• Exact (and sometimes approximate) inference of
probabilities for an arbitrary BN is NP-hard
• There are numerous methods for probabilistic
inference in BN (for instance, Monte Carlo), which
have been shown to be useful in many cases