ML 3
ML 3
Tuesday (10am-11am)
Wednesday (10am-11am & 3pm-4pm)
Friday (9am-10am, 11am-12am, 2pm-3pm)
Dr. Srinivasa L. Chakravarthy
&
Smt. Jyotsna Rani Thota
Department of CSE
GITAM Institute of Technology (GIT)
Visakhapatnam – 530045
Email: [email protected] & [email protected]
Department of CSE, GIT 1
20 August 2020
EID 403 and machine learning
Course objectives
20 August 2020 4
&
● Probably Learning
● Sample Complexity
● Finite and Infinite Hypothesis space
● Mistake Bound model of Learning
7
Introduction
This chapter introduces
The Goal in this chapter is to answer some questions with computational theory
such as-
3. Mistake bound- How many training examples will the learner misclassify
before converging to a successful hypothesis?
Probably learning an approximately correct hypothesis-
Error of a Hypothesis-
In order to identify , how closely the learner’s output hypothesis h approximates the
actual target concept c-
we need to know about true error of a hypothesis h w.r.t target concept c, i.e.,
Probably learning an approximately correct hypothesis(cont.)
Error of a Hypothesis-(cont.)
PAC Learnability-
2. Given that the training examples are drawn randomly, there is always a
nonzero probability that the training examples encountered by learner will be
misleading.
Probably learning an approximately correct hypothesis(cont.)
PAC Learnability-(cont.)
PAC Learnability-(cont.)
The growth in the required training examples with problem size is called as-
-“Sample Complexity” of learning problem.
The general bound on the sample complexity for a very broad class of learners,
called consistent learners. A learner is consistent if its output hypotheses perfectly
fits the training data.
Sample complexity for Finite hypothesis spaces-(cont.)
As we know that, version space VSH,D to be the set f all hypotheses h ∈ H that
correctly classify the training examples D.
So, every consistent learner outputs a hypothesis belonging to the version space.
Use theorem-
Sample complexity for Finite hypothesis spaces-(cont.)
Agnostic Learning
If H does not contains the target concept c, then a zero-error hypothesis cannot
always be found.
In this case, we might ask our learner to output the hypothesis from H that has the
minimum error over the training examples.
Sample complexity for PAC leaning grows as the logarithm of size of the
hypothesis space but, the drawbacks to characterizing sample complexity in
terms of |H| are-
Note- For any finite H, VC(H) ≤ log2|H|. To see this suppose VC(H)=d, then H will
require 2d distinct hypotheses to shatter d instances.
Instance1: 100
Instance2: 010
Instance3: 001
Sample complexity for Infinite hypothesis spaces-(cont.)
Example-VC dimension for linear decision surfaces(cont.)
The answer is- Here the no.of training examples grows logarithmically in
Here the no.of training examples grows log times linear in (1/ ∈)
Mistake Bound model
So far the discussion is about-
Generally, we might count the number of mistakes made before PAC learning the
target concept.
So, In the examples below, we consider the number of mistakes made before
learning the target concept exactly means converging to a hypothesis such that
h(x)=c(x).
Mistake Bound for FIND-S Algorithm
FIND-S begins with most specific hypothesis and generalizes the hypotheses
space in order to observe positive training examples.
In this case,if an instance is correctly classified then the version space will be
reduced to at most half its current size
Optimal Mistake Bounds