0% found this document useful (0 votes)
16 views

Pert20 - Knowledge in Learning

Uploaded by

82gfmcz5fs
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
16 views

Pert20 - Knowledge in Learning

Uploaded by

82gfmcz5fs
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 39

Course : Artificial Intelligence (COMP6065)

Non-official Slides

Knowledge in Learning

Session 20

Revised by Williem, S. Kom., Ph.D.


1
Learning Outcomes
At the end of this session, students will be able to:

• LO 5 : Apply various techniques to an agent when acting under


certainty

• LO 6 : Apply how to process natural language and other


perceptual signs in order that an agent can interact
intelligently with the world

2
Outline
1. A Logical Formulation of Learning

2. Knowledge in Learning

3. Explanation Based Learning

4. Learning Using Relevance Information

5. Inductive Logic Programming

6. Summary

3
A Logical Formulation of Learning
• In this session, the hypotheses are represented by a set of
logical sentences

– Allows for incremental construction of hypotheses

– Allows for prior knowledge

• Example

– The restaurant learning problem can be described as

4
A Logical Formulation of Learning
• Study case

– Change the example in logical representation

5
A Logical Formulation of Learning
• Thus, a decision tree can be
interpreted in logical
representation

6
A Logical Formulation of Learning
• As the example arrive, hypotheses that are not consistent with
the examples can be ruled out

• What is the meaning when an example is consistent?

– False negative: When the hypothesis says that it is


negative, but in fact it is positive

– False positive: When the hypothesis says that it is positive,


but in fact it is negative

7
A Logical Formulation of Learning

Hypothesis False negative False positive

8
A Logical Formulation of Learning
• How to find the logically consistent hypothesis?

– Current-best-hypothesis search

• To maintain a single hypothesis, and to adjust it as new


examples arrive

– Least-commitment search

• Candidate elimination algorithm has an incremental


property, where one never has to god back and
reexamine the old examples

9
Current-best-hypothesis Search
• Suppose we have some hypothesis hr

– As long as each new example is consistent, we do nothing

– If there is a false negative example?

• Generalization: We extend the hypothesis region

– If there is a false positive example?

• Specialization: We reduce the hypothesis region

10
Current-best-hypothesis Search
(extensions of predictor Hr)

Initial False False


hypothesis negative positive
a generalization a specialization
Generalization e.g. via dropping conditions
Alternate(x)Patrons(x, Some)  Patrons(x, Some)

Specialization e.g. via adding conditions or via removing disjuncts


Alternate(x)  Alternate(x)Patrons(x, Some)
11
Least-commitment search
• Assuming the original hypothesis space is the right answer, the
reduced disjunction must still contain the right answer because
only incorrect hypotheses have been removed

• The set of hypotheses remaining are called version space

• The learning algorithm is called version space learning or


candidate elimination algorithm

• The problem is how to deal with all remaining hypotheses?

– HUGE!

12
Least-commitment search
• The problem is similar to how we represent all numbers between
1 and 2

– Using boundary!!

• The boundary in the version space is generalization/specialization

– A most general boundary (G-set)

– A most specific boundary (S-set)

• Everything in between is guaranteed to be consistent with the


examples

13
Version Space

14
Version Space Learning
• Algorithm:

– False positive for Si: Si is too general, and there are no


consistent specializations for Si, so throw Si out of S-Set

– False negative for Si: Si is too specific, so replace it with


all its immediate generalizations.

– False positive for Gi: Gi is too general, so replace it with


all its immediate specializations.

– False negative for Gi: Gi is too specific, but there are no


consistent generalizations of Gi, so throw Gi out of G-Set
15
Version Space Learning
• We continue for each example until:

– We have exactly one hypothesis left

– The version space collapses – either S-set or G-set empty

– We run out of examples and have several hypotheses


remaining

16
Candidate Elimination Algorithm
• Given : A representation language and a set of positive
and negative examples expressed in that language.

• Compute : A concept description that is consistent with all


the positive examples and none of the negative examples.

1. Initialize G to contain one element : the null description (all


features are variables)

2. Initialize S to contain one element : the first positive


example

17
Candidate Elimination Algorithm
3. Accept a new training example.

• If it is a positive example

– Remove from G any descriptions that do not cover the


example

– Update the S set to contain the most specific set of


descriptions in the version space that cover the example
and the current element of the S set

18
Candidate Elimination Algorithm
• If it is a negative example
– Remove from S any descriptions that cover the example
– Update the G set to contain the most general set of
descriptions in the version space that do not cover the
example
• If S and G are both singleton sets, then if they are identical,
output their value and halt.
• If they are both singleton sets but they are different, then the
training cases were inconsistent. Output this result and halt.
• Otherwise, go to step 3.
19
Candidate Elimination
Algorithm
G : { obj(X,Y,Z) }, S: { } Pos : obj(small,red,ball)

Initialize G to be the most general concept in the space, and


Initialize S to the first pos. training instance

G : { obj(X,Y,Z) }, S :{ obj(small, red, ball) }

20
Candidate Elimination
Algorithm
G : { obj(X,Y,Z) }, S: { } Pos : obj(small,red,ball)
Initialize G to be the most general concept in the space, and
Initialize S to the first pos. training instance
G : { obj(X,Y,Z) }, S :{ obj(small, red, ball) } Neg : obj(small,blue,ball)

Delete all members of S that match n


For each g that match n, replace g with its most general specialization
that do not match n

G : { obj(X,red,Z) }, S :{ obj(small, red, ball) }

21
Candidate Elimination
Algorithm
G : { obj(X,Y,Z) }, S: { } Pos : obj(small,red,ball)
Initialize G to be the most general concept in the space, and
Initialize S to the first pos. training instance
G : { obj(X,Y,Z) }, S :{ obj(small, red, ball) } Neg : obj(small,blue,ball)

Delete all members of S that match n


For each g that match n, replace g with its most general specialization
that do not match n

G : { obj(X,red,Z) }, S :{ obj(small, red, ball) } Pos : obj(large,red,ball)

Delete all members of G that fail to match p


For every s, if s does not match p, replace s with its most specific
generalization that match p

G : { obj(X,red,Z) }, S :{ obj(X, red, ball) }

22
Candidate Elimination
Algorithm
G : { obj(X,Y,Z) }, S: { } Pos : obj(small,red,ball)
Initialize G to be the most general concept in the space, and
Initialize S to the first pos. training instance
G : { obj(X,Y,Z) }, S :{ obj(small, red, ball) } Neg : obj(small,blue,ball)

Delete all members of S that match n


For each g that match n, replace g with its most general specialization
that do not match n

G : { obj(X,red,Z) }, S :{ obj(small, red, ball) } Pos : obj(large,red,ball)

Delete all members of G that fail to match p


For every s, if s does not match p, replace s with its most specific
generalization that match p

G : { obj(X,red,Z) }, S :{ obj(X, red, ball) } Neg: obj(large,red,cube)

Delete all members of S that match n


For each g that match n, replace g with its most general specialization
that do not match n
23
G : { obj(X,red,ball) }, S :{ obj(X, red, ball) }
Candidate Elimination
Algorithm
Positive and Negative Examples of the Concept “Japanese economy car”

origin : Japan origin : Japan origin : Japan


mfr : Honda mfr : Toyota mfr : Toyota
color : Blue color : Green color : Blue
decade : 1980 decade : 1970 decade : 1990
type : Economy type : Sports type : Economy
(+) (-) (+)
origin : USA origin : Japan
mfr : Chrysler mfr : Honda
color : Red color : White
decade : 1980 decade : 1980
type : Economy type : Economy
(-) (+)

24
Candidate Elimination
Algorithm
A Version Space Example
• G = {(x1,x2,x3,x4,x5)}
S = {(Japan, Honda, Blue, 1980, Economy)}

• G = {(x1, Honda,x3,x4,x5) ,(x1,x2,Blue,x4,x5),


(x1,x2,x3,1980,x5),(x1,x2,x3,x4,Economy)}
• S = {(Japan, Honda, Blue, 1980, Economy)}

• G = {(x1,x2, Blue, x4,x5), (x1,x2,x3,x4,Economy)}


• S = {(Japan, x2, Blue, x4, Economy)}

• G = {(x1,x2, Blue, x4,x5), (x1,x2, Blue,x4,Economy),


(Japan,x2,x3,x4, Economy)}
• S = {(Japan,x2, Blue,x4, Economy)}

• G = {(Japan,x2,x3,x4,Economy)}
• S = {(Japan,x2,x3,x4,Economy)} 25
Knowledge in Learning
• Let Descriptions denote the conjunction of all the example
descriptions in the training set, and let Classifications denote
the conjunction of all the example classification.

• Then a Hypothesis that "explains the observations" must


satisfy the following property (recall that |= means "logically
entails"):

Hypothesis ۸ Descriptions |= Classifications

26
Knowledge in Learning

27
Knowledge in Learning
• There are three kinds of learning using prior knowledge:

– Explanation-based learning (EBL)

– Relevance-based learning (RBL)

– Knowledge-based inductive learning (KBIL)

• Inductive logic programming (ILP)

28
Explanation Based Learning
• A method for extracting general rules from individual
observations

• As example, consider the problem of differentiation!

– X2 = 2X

• Compare the student with and without calculus


knowledge

• Which one is faster?

29
Explanation Based Learning
• The technique of memoization has long been used in
computer science to speed up programs by saving the results
of computation.

• The basic idea of memo functions is to accumulate a database


of input—output pairs; when the function is called, it first
checks the database to see whether it can avoid solving the
problem from scratch.

• Explanation –based learning takes this a good deal further, by


creating general rules that cover an entire class of cases.
30
Learning Using Relevance Information
• A method for extracting general rules from the relevance
between prior knowledge and observations to explain the
observations

• RBL does not produce hypotheses that go beyond the logical


content of the background knowledge

– Cannot create a new knowledge from scratch


31
Learning Using Relevance Information
• Functional dependencies or determinations occur so
commonly in certain kinds of applications

• Determinations specify a sufficient basis vocabulary from which


to construct hypotheses concerning the target predicate

32
Inductive Logic Programming
• A method for extracting general rules from the background
knowledge and new hypothesis to explain the examples

• Inductive logic programming (ILP) combines inductive


methods with the power of first-order representations,
concentrating in particular on the representation of
hypotheses as logic programs.

33
Inductive Logic Programming
• It has gained popularity for three reasons :
1. ILP offers a rigorous approach to the general knowledge-
based inductive learning problem.

2. ILP offers complete algorithms for inducing general, first-


order theories from examples, which can therefore learn
successfully in domains where attribute-based algorithms
are hard to apply.

3. Inductive logic programming produces hypotheses that


are (relatively) easy for humans to read.
34
Inductive Logic Programming
• Example

– parent_of(charles,george)

– parent_of(george,diana)

– parent_of(bob,harry)

– parent_of(harry,elizabeth)

– grandparent_of(X,Y) : parent_of(X,Z), parent_of(Z,Y)


• Question query Answer

• grandparent_of(X,Y)? • grandparent_of(charles,diana)
• grandparent_of(bob,elizabeth)
35
Summary
• The use of prior knowledge in learning leads to a picture of
cumulative learning

• Prior knowledge helps learning by

– Eliminating otherwise consistent hypotheses

– Filling in the explanation examples, allowing shorter


hypotheses

36
References
• Stuart Russell, Peter Norvig. 2010. Artificial Intelligence : A
Modern Approach. Pearson Education. New Jersey.
ISBN:9780132071482

• http://aima.cs.berkeley.edu

37
Version Space

Case Study

38
Candidate Elimination
Algorithm
Exercises : How About The Concept of Elephant ?

39

You might also like