Ba Yes I An Reasoning
Ba Yes I An Reasoning
Knowledge
Representation and
Reasoning
Chapter 10.1-10.2, 10.6
Adapted from slides by
Tim Finin and
Marie desJardins.
Abduction
Abduction is a reasoning process that tries to form plausible
explanations for abnormal observations
Abduction is distinctly different from deduction and induction
Abduction is inherently uncertain
Abduction
Definition (Encyclopedia Britannica): reasoning that derives
an explanatory hypothesis from a given set of facts
The inference result is a hypothesis that, if true, could
explain the occurrence of the given facts
Examples
Dendral, an expert system to construct 3D structure of
chemical compounds
Fact: mass spectrometer data of the compound and its
chemical formula
KB: chemistry, esp. strength of different types of bounds
Reasoning: form a hypothetical 3D structure that satisfies the
chemical formula, and that would most likely produce the
given mass spectrum
Abduction: rule:
observation:
explanation:
Induction: case:
These balls are from the box
observation:
These balls are black
hypothesized rule: All ball in the box are black
Deduction reasons from causes to effects
Abduction reasons from effects to causes
Induction reasons from specific cases to general rules
Whenever
A then B
------------Possibly
A => B
Characteristics of abductive
reasoning
Characteristics of abductive
reasoning (cont.)
Characteristics of abductive
reasoning (cont.)
Reasoning is non-monotonic
That is, the plausibility of hypotheses can
increase/decrease as new facts are collected
In contrast, deductive inference is monotonic: it never
change a sentences truth value, once known
In abductive (and inductive) reasoning, some
hypotheses may be discarded, and new ones formed,
when new observations are made
Sources of uncertainty
Uncertain inputs
Missing data
Noisy data
Uncertain knowledge
Uncertain outputs
Bayesian reasoning
Probability theory
Bayesian inference
Use probability theory and information about independence
Reason diagnostically (from evidence (effects) to conclusions
(causes)) or causally (from causes to effects)
Bayesian networks
Compact representation of probability distribution over a set of
propositional random variables
Take advantage of independence relationships
Rule-based methods
Certainty factors (Mycin): propagate simple models of belief
through causal or diagnostic rules
Evidential reasoning
Dempster-Shafer theory: Bel(P) is a measure of the evidence for P;
Bel(P) is a measure of the evidence against P; together they define
a belief interval (lower and upper bounds on confidence)
Fuzzy reasoning
Fuzzy sets: How well does an object satisfy a vague property?
Fuzzy logic: How true is a logical statement?
Uncertainty tradeoffs
Bayesian networks: Nice theoretical properties combined
with efficient reasoning make BNs very popular; limited
expressiveness, knowledge engineering challenges may
limit uses
Nonmonotonic logic: Represent commonsense reasoning,
but can be computationally very expensive
Certainty factors: Not semantically well founded
Dempster-Shafer theory: Has nice formal properties, but
can be computationally expensive, and intervals tend to
grow towards [0,1] (not a very useful conclusion)
Fuzzy reasoning: Semantics are unclear (fuzzy!), but has
proved very useful for commercial applications
CS 63
Bayesian Reasoning
Chapter 13
Adapted from slides by
Tim Finin and
Marie desJardins.
Outline
Probability theory
Bayesian inference
From the joint distribution
Using independence/factoring
From sources of evidence
Sources of uncertainty
Uncertain inputs
Missing data
Noisy data
Uncertain knowledge
Uncertain outputs
0 P(a) 1
P(true) = 1 ; P(false) = 0
ab
Probability theory
Random variables
Domain
(Alarm=True Burglary=True
Earthquake=False) or equivalently
(alarm burglary earthquake)
P(Burglary) = 0.1
P(Alarm, Burglary) =
alarm
alarm
burglary
0.09
0.01
burglary
0.1
0.8
Product rule:
P(a b) = P(a | b) P(b)
Marginalizing:
P(B) = aP(B, a)
P(B) = aP(B | a) P(a)
(conditioning)
alarm
earthquake
earthquake
earthquake
earthquake
burglary
0.01
0.08
0.001
0.009
burglary
0.01
0.09
0.01
0.79
smart
p(smart
study prep) study study
study
study
prepared
0.432
0.16
0.084
0.008
prepared
0.048
0.16
0.036
0.072
Queries:
What is the prior probability of smart?
What is the prior probability of study?
What is the conditional probability of prepared, given
study and smart?
Independence
When two sets of propositions do not affect each others
probabilities, we call them independent, and can easily
compute their joint and conditional probability:
Independent (A, B) P(A B) = P(A) P(B), P(A | B) = P(A)
Exercise: Independence
smart
smart
p(smart
study prep) study study
study
study
prepared
0.432
0.16
0.084
0.008
prepared
0.048
0.16
0.036
0.072
Queries:
Is smart independent of study?
Is prepared independent of study?
Conditional independence
Absolute independence:
A and B are independent if and only if P(A B) = P(A) P(B);
equivalently, P(A) = P(A | B) and P(B) = P(B | A)
smart
p(smart
study prep) study study
study
study
prepared
0.432
0.16
0.084
0.008
prepared
0.048
0.16
0.036
0.072
Queries:
Is smart conditionally independent of prepared, given
study?
Is study conditionally independent of prepared, given
smart?
Bayess rule
Bayess rule is derived from the product rule:
P(Y | X) = P(X | Y) P(Y) / P(X)
Bayesian inference
In the setting of diagnostic/evidential reasoning
P(E j | H i )
E1
H i P(Hi )
Ej
hypotheses
Em
evidence/manifestations
P(Hi )
P(E j | Hi )
P(Hi | E j )
E1, , Em
H1, , Hn
Ej and Hi are binary; hypotheses are mutually exclusive (nonoverlapping) and exhaustive (cover all possible cases)
Conditional probabilities:
P(Ej | Hi), i = 1, , n; j = 1, , m
Limitations of simple
Bayesian inference
Cannot easily handle multi-fault situation, nor cases where
intermediate (hidden) causes exist:
Disease D causes syndrome S, which causes correlated
manifestations M1 and M2