0% found this document useful (0 votes)
17 views64 pages

Uncertainity Measure

Probability theory provides a framework for quantifying uncertainty and reasoning about events. It allows representing uncertain data, combining pieces of uncertain data, and making inferences using uncertain data. Probability is a number between 0 and 1 that reflects the likelihood of an event occurring. Key concepts in probability theory include sample spaces, events, unions and intersections of events, mutually exclusive and independent events, axioms of probability, and conditional probability. Bayes' theorem relates conditional probabilities and allows updating beliefs based on evidence. It plays an important role in probabilistic reasoning.

Uploaded by

Narender Narru
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
17 views64 pages

Uncertainity Measure

Probability theory provides a framework for quantifying uncertainty and reasoning about events. It allows representing uncertain data, combining pieces of uncertain data, and making inferences using uncertain data. Probability is a number between 0 and 1 that reflects the likelihood of an event occurring. Key concepts in probability theory include sample spaces, events, unions and intersections of events, mutually exclusive and independent events, axioms of probability, and conditional probability. Bayes' theorem relates conditional probabilities and allows updating beliefs based on evidence. It plays an important role in probabilistic reasoning.

Uploaded by

Narender Narru
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 64

Probability Theory:

Uncertainty Measure
Uncertainty
● Most intelligent systems have some degree of
uncertainty associated with them.
● Uncertainty may occur in KBS because of the
problems with the data.
− Data might be missing or unavailable.
− Data might be present but unreliable.
− The representation of the data may be imprecise or
inconsistent.
Cont…
● For any form of uncertainty scheme, we must be
concerned with three issues.
− How to represent uncertain data?
− How to combine two or more pieces of uncertain data?
− How to draw inference using uncertain data?
● Probability is the oldest theory with strong
mathematical basis.
● Other methods for handling uncertainty are Bayesian
belief network, Certainty factor theory etc.
Probability Theory
● Probability is a way of turning opinion or expectation into
numbers.
● It lies between 0 to 1 that reflects the likelihood of an
event.
● The chance that a particular event will occur = the
number of ways the event can occur divided by the total
number of all possible events.
Example: The probability of throwing two successive
heads with a fair coin is 0.25
− Total of four possible outcomes are :
HH, HT, TH & TT
− Since there is only one way of getting HH,
probability = ¼ = 0.25
Range of
Probability
1 Certain

.5

0 Impossible

5
Event
Event: Every non-empty subset A (of sample space S) is
called an event.
− null set  is an impossible event.
− S is a sure event
● P(A) is notation for the probability of an event A.
● P() = 0 and P(S) = 1
● The probabilities of all events S = {A1, A2, …, An}
must sum up to certainty i.e. P(A1) + … + P(An) = 1
● Since the events are the set, it is clear that all set
operations can be performed on the events.
● If A and B are events, then
− A  B ; A B and A' are also events.
− A - B is an event "A but not B
− Events A and B are mutually exclusive, if A  B=
Union of
Sets
• The union of two sets contains an instance of each element of the two
sets.

X  1,4,7,9
Y  2,3,4,5,6
X  Y  1,2,3,4,5,6,7,9
X Y

C   I B M , D E C , Apple
F   A p p l e , G r a p e , Lime

C  F   I B M , D E C , Apple, Grape,

Lime

7
Intersection of Sets
• The intersection of two sets contains only those element common to the

two sets.
X  1,4,7,9 
Y  2,3,4,5,6  X Y

X  Y   4

C   IBM , DEC,
Apple F   A p p l e ,

G r a p e , Lime C  F  
Apple

8
Mutually Exclusive Events
• Events with no common outcomes
• Occurrence of one event precludes the occurrence of the other event

C   I B M , D E C , Apple
F   Grape, Lime
C  F   

X
X  1,7,9 Y
Y  2,3,4,5,6
X  Y   
P( X  Y ) 
0
9
Independent
Events
• Occurrence of one event does not affect the occurrence or nonoccurrence
of the other event
• The conditional probability of X given Y is equal to the marginal probability
of X.
• The conditional probability of Y given X is equal to the marginal probability
of Y.

P( X|Y )  P( X) and P(Y| X)  P(Y )

10
Axioms of Probability
● Let S be a sample space, A and B are events.
− P(A)  0
− P(S) = 1
− P(A’ ) = 1 - P(A)
− P(A  B ) = P(A) + P(B) – P(A  B)
− If events A and B are mutually exclusive, then
P(A  B ) = P(A) + P(B),
● In general, for mutually exclusive events A1,…,An in S
P(A1  A2 …  An ) = P(A1) + P(A2) + …+ P(An)
Joint Probability
● Joint Probability of the occurrence of two independent events is
written as P (A and B) and is defined by
P(A and B) = P(A  B) = P(A) * P(B)
Example: We toss two fair coins separately.
Let P(A) = 0.5 , Probability of getting Head of first coin
P(B) = 0.5, Probability of getting Head of second coin
● Probability (Joint probability) of getting Heads on both the coins is
= P(A and B)
= P(A) * P(B) = 0.5 X 0.5 = 0.25
● The probability of getting Heads on one or on both of the coins i.e.
the union of the probabilities P(A) and P(B) is expressed as
P(A or B) = P(A  B) = P(A) + P(B) - P(A) * P(B)
= 0.5 X 0.5 - 0.25
= 0.75
Joint Probability Distribution
 Joint Probability Distribution of 2 variables A
and B

Joint Probabilities A A’

B 0.2 0.12
B’ 0.65 0.03

 Joint Probability Distribution of n variables


have 2n entries in the Joint Probability
Distribution Table
Problem

 •A company has 140 employees, of which 30


are supervisors.
 • Eighty of the employees are married, and
20% of the married employees are
supervisors.
 •If a company employee is randomly
selected, what is the probability that the
employee is married and is a supervisor?


Conditional Probability
● It relates the probability of one event to the occurrence of another
i.e. probability of the occurrence of an event H given that an
event E is known to have occurred.
● Probability of an event H (Hypothesis), given the occurrence of
an event E (evidence) is denoted by P(H | E) and is defined as
follows:

Number of events favorable to H


which are also favorable to E
P(H | E) =
No. of events favorable to E

P(H and E)
=
P(E)
Example-1
● What is the probability of a person to be male if
person chosen at random is 80 years old?

● The following probabilities are given


− Any person chosen at random being male is about 0.50
− probability of a given person be 80 years old chosen at
random is equal to 0.005
− probability that a given person chosen at random is both
male and 80 years old may be = 0.5 * 0.005 = 0.0025
Example-1 Soln
● The probability that an 80 years old
person chosen at random is male is
calculated as follows:
P(X is male | Age of X is 80)
= [P(X is male and the age of X is 80)] /
[P(Age of X is 80)]
= 0.0025 / 0.005 = 0.5
Conditional Probability with Multiple
Evidences

● If there are n evidences and one hypothesis, then


conditional probability is defined as follows:

P(H and E1 … and En)


P(H | E1 and … and En) =
P(E1 and … and En)
Bayes’ Theorem
· Bayes theorem provides a mathematical model for this
type of reasoning where prior beliefs are combined
with evidence to get estimates of uncertainty.
· This approach relies on the concept that one should
incorporate the prior probability of an event into the
interpretation of a situation.
· It relates the conditional probabilities of events.
· It allows us to express the probability P(H | E) in terms
of the probabilities of P(E | H), P(H) and P(E).
P(E|H) * P(H)
P(H|E) =
P(E)
Proof of Bayes’ Theorem
● Bayes’ theorem is derived from conditional probability.
Proof: Using conditional probability
P(H|E) = P(H and E) / P(E)
 P(H|E) * P(E) = P(H and E) (1)
Also P(E|H) = P(E and H) / P(H)
 P(E|H) * P(H) = P(E and H) (2)

From Eqs (1) and (2), we get


P(H|E) * P(E) = P(E|H) * P(H)
Hence, we obtain
P(E|H) * P(H)
P(H|E) =
P(E)
Extension of Bayes’ Theorem
● Consider one hypothesis H and two evidences E1 and
E2.
● The probability of H if both E1 and E2 are true is
calculated by using the following formula:

P(E1| H) * P(E2| H) * P(H)


P(H|E1 and E2) =
P(E1 and E2)
Contd..
● Consider one hypothesis H and Multiple evidences E1,
…., En.
● The probability of H if E1,…, En are true is calculated
by using the following formula:

P(E1| H) * … * P(En | H) * P(H)


P(H|E1 and … and En) =
P(E1 and … and En)
Example-2
Marie's marriage is tomorrow. In recent years,
each year it has rained only 5 days.
The weatherman has predicted rain for
tomorrow.
When it actually rains, the weatherman correctly
forecasts rain 90% of the time.
When it doesn't rain, the weatherman
incorrectly forecasts rain 10% of the time.
The question : What is the probability that it
will rain on the day of Marie's wedding?
Example 3 -Clinic Trial
In a clinic, the probability of the patients having K

virus is 0.15. A blood test done on patients :


If patient has virus, then the test is +ve with
probability 0.95.
If the patient does not have the virus, then the test is
+ve with probability 0.02.
Find :
If the test is +ve what are the probabilities
that the patient
(i) Has The Virus (ii)Does not have Virus
If the test is -ve what are the probabilities
that the patient
(i) Has The Virus (ii)Does not have Virus
Example-4
Example-4
● Find whether Bob has a cold (hypotheses) given that
he sneezes (the evidence) i.e., calculate P(H | E).
● Suppose that we know / given the following.
P(H) = P (Bob has a cold) = 0.2
P(E | H)= P(Bob was observed sneezing
| Bob has a cold) = 0.75
P(E | ~H)= P(Bob was observed sneezing
| Bob does not have a cold) =
0.2
Now
P(H | E) = P(Bob has a cold | Bob was observed sneezing)
= [ P(E | H) * P(H) ] / P(E)
Cont…
● We can compute P(E) as follows:
P(E) = P( E and H) + P( E and ~H)
= P(E | H) * P(H) + P(E | ~H) * P(~H)
= (0.75)(0.2) + (0.2) (0.8) = 0.31
− Hence P(H | E) = [(0.75 * 0.2)] / 0.31 = 0.48387
− We can conclude that “Bob’s probability of having a cold given
that he sneezes” is about 0.5
● Further it can also determine what is his probability of
having a cold if he was not sneezing?
P(H | ~E) = [P(~E | H) * P(H)] / P(~E)
= [(1 – 0.75) * 0.2] / (1 – 0.31)
= 0.05 / 0.69 = 0.072
− Hence “Bob’s probability of having a cold if he was not
sneezing” is 0.072
Advtg and Disadvtg
 Strong theoretical foundation.
 Has well defined semantics for decision making.

 The system using Bayesian approach needs


quite a large amount of probability data to
construct a knowledge base.
Bayesian Belief Network
● Joint probability distribution of two variables A and B
are given in the following Table

Joint Probabilities A A’
B 0.20 0.12
B’ 0.65 0.03
● Joint probability distribution for n variables require 2n
entries with all possible combinations.
● The time and storage requirements for such
computations become impractical as n grows.
Contd..
● Inferring with such large numbers of probabilities
does not seem to model human process of
reasoning.
● Human tends to single out few propositions which are
known to be causally linked when reasoning with
uncertain beliefs.
● This leads to the concept of forming belief network
called a Bayesian belief network.
● It is a probabilistic graphical model that encodes
probabilistic relationships among set of variables with
their probabilistic dependencies.
● This belief network is an efficient structure for storing
joint probability distribution.
Definition of BBN
● It is a acyclic (with no cycles) directed graph where
the nodes of the graph represent evidence or
hypotheses and arc connecting two nodes represents
dependence between them.
● If there is an arc from node X to another node Y (i.e.,
X Y), then X is called a parent of Y, and Y is a child
of X.
● The set of parent nodes of a node Xi is represented
by parent_nodes(Xi).
Joint Probability of n variables
● Joint probability for ‘n’ variables (dependent or
independent) is computed as follows.
● For the sake of simplicity we write P(X1 , … , Xn)
instead of P(X1 and … and Xn).

P(X1 , … ,Xn) = P(Xn | X1 ,…, Xn-1) * P(X1 , … , Xn-1)


Or
P(X1 , … , Xn) = P(Xn | X1 , … , Xn-1) * P(Xn-1 | X1 , … , Xn-2) * …. * P(X2 | X1) * P(X1)
Joint Probability of ‘n’ Variables using B-
Network
● In Bayesian Network, the joint probability distribution
can be written as the product of the local distributions
of each node and its parents such as:

n
P(X1, … , Xn) =  P(Xi | parent_nodes(Xi))
i =1
● This expression is reduction of joint probability
formula of ‘n’ variables as some of the terms
corresponding to independent variables will not be
required.
Cont…
● If node Xi has no parents, its probability distribution
is said to be unconditional and it is written as P(Xi)
instead of P(Xi | parent_nodes(Xi)).
● Nodes having parents are called conditional.
● If the value of a node is observed, then the node is
said to be an evidence node.
● Nodes with no children are termed as hypotheses
node and nodes with no parents are called
independent nodes.
Example

● The following graph is a Bayesian belief network.


− Here there are four nodes with {A, B} representing
evidences and {C, D} representing hypotheses.
− A and B are unconditional nodes and C and D are
conditional nodes.

A B

C D

Bayesian Belief Network


Cont...
● To describe above Bayesian network, we should
specify the following probabilities.

P(A) = 0.3
P(B) = 0.6
P(C|A) = 0.4
P(C|~A) = 0.2
P(D|A, B) = 0.7
P(D|A, ~B) = 0.4
P(D|~A, B) = 0.2
P(D|~A, ~B) = 0.01
Cont…
● They can also be expressed as conditional probability
tables as follows:

Conditional Probability Tables

P(A) P(B) A P(C) A B P(D)


0.3 0.6 T 0.4 T T 0.7
F 0.2 T F 0.4
F T 0.2
F F 0.01
Cont…
● Using Bayesian belief network on previous slide,
only 8 probability values in contrast to 16 values are
required in general for 4 variables {A, B, C, D} in joint
distribution probability.
● Joint probability using Bayesian Belief Network is
computed as follows:

P(A, B, C, D) = P(D|A, B) * P(C|A) * P(B) * P(A)


= 0.7 * 0.4 * 0.6 * 0.3 = 0.0504
Example of Simple B-Network
● Suppose that there are three events namely
earthquake, burglary or tornado which could cause
ringing of alarm in a house.
● This situation can be modeled with Bayesian network
as follows.
● All four variables have two possible values T (for true)
and F (for false).
− Here the names of the variables have been abbreviated
to A = Alarm, E = Earthquake, and B = Burglary and T =
Tornado.
Cont…

Earthquake Tornado Burglary

Alarm ringing
Cont…
● Table contains the
probability values Conditional Probability Tables
representing
complete Bayesian
P(E) P(B) E B Tor P(A)
belief network. 0.4 0.7 T T T 1.0
● Prior probability of T T F 0.9
‘earthquake’ is 0.4
and if it is E P(Tor) T F T 0.95
earthquake then T 0.8 T F F 0.85
probability of
‘tornado’ is 0.8. and
F 0.5 F T T 0.89
if not then the F T F 0.7
probability of F F T 0.87
‘earthquake’ is 0.5.
F F F 0.3
Cont…
● The joint probability is computed as follows:
P(E, B, T, A) = P(A| E, B, T) * P(T|E) * P(E) *
P(B)
= 1.0 * 0.8 * 0.4 * 0.7 = 0.214
● Using this model one can answer questions using the
conditional probability formula as follows:
− "What is the probability that it is earthquake, given the alarm
is ringing?" P(E|A)
− "What is the probability of burglary, given the alarm is
ringing?" P(B|A)
− "What is the probability of ringing alarm if both earthquake
and burglary happens?" P(A|E, B)
Certainty Factor
 Developed by MYCIN
 Certainty Factor
CF(H , E) = MB(H , E) – MD(H , E)
CF is the certainty factor in the hypothesis H
due to evidence E
MB is the measure of increased belief in H
due to E
MD is the measure of increased disbelief in H
due to E
Characteristics
Two Evidences supporting hypothesis
● Case1: Incrementally acquired evidence
E1  H, E2  H
● Let us first compute MB(H, E1 and E2) and
MD(H, E1 and E2)
0, if MD[H, E1 and E2] = 1
MB[H, E1 and E2] =
MB[H, E1] + MB[H, E2] * (1- MB[H, E1]), otherwise

0, if MB[H, E1 and E2] = 1


MD[H, E1 and E2] =
MD[H, E1] + MD[H, E2] * (1- MD[H, E1]), otherwise

● CF[H,E1andE2] = MB[H,E1andE2] – MD[H,E1andE2]


Example
● Suppose we make an initial observation E1 that
confirms our belief in H with MB[H, E1) = 0.4 and
MD(H, E1) = 0. Consider second observation E2 that
also confirms H with MB[H, E2) = 0.3. Then CF(H, E1)
= 0.4

MB(H, E1 and E2) = MB(H, E1) + MB(H, E2) * (1 – MB(H,


E1))
= 0.4 + 0.3 * (1-MB(H, E1))
= 0.4 +0.18 = 0.58
and
MD(H, E1 and E2) = 0.0
Therefore,
CF(H, E1 and E2) = 0.58

● Here we notice that slight confirmatory evidence can


larger certainty factor.
Cont…
● Case 2: There are two hypotheses H1 and H2 based
on the same evidence E. Find CF for conjunction and
disjunction of hypotheses.
Dempster–Shafer Theory
● It is a mathematical theory of evidence.
● It allows one to combine evidence from different
sources and arrive at a degree of belief.
● Belief function is basically a generalization of the
Bayesian theory of probability.
● Belief functions allow us to base degrees of belief or
confidence for one event on probabilities of related
events, whereas Bayesian theory requires
probabilities for each event.
● These degrees of belief may or may not have the
mathematical properties of probabilities.
Cont…
● The difference between them will depend on how
closely the two events are related.
● It also uses numbers in the range [0, 1] to indicate
amount of belief in a hypothesis for a given piece of
evidence.
● Degree of belief in a statement depends upon the
number of answers to the related questions
containing the statement and the probability of each
answer.
● In this formalism, a degree of belief (also referred to
as a mass) is represented as a belief function rather
than a Bayesian probability distribution
Dempster Theory Formalism
● Let U be the universal set of all hypotheses,
propositions, or statements under consideration.
● The power set P(U), is the set of all possible subsets of U,
including the empty set represented by .
● The theory of evidence assigns a belief mass to each subset
of the power set.
● A function m: P(U)  [0,1] is called a basic belief
assignment (BBA) function. It satisfies the following
axioms:
− m() = 0 ;  m(A) = 1, A  P(U)
● The value of m(A) is called mass assigned to A on the unit
interval.
● It makes no additional claims about any subsets of A,
each of which has, by definition, its own mass.
Dempster's Rule of Combination
● The original combination rule, known as Dempster's
rule of combination, is a generalization of Bayes' rule.
● Assume that m1 and m2 are two belief functions
used for representing multiple sources of evidences
for two different hypotheses.
● Let A, B  U, such that m1(A) ≠ 0, and m2(B) ≠ 0.
● The Dempster's rule for combining two belief
functions to generate an m3 function may be defined
as:
m3() = 0
 A  B = C (m1(A) * m2(B))
m3(C) =
1 -  A  B =  (m1(A) * m2(B))
Cont…
● This belief function gives new value when applied on
the set C = A  B.
● The combination of two belief functions is called the
joint mass.
− Here m3 can also be written as (m1  m2).
● The expression [  A  B =  (m1(A) * m2(B))] is called
normalization factor.
− It is a measure of the amount of conflict between the two
mass sets.
● The normalization factor has the effect of completely
ignoring conflict and attributing any mass associated
with conflict to the null set.
Example : Diagnostic System
● Suppose we have mutually exclusive hypotheses
represented by a set U = {flu, measles, cold, cough}.
● The goal is to assign or attach some measure of belief
to the elements of U based on evidences.
− It is not necessary that particular evidence is supporting some
individual element of U but rather it may support subset of U.
− For example, an evidence of ‘fever’ might support {flu,
measles}.
● So a belief function ‘m’ is defined for all subsets of U.
● The degree of belief to a set will keep on changing if
we get more evidences supporting it or not.
Example : Cont…
● Initially assume that we have no information about
how to choose hypothesis from the given set U.
● So assign m for U as 1.0 i.e., m(U) = 1.0
− This means we are sure that answer is somewhere in the
whole set U.
● Suppose we acquire evidence (say fever) that
supports the correct diagnosis in the set {flu, measles}
with its corresponding ‘m’ value as 0.8.
● Then we get m({flu, measles}) = 0.8 and m(U) = 0.2.
Cont…
● Let us define two belief functions m1 and m2 based
on evidence of fever and on evidence of headache
respectively as follows:
m1({flu, measles}) = 0.8
m1(U) = 0.2
m2({flu, cold}) = 0.6
m2(U) = 0.4
● We can compute their combination m3 using these
values.
Cont…

Combination of m1 and m2 m2({flu, cold}) = 0.6 m2(U) = 0.4

m1({flu, measles}) = 0.8 m3({flu}) = 0.48 m3({flu, measles}) = 0.32


m1(U) = 0.2 m3({flu, cold}) = 0.12 m3(U) = 0.08
● Now previous belief functions are modified to m3 with
the following belief values and are different from
earlier beliefs.
m3({flu}) = 0.48
m3({flu, cold}) = 0.12
m3({flu, measles}) = 0.32
m3(U) = 0.08
● Further, if we have another evidence function m4 of
sneezing with the belief values as:
m4({cold, cough}) = 0.7
m4(U) = 0.3
● Then the combination of m3 and m4 gives another
belief function as follows:
Cont…
Combination of m3 and m4 m4({cold, cough}) = 0.7 m4(U) = 0.3

m3({flu}) = 0.48 m5(  ) = 0.336 m5({flu}) = 0.114


m3({flu, cold)) = 0.12 m5({cold}) = 0.084 m5({flu, cold)) = 0.036
m3({flu, measles})= 0.32 m5( ) = 0.224 m5({flu, measles})= 0.096
m3(U) = 0.08 m5({cold, cough}) = 0.056 m5(U) = 0.024

You might also like