Ai PPT 3
Ai PPT 3
Probabilistic
reasoning in Artificial intelligence
Chandigarh Group of Colleges, Landran
UNCERTAINTY
CAUSES OF UNCERTAINTY
PROBABILISTIC REASONING
In probabilistic reasoning, there are two ways to solve problems with uncertain
knowledge:
• Bayes' rule
• Bayesian Statistics
Chandigarh Group of Colleges, Landran
PROBABILITY
• Probability can be defined as a chance that an uncertain event will occur. It is the
numerical measure of the likelihood that an event will occur. The value of probability
always remains between 0 and 1 that represent ideal uncertainties.
1. 0 ≤ P(A) ≤ 1, where P(A) is the probability of an event A.
2. P(A) = 0, indicates total uncertainty in an event A.
3. P(A) =1, indicates total certainty in an event A.
Chandigarh Group of Colleges, Landran
PROBABILITY CONT.
We can find the probability of an uncertain event by using the below formula.
CONDITIONAL PROBABILITY
• Conditional probability is a probability of occurring an event when another event has already
happened.
• Let's suppose, we want to calculate the event A when event B has already occurred, "the
probability of A under the conditions of B", it can be written as:
VENN DIAGRAM
Chandigarh Group of Colleges, Landran
• Example:
• In a class, there are 70% of the students who like English and 40% of the students
who likes English and mathematics, and then what is the percent of students
those who like English also like mathematics?
• Solution:
• Let, A is an event that a student likes Mathematics
• B is an event that a student likes English
• Hence, 57% are the students who like English also like Mathematics.
Chandigarh Group of Colleges, Landran
BAYES' THEOREM
• Bayes' theorem is also known as Bayes' rule, Bayes' law, or Bayesian reasoning,
which determines the probability of an event with uncertain knowledge.
• In probability theory, it relates the conditional probability and marginal probabilities
of two random events.
• Bayes' theorem was named after the British mathematician Thomas Bayes.
The Bayesian inference is an application of Bayes' theorem, which is fundamental to
Bayesian statistics.
• It is a way to calculate the value of P(B|A) with the knowledge of P(A|B).
• Bayes' theorem allows updating the probability prediction of an event by observing
new information of the real world.
Chandigarh Group of Colleges, Landran
EXAMPLE
• If cancer corresponds to one's age then by using Bayes' theorem, we can
determine the probability of cancer more accurately with the help of age.
• Bayes' theorem can be derived using product rule and conditional probability of
event A with known event B:
• As from product rule we can write:
P(A ⋀ B)= P(A|B) P(B) or
• Similarly, the probability of event B with known event A:
P(A ⋀ B)= P(B|A) P(A)
• Equating right hand side of both the equations, we will get:
Chandigarh Group of Colleges, Landran
EXAMPLE CONT.
• The above equation (a) is called as Bayes' rule or Bayes' theorem. This equation is basic of most
modern AI systems for probabilistic inference.
• It shows the simple relationship between joint and conditional probabilities. Here,
• P(A|B) is known as posterior, which we need to calculate, and it will be read as Probability of
hypothesisA when we have occurred an evidence B.
• P(B|A) is called the likelihood, in which we consider that hypothesis is true, then we calculate the
probability of evidence.
• P(A) is called the prior probability, probability of hypothesis before considering the evidence
• P(B) is called marginal probability, pure probability of an evidence.
• In the equation (a), in general, we can write P (B) = P(A)*P(B|Ai), hence the Bayes' rule can be
written as
• Where A1, A2, A3,........, An is a set of mutually exclusive and exhaustive events.
Chandigarh Group of Colleges, Landran
EXAMPLE -1:
Question: what is the probability that a patient has diseases meningitis with a stiff
neck?
Given Data:
– A doctor is aware that disease meningitis causes a patient to have a stiff neck, and it occurs
80% of the time. He is also aware of some more facts, which are given as follows:
– The Known probability that a patient has meningitis disease is 1/30,000.
– The Known probability that a patient has a stiff neck is 2%.
• Let a be the proposition that patient has stiff neck and b be the proposition that patient
has meningitis. , so we can calculate the following as:
– P(a|b) = 0.8
– P(b) = 1/30000
– P(a)= .02
•
• Hence, we can assume that 1 patient out of 750 patients has meningitis disease with a
stiff neck.
Chandigarh Group of Colleges, Landran
EXAMPLE -2: a standard deck of playing cards, a single card is drawn. The
Question: From
probability that the card is king is 4/52, then calculate posterior probability
P(King|Face), which means the drawn face card is a king card.
Solution:
IN ARTIFIC ial
INTELLIGENCE
Following are some applications of Bayes' theorem:
• It is used to calculate the next step of the robot when the already executed step is
given.
• Bayes' theorem is helpful in weather forecasting.
• It can solve the Monty Hall problem.
BAYESIAN BELIEF NETWORK IN
ARTIFICIAL INTELLIGENCE
• Bayesian belief network is key computer technology for dealing with probabilistic events
and to solve a problem which has uncertainty. We can define a Bayesian network as:
• "A Bayesian network is a probabilistic graphical model which represents a set of variables
and their conditional dependencies using a directed acyclic graph."
• It is also called a Bayes network, belief network, decision network, or Bayesian model.
• Bayesian networks are probabilistic, because these networks are built from a probability
distribution, and also use probability theory for prediction and anomaly detection.
• It can also be used in various tasks including prediction, anomaly detection, diagnostics,
automated insight, reasoning, time series prediction, and decision making under
uncertainty.
Chandigarh Group of Colleges, Landran
it can be written as the following way in terms of the joint probability distribution.
= P[x1| x2, x3,....., xn]P[x2, x3,....., xn]
= P[x1| x2, x3,....., xn]P[x2|x3,....., xn]....P[xn-1|xn]P[xn].
• In general for each variable Xi, we can write the equation as:
P(Xi|Xi-1,........., X1) = P(Xi |Parents(Xi ))
Chandigarh Group of Colleges, Landran
EXAMPLE
Chandigarh Group of Colleges, Landran
• In the above figure, we have an alarm ‘A’ – a node, say installed in a h ouse of a
person ‘gfg’, which rings upon two probabilities i.e burglary ‘B’ an d fire ‘F’,
which are – parent nodes of the alarm node. The alarm is the parent node of two
probabilities P1 calls ‘P1’& P2 calls ‘P2’person nodes.
• Upon the instance of burglary and fire, ‘P1’ and ‘P2’ call person ‘gfg’,
respectively. But, there are few drawbacks in this case, as sometimes ‘P1’ may
forget to call the person ‘gfg’, even after hearing the alarm, as he has a tendency
to forget things, quick. Similarly, ‘P2’, sometimes fails to call the person ‘gfg’,
as he is only able to hear the alarm, from a certain distance.
Chandigarh Group of Colleges, Landran
• Q) Find the probability that ‘P1’ is true (P1 has called ‘gfg’), ‘P2’ is true (P2 h as
called ‘gfg’) when the alarm ‘A’ rang, but no burglary ‘B’ and fire ‘F’ has
occurred.
• => P ( P1, P2, A, ~B, ~F) [ where- P1, P2 & A are ‘true’ events and ‘~B’ & ‘~F’
are ‘false’events]
• [ Note: The values mentioned below are neither calculated nor computed. They
have observed values ]
• Burglary ‘B’ –
• P (B=T) = 0.001 (‘B’ is true i.e burglary has occurred)
• P (B=F) = 0.999 (‘B’ is false i.e burglary has not occurred)
• Fire ‘F’ –
• P (F=T) = 0.002 (‘F’ is true i.e fire has occurred)
• P (F=F) = 0.998 (‘F’is false i.e fire has not occurred)
Chandigarh Group of Colleges, Landran
ALARM ‘A’
–
B F P (A=T) P (A=F)
T T 0.95 0.05
T F 0.94 0.06
F T 0.29 0.71
F F 0.001 0.999
•The alarm ‘A’ node can be ‘true’ or ‘false’ ( i.e may have rung or may not have
rung). It has two parent nodes burglary ‘B’ and fire ‘F’ which can be ‘true’ or ‘false’
(i.e may have occurred or may not have occurred) depending upon different
conditions.
Chandigarh Group of Colleges, Landran
PERSON ‘P1’
–
A P (P1=T) P (P1=F)
T 0.95 0.05
F 0.05 0.95
•The person ‘P1’node can be ‘true’or ‘false’(i.e may have called the person ‘gfg’
or not) . It has a parent node, the alarm ‘A’, which can be ‘true’or ‘false’(i.e may
have rung or may not have rung ,upon burglary ‘B’or fire ‘F’).
Chandigarh Group of Colleges, Landran
PERSON ‘P2’
–
A P (P2=T) P (P2=F)
T 0.80 0.20
F 0.01 0.99
•The person ‘P2’ node can be ‘true’ or false’ (i.e may have called the person ‘gfg’
or not). It has a parent node, the alarm ‘A’, which can be ‘true’ or ‘false’ (i.e may
have rung or may not have rung, upon burglary ‘B’or fire ‘F’).
Chandigarh Group of Colleges, Landran