0% found this document useful (0 votes)
56 views

CBSE Class 12 Maths Notes Chapter 13 Probability

This document provides an overview of key concepts in probability, including: 1) Events are subsets of outcomes from random experiments and can be equally likely, mutually exclusive, or exhaustive. 2) The probability of an event is the number of favorable outcomes divided by the total number of outcomes. 3) Conditional probability is the probability of one event given that another event has occurred. 4) Independent events have probabilities that are unaffected by each other, while dependent event probabilities are related.

Uploaded by

lotusraghu5259
Copyright
© © All Rights Reserved
Available Formats
Download as DOC, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
56 views

CBSE Class 12 Maths Notes Chapter 13 Probability

This document provides an overview of key concepts in probability, including: 1) Events are subsets of outcomes from random experiments and can be equally likely, mutually exclusive, or exhaustive. 2) The probability of an event is the number of favorable outcomes divided by the total number of outcomes. 3) Conditional probability is the probability of one event given that another event has occurred. 4) Independent events have probabilities that are unaffected by each other, while dependent event probabilities are related.

Uploaded by

lotusraghu5259
Copyright
© © All Rights Reserved
Available Formats
Download as DOC, PDF, TXT or read online on Scribd
You are on page 1/ 5

CBSE Class 12 Maths Notes Chapter 13

Probability
Event: A subset of the sample space associated with a random experiment is
called an event or a case.
e.g. In tossing a coin, getting either head or tail is an event.

Equally Likely Events: The given events are said to be equally likely if none of
them is expected to occur in preference to the other.
e.g. In throwing an unbiased die, all the six faces are equally likely to come.

Mutually Exclusive Events: A set of events is said to be mutually exclusive, if


the happening of one excludes the happening of the other, i.e. if A and B are
mutually exclusive, then (A ∩ B) = Φ
e.g. In throwing a die, all the 6 faces numbered 1 to 6 are mutually exclusive,
since if any one of these faces comes, then the possibility of others in the same
trial is ruled out.

Exhaustive Events: A set of events is said to be exhaustive if the performance


of the experiment always results in the occurrence of at least one of them.
If E1, E2, …, En are exhaustive events, then E1 ∪ E2 ∪……∪ En = S.
e.g. In throwing of two dice, the exhaustive number of cases is 6 2 = 36. Since any
of the numbers 1 to 6 on the first die can be associated with any of the 6
numbers on the other die.

Complement of an Event: Let A be an event in a sample space S, then the


complement of A is the set of all sample points of the space other than the
sample point in A and it is denoted by A’or A¯.
i.e. A’ = {n : n ∈ S, n ∉ A]
Note:
(i) An operation which results in some well-defined outcomes is called an
experiment.
(ii) An experiment in which the outcomes may not be the same even if the
experiment is performed in an identical condition is called a random experiment.

Probability of an Event
If a trial result is n exhaustive, mutually exclusive and equally likely cases and m
of them are favourable to the happening of an event A, then the probability of
happening of A is given by
Note:
(i) 0 ≤ P(A) ≤ 1
(ii) Probability of an impossible event is zero.
(iii) Probability of certain event (possible event) is 1.
(iv) P(A ∪ A’) = P(S)
(v) P(A ∩ A’) = P(Φ)
(vi) P(A’)’ = P(A)
(vii) P(A ∪ B) = P(A) + P(B) – P(A ∩ S)

Conditional Probability: Let E and F be two events associated with the same


sample space of a random experiment. Then, probability of occurrence of event
E, when the event F has already occurred, is called a conditional probability of
event E over F and is denoted by P(E/F).

Similarly, conditional probability of event F over E is given as

Properties of Conditional Probability: If E and E are two events of sample


space S and G is an event of S which has already occurred such that P(G) ≠ 0,
then
(i) P[(E ∪ F)/G] = P(F/G) + P(F/G) – P[(F ∩ F)/G], P(G) ≠ 0
(ii) P[(E ∪ F)/G] = P(F/G) + P(F/G), if E and F are disjoint events.
(iii) P(F’/G) = 1 – P(F/G)
(iv) P(S/E) = P(E/E) = 1

Multiplication Theorem: If E and F are two events associated with a sample


space S, then the probability of simultaneous occurrence of the events E and F is
P(E ∩ F) = P(E) . P(F/E), where P(F) ≠ 0
or
P(E ∩ F) = P(F) . P(F/F), where P(F) ≠ 0
This result is known as multiplication rule of probability.

Multiplication Theorem for More than Two Events: If F, F and G are three events
of sample space, then

Independent Events: Two events E and F are said to be independent, if


probability of occurrence or non-occurrence of one of the events is not affected
by that of the other. For any two independent events E and F, we have the
relation
(i) P(E ∩ F) = P(F) . P(F)
(ii) P(F/F) = P(F), P(F) ≠ 0
(iii) P(F/F) = P(F), P(F) ≠ 0
Also, their complements are independent events,
i.e. P(E¯ ∩ F¯) = P(E¯) . P(F¯)
Note: If E and F are dependent events, then P(E ∩ F) ≠ P(F) . P(F).
Three events E, F and G are said to be mutually independent, if
(i) P(E ∩ F) = P(E) . P(F)
(ii) P(F ∩ G) = P(F) . P(G)
(iii) P(E ∩ G) = P(E) . P(G)
(iv)P(E ∩ F ∩ G) = P(E) . P(F) . P(G)
If atleast one of the above is not true for three given events, then we say that the
events are not independent.
Note: Independent and mutually exclusive events do not have the same
meaning.

Baye’s Theorem and Probability Distributions


Partition of Sample Space: A set of events E1, E2,…,En is said to represent a
partition of the sample space S, if it satisfies the following conditions:
(i) Ei ∩ Ej = Φ; i ≠ j; i, j = 1, 2, …….. n
(ii) E1 ∪ E2 ∪ …… ∪ En = S
(iii) P(Ei) > 0, ∀ i = 1, 2,…, n

Theorem of Total Probability: Let events E1, E2, …, En form a partition of the


sample space S of an experiment.If A is any event associated with sample space
S, then

Baye’s Theorem: If E1, E2,…,En are n non-empty events which constitute a


partition of sample space S, i.e. E1, E2,…, En are pairwise disjoint E1 ∪ E2 ∪ …….
∪ En = S and P(Ei) > 0, for all i = 1, 2, ….. n Also, let A be any non-zero event,
the probability

Random Variable: A random variable is a real-valued function, whose domain is


the sample space of a random experiment. Generally, it is denoted by capital
letter X.
Note: More than one random variables can be defined in the same sample
space.

Probability Distributions: The system in which the values of a random variable


are given along with their corresponding probabilities is called probability
distribution.
Let X be a random variable which can take n values x 1, x2,…, xn.
Let p1, p2,…, pn be the respective probabilities.
Then, a probability distribution table is given as follows:

such that P1 + p2 + P3 +… + pn = 1


Note: If xi is one of the possible values of a random variable X, then statement X
= xi is true only at some point(s) of the sample space. Hence ,the probability that
X takes value x, is always non-zero, i.e. P(X = xi) ≠ 0

Mean and Variance of a Probability Distribution: Mean of a probability


distribution is

Bernoulli Trial: Trials of a random experiment are called Bernoulli trials if they


satisfy the following conditions:
(i) There should be a finite number of trials.
(ii) The trials should be independent.
(iii) Each trial has exactly two outcomes, success or failure.
(iv) The probability of success remains the same in each trial.

Binomial Distribution: The probability distribution of numbers of successes in


an experiment consisting of n Bernoulli trials obtained by the binomial expansion
(p + q)n, is called binomial distribution.
Let X be a random variable which can take n values x 1, x2,…, xn. Then, by
binomial distribution, we have P(X = r) = nCr prqn-r
where,
n = Total number of trials in an experiment
p = Probability of success in one trial
q = Probability of failure in one trial
r = Number of success trial in an experiment
Also, p + q = 1
Binomial distribution of the number of successes X can be represented as

Mean and Variance of Binomial Distribution


(i) Mean(μ) = Σ xipi = np
(ii) Variance(σ2) = Σ xi2 pi – μ2 = npq
(iii) Standard deviation (σ) = √Variance = √npq
Note: Mean > Variance

You might also like