week 2 lecture notes
week 2 lecture notes
Probability Models
Prof. Abhinava Tripathi
Introduction
Introduction
• Probability models empower statisticians to draw inferences from sample
data
1. The experimental outcomes are well defined, and in many cases can even be listed
prior to conducting the experiment
2. On any single repetition or trial of the experiment, one and only one of the possible
experimental outcomes will occur
3. The experimental outcome that occurs on any trial is determined solely by chance
Random Experiments and Sample Space
• Whether it’s cricket, football or soccer or
basketball, tossing a coin is a very common
feature
• When you toss it, the chance of heads is the same as the chance of tails
• What do you think will be the chance that the captain wins?
Random Experiments and Sample Space
• You can have only 2 outcomes: heads or tails
• The sample space for a random experiment is the set of all experimental
outcomes
• The set of possible outcomes {Head, Tails} is known as the sample space
• In our example, the probability of the captain winning is known as the event
Probability Events
Probability Events
• Probability is a numerical measure of the likelihood that an event will occur
• certain
• to occur
Probability Events
• Let us consider the example of rolling dice with 6 faces
• What about the probability of getting an even number, that is, 2, 4, 6: 3/6 =
0.5.
Probability Events
• In this example, Rolling a dice is a random experiment
• ‘Rolling a one’ or the ‘Rolling an even number’ are the events for which we
find the probability
• The probability formula we discussed may not apply at all the places
Probability Events
• Let’s consider the marks scored out of 100 by a student in a math test
• What can be the outcomes of the test: any number between 0 and 100
• You can have a test that is designed such that people get at least a passing
grade
• For example, if there were 50 students in the class who took the math test
• In our case, the experiment is repeated for 50 students and the Frequency of
favorable outcomes is 10: probability is 10/50= 0.20
Probability Events
• The formula: “Number of favorable outcome divided by Total outcomes” worked for
the coin toss and rolling a die experiments because different outcomes of
experiments do not depend on what happened in the past
• The different outcomes the case of coin toss and in case of die roll are equally likely
• That is not the case with the Math exam, which is designed in a way that one
outcome, say, scoring 0, is not as likely as scoring a 100, or scoring 60 or 80
Probability Definition
Probability Definition
• We calculate probability when we want to understand the chances of an event
happening in a random experiment
• It might or might not rain on a given day; hence there are two possible events
• Often, the news on weather channel tells you that there is a 20% or 30%
chance of rainfall on a given day
• Let us flip two coins together: what would be the sample space here
• If we use H to denote head and T to denote tail, then then our sample space
becomes: S= HH, HT, TH, and TT
Combining two or more experiments
• In this experiment, the total number of possible outcomes are 2x2=4
• Let us generalize this with a random event in k-steps, where the possible
outcomes are n1, n2,…nk
• Event B, is represented by the area that is not in A, which is the white portion
within the rectangle
• This takes us to the basic rule of probability, that is, the sum of the probabilities of
all the events always add up to 1
Intersection Events
Intersection Events
• An event is a subset of all the
possible outcomes of a random
experiment
• It can be broken down into three simple events, ‘Rolling a one’, ‘Rolling a three’
and ‘Rolling a five’
• For Event A, the sample outcomes are (5, 2), (5, 3) and (5, 4)
• For Event B, the sample outcomes are (5, 2), (6, 2) and (7, 2)
• Consider event A where first die has number 1 and event B, where second dice has
the number 5
Intersection Events
• Each outcome has two values as they represent the individual outcomes of each of
the two dice
• There are six possible outcomes for event A: {(1, 1), (1, 2), (1, 3), (1, 4), (1, 5), (1,
6)}
• Again the possible outcomes for event B are: {(1, 5), (2, 5), (3, 5), (4, 5), (5, 5), (6,
5)}
• Event A, where the first dice rolls 1 (6 possible outcomes): {(1, 1), (1, 2), (1, 3), (1,
4), (1, 5), (1, 6)}
• Event B, where the second dice rolls 5 (6 possible outcomes):{(1, 5), (2, 5), (3, 5),
(4, 5), (5, 5), (6, 5)}
Union of Events
• Event A, where the first dice rolls 1 (6 possible
outcomes): {(1, 1), (1, 2), (1, 3), (1, 4), (1, 5), (1, 6)}
• All the possible AUB outcomes are listed here: {(1, 1), (1, 2), (1, 3), (1, 4), (1, 5),
(1, 6), (2, 5), (3, 5), (4, 5), (5, 5), (6, 5) }
Mutually Exclusive Events
Mutually Exclusive Events
• We can say that two events are mutually exclusive when
they do not occur at the same time; that is why they are
also called as disjoint events
• In the Venn diagram shown on the screen, you can see that events A and B do not
have an overlapping region
Mutually Exclusive Events
• This is a clear indication that the two events are mutually
exclusive or disjoint
• Event A is the roll where the first die has the number 1. Sample space: (1, 1), (1,
2), (1, 3), (1, 4), (1, 5), (1, 6)
• Event B is Rolls where the first die has the number 3 and the second die has
number 5. Sample space: {(3, 5)}.
Mutually Exclusive Events
• Event A is Rolls where the first die has the number 1.
Sample space: (1, 1), (1, 2), (1, 3), (1, 4), (1, 5), (1, 6)
• Event B is Rolls where the first die has the number 3 and
the second die has number 5. Sample space: {(3, 5)}.
• There is no intersection between the two events; hence, these two events are
mutually exclusive
Probability for complex events
Probability for complex events
• Let us work through some classic problems related to complex events
• In how many ways you can draw 2 balls from the bin with replacement?
• Similarly, you can draw 4 ways in the second draw. Hence, you can have 4*4 = 16
ways of drawing the 2 balls with replacement
Probability for complex events
• How many ways can you draw 2 balls with replacement such that both
are yellow?
• In each draw you have only 1 yellow ball. So there is only 1 way in
which both the draws are yellow balls
• The probability that all these four events will happen together is = P(A)*
P(B)*P(C)*P(D)
Probability for complex events
• The multiplication rule can also be applied for dependent events and the formula
changes slightly
• Independent events are such that the occurrence of one event is not dependent on
the other event
Probability for complex events
• What would happen to the probability of A and B happening together if the events
are mutually exclusive?
• The bin now has 20 balls and that out of all the 20 balls, there are 4 yellow balls, 5
green balls, 6 red balls and 5 blue balls
• With multiplication rule: P(A∩B) = P(A) * P(B) = 4/20 *4/20 = 1/25 = 0.04
• What is the probability of drawing two yellow balls if we don't replace that ball after
we have drawn it?
Probability for complex events
• What is the probability of drawing two yellow balls if we don't replace that ball after
we have drawn it?
• Since there is no replacement, the 2nd draw becomes dependent on the 1st draw
• Event B is conditional on event A due to the act of not drawing with replacement
• The total number of ways we can draw a green ball in draw 1 and a blue ball in draw
2 is 5 x 5 = 25
• We can define the events as follows: (a) Event A as the Probability of drawing a
yellow ball in draw 1; and (b) Event B as the Probability of drawing a yellow ball in
draw 2
• Since we are doing this without replacement, we can use the conditional probability
formula: P(B given A) = P(A በ B) / P(A)
Conditional Probabilities
• We get P(A በ B) = P(B given A) * P(A); where P(A) = 4/20
• Here P(B/A): the probability of drawing a yellow ball in the 2nd draw given we have
• So, we have 19 balls left. Since, we have already drawn a yellow ball, 3 yellow balls
• If you look at the formula of P(A በ B) for dependent and independent events, the only
• Second, the condition for the conditional probability formula P(B given A) = P(A በ B)
/ P(A) to work is that the denominator is non zero, that is P(A) is not 0.
Addition Rule of probability
Addition Rules of probability
• Consider a simple Venn Diagram problem: Your city, which has 100 households,
there are two newspapers. Times and Daily news. The circulation departments
report that 25 of the city’s households have a subscription to the Times and 35
subscribe to the Daily News. A survey reveals that 6 of all households subscribe to
newspaper?
Addition Rules of probability
• A is the number of households reading
the Times and B is the number of
households reading the Daily News
• The shaded region that overlaps A with B, that is A ∩ B indicates the households
which have subscription to both
• The shaded region that overlaps A with B, that is A ∩ B indicates the households
which have subscription to both
• If A and B are mutually exclusive, then P(A ∩ B) =0: P(A U B) = P(A) + P(B).
Joint and Marginal probability
Joint and Marginal probability
• The three major types of probability include: (a) Joint probability, (b) Marginal
probability, and (c) Conditional probability
Fund manager graduated from a top-30 MBA Fund does not outperform the
100
program market
Fund manager did not graduate from a top-30 Fund does not outperform the
244
MBA program market
Joint and Marginal probability
Now before we calculate the different types of probabilities, we need to first convert this
into a probability table
Fund manager graduated from a top-30 MBA Fund does not outperform the
0.34
program market
Fund manager did not graduate from a top-30 Fund does not outperform the
0.61
MBA program market
Joint and Marginal probability
We have also given notations to each of these events for simplicity
A1 B1 0.1
A1 B2 0.34
A2 B1 0.04
A2 B2 0.61
• The joint probability of event A1 and B1 is equal to P(A1 intersection B1) which is 0.11
• The joint probability of events A1 and B2 is equal to 0.34
• The joint probability of event A2 and B1 is equal to 0.04
• And the joint probability of events A2 and B2 is equal to 0.61
Joint and Marginal probability
Next let us understand the concept of marginal probability
A1 B1 0.1
A1 B2 0.34
A2 B1 0.04
A2 B2 0.61
A1 B1 0.1
A1 B2 0.34
A2 B1 0.04
A2 B2 0.61
• The following formula is used: P(B1 given A1) = P(A1 በ B1) / P(A1)
• The probability that a manager graduated from a top MBA program: P(A1) = 0.10 +
0.34 = 0.44
• Here, the joint probability P(A1 በ B1) from the table as 0.1
• P(A given B) = P(A) since P(A1 በ B1) = P(A) * P(B) for independent events
• If the event, ‘Mutual Fund outperforms the market’ and the event, ‘Fund manager is
from a top MBA school’ are independent of each other or not?
Fund manager graduated from a top-30 MBA Fund does not outperform the
0.34
program market
Fund manager did not graduate from a top-30 Fund does not outperform the
0.61
MBA program market
Bayes Theorem
• We calculated the probability that a mutual fund will outperform the market given the
fund manager graduated from a top MBA program
• Also, P(B given A) P(A) = P(A intersection B)= P(A given B) P(B)
• Consider the following example, past research suggests that the probability of a
middle-aged female developing breast cancer is 0.01 or 1% (Event B)
• A person can be tested positive for breast cancer but in actuality, one may not have it,
and vice-versa
• Here, B is the event that a middle-aged female develops breast cancer; P(B)=0.01
Bayes Theorem
• So effectively, we are also asking how reliable is the testing methodology?
• Thus, the probability that we are interested in measuring is basically, P(B given A)
• In a sample of women who already had breast cancer, it was found that only 90% of
these women tested positive. This means that the probability of a woman testing
positive given she has breast cancer is 0.9: P(A given B) = 0.9.
Bayes Theorem
• From the Bayes rule, P(B given A) = P(A given B) P(B) / P(A)
• We have P(A given B) and P(B) with us; we need to calculate P(A)
• B’, which is the complement of B will be an event that a woman does not have breast
cancer
• We have P(A given B) and P(B) with us; we need to calculate P(A)
• B’, which is the complement of B will be an event that a woman does not have breast
cancer
• P(A) = P(A በ B) + P(A በ B’)= P(A given B)* P(B) + P(A given B’) *P(B’)
• P(B│A)= (P(A given B) P(B)) / (P(A given B) P(B) + P(A given B’) P(B’) )
• There is only a 10.2% chance that a woman will develop breast cancer if she tests
positive
• Bayes theorem : P(B│A)= (P(A given B) P(B)) / (P(A given B) P(B) + P(A given B’)
P(B’) )
B B’
• Let us understand the significance
A P(A∩B) P(A∩B’)
of this equation A’ P(A’∩B) P(A’∩B’)
Bayes Theorem
• We start off with a set of initial or prior probabilities
• Prior probability is the probability of an event before we did our experiment and
collected our new data
• In our case, the probability of event B, which was the probability of a female
developing breast cancer is the prior probability
• We obtained new information about all the events for which we had the prior
probabilities
• This information was given to us in the form of the conditional probabilities, P(A given
B) and P(A given B’)
Bayes Theorem
• Our objective was basically to update these our prior knowledge about females developing
breast cancer with this new information and calculate P(B given A)
• This new probability that we calculated is referred to as posterior probabilities and Bayes’
theorem is used for making these probability calculations
• In our case, the probability that a woman has or will develop breast cancer if she tests positive
was the posterior probability
• P(B│A)= (P(A given B) P(B)) / (P(A given B) P(B) + P(A given B’) P(B’) )
INDIAN INSTITUTE OF TECHNOLOGY KANPUR
Thanks!
INDIAN INSTITUTE OF TECHNOLOGY KANPUR
Probability Models
Prof. Abhinava Tripathi
Recap: Bayes Theorem
Recap: Bayes Theorem
• Bayes theorem can be used to calculate posterior probabilities and the
formula is
• P(B│A)= (P(A given B) P(B)) / P(A given B) P(B) + P(A given B’) P(B’)
• In the earlier breast cancer example, there were only two possibilities, either
the female has breast cancer (Event B, P(B)) or she doesn’t have (Event B’,
P(B’))
• What if there are more than two possibilities, i.e., B1, B2, and B3
Recap: Bayes Theorem
• Consider Event Bi, where a patient is treated at hospital i (1, 2, 3). Thus, the
probability that a patient was treated in hospital 1 is P(B1) which is 0.6.
Similarly, the probability that a patient was treated in hospital 2 is P(B2)
which is 0.3. And the probability that a patient was treated in hospital 3 is
P(B3) which is 0.1. So these are our prior probabilities
• Here we have B1, B2 and B3 such that the probabilities of all these 3 events
sum upto 1
• Let’s say a malpractice suit is filed today. Then what is the probability that the
suit was filed against hospital 1?
• Remember that we also have the prior probabilities that the patient was
admitted to hospital 1, 2, 3. So we have: P(B1) = 0.6; P(B2) = 0.3; P(B3) =
0.1
• Given the information that a malpractice suit is being filed against one of the
hospitals, what is the probability that hospital is hospital 1 or B1? In other
words, we want to calculate P(B1 given A)
Recap: Bayes Theorem
• Recall our earlier equation of Bayes theorem
• We already know P(A given B1) and P(B1). What we don’t have with us is P(A)
B B’
A P(A∩B) P(A∩B’)
A’ P(A’∩B) P(A’∩B’)
• We have the joint probability of A and B1 as P(A∩B1), the joint probability of A and
B2 as P(A∩B2) and so on
Recap: Bayes Theorem
• We can rewrite the joint probability as follows
B1 B2 B3
A P(A∩B1) P(A∩B2) P(A∩B3)
A’ P(A’∩B1) P(A’∩B2) P(A’∩B3)
• P(A∩B1) = P(A given B1)P(B1); P(A∩B2) = P(A given B2)P(B2); P(A∩B3) = P(A
given B3)P(B3)
• P(A) = P(A given B1)P(B1) + P(A given B2)P(B2) + P(A given B3)P(B3)
• Also, P(A) = P(A given B1)P(B1) + P(A given B2)P(B2) + P(A given B3)P(B3)
Recap: Bayes Theorem
Thus we obtained the expression for conditional probabilities as shown here
• P(B1│A)= (P(A given B1) P(B1)) / P(A given B1)P(B1) + P(A given B2)P(B2)
+ P(A given B3)P(B3)
• P(B2│A)= (P(A given B2) P(B2)) / P(A given B1)P(B1) + P(A given B2)P(B2)
+ P(A given B3)P(B3)
• P(B3│A)= (P(A given B3) P(B3)) / (P(A given B1)P(B1) + P(A given
B2)P(B2) + P(A given B3)P(B3))
Recap: Bayes Theorem
Let us do the numbers now
• Thus, if a malpractice suit takes place, then there is a 35.29 percent chance that
this suit took place against hospital 1
• For any event ‘i’, P(Bi given A) = P(A given Bi) P(Bi) / P(A)
• We can generalize the expression for P(A): (P(A given B1) P(B1) + P(A
given B2) P(B2) + … + P(A given Bn) P(Bn) ), thus we obtain
• P(Bi given A) = P(A given Bi) P(Bi) / (P(A given B1) P(B1) + P(A given
B2) P(B2) + … + P(A given Bn) P(Bn) )
Introduction to Random Variables and
Probability Distributions
Introduction
• In the previous discussions, we developed an understanding of basic
concepts of probability and other concepts such as Bayes’ theorem and its
applications
• In the next set of topics, we will discuss probability distributions and their
properties
• First we will learn about the concept of Random Variables and Probability
Distributions
Introduction
• The next concept that we will see is that of the Expected value and Variance
• We will also discuss a discrete distribution, that is, the binomial distribution
Random Variables: I
Random Variables
• “The house always wins”: One or two people may end up winning large sums
of money, but the machines are designed such that the rest of the people
collectively lose more
• Consider a bag filled with 3 balls, 2 red and 1 blue; Each participant had to
take out a ball, note its color, and then put it back in
• A participant who got a red ball all 4 times would receive 150 rupees; the
participants who got any other result would have to pay 10 rupees
Random Variables
• We’ll approach the problem in three steps
• The one outcome all the players would want is to have all four ball being red
• But there are other possible out comes, what are they?
Random Variables
Let us look at all the possible outcomes
• We could get 4 blue balls - there’s only one outcome in which this happens.
• We could get 3 blue balls and 1 red ball - this could happen in 4 ways - RBBB,
BRBB, BBRB and BBBR.
• We could also get 2 blue balls and 2 red balls - this could happen in 6 ways.
• What is the profit or loss you should expect while running such a business?
• In the previous example, what are the possible outcomes with 3 red balls: X=3
• In statistical terms, this ‘X’ converts possible outcomes to a number and is a random
variable
Random Variables
• Such random variables can be defined in multiple ways
• For example, the number of blue balls that we drew from the bag or the number of
red balls minus the number of blue balls that we have drawn from the bag
• The right way to choose the variable depends upon the nature of information you are
interested in
• If we want to know whether we will win or lose, we need to find different values of X
and their probabilities
Probability Distributions
Probability Distributions
• We defined ‘X’ as the random event and number of red balls drawn
• So, for an outcome where we draw all the four blue balls only, X=0
• There are four possible outcomes if we are to draw one red ball then X=4
• In this manner, we have brought down our 16 outcomes into 5 groups where
the random variable takes the value from 0 to 4
• To find out whether we loose money or win, we need to find the likelihood of
each of these values
Probability Distributions
• The following results are obtained
X Individual runs Probabilities
0 2 0.02
1 7 0.07
2 32 0.32
3 43 0.43
4 16 0.16
Total 100 1
• The probability of any outcome is the number of favorable outcomes divided by the
total number of outcomes
• On the X axis here, we have our 5 random variable values and on the y axis, we have
our probability values
Probability Distributions
• Each value represents the probability of getting a certain number of red balls
• Using the probability distribution, we can answer the question would we in the long
run make money or lose money’
Expected Value
Expected Value
• Remember the red ball game; if somebody played it 1000 times, what will be
their on average win or loss
• In 1000 draws, it is expected that 1 red ball draw will occur 70 times
• Alternatively, we can say that we obtain 2.64 red balls per experiment
• This expected value means that if you play this game infinite number of
times, the average number of red balls per draw you expect is 2.64
• In our case, a more efficient random variable would be the expected amount
won or lost in a game
Expected Value
• So we can define: X = money won after playing the game once
• For this variable, we have two values: 150 for winning (getting 4-red balls)
and -10 for losing (any other outcome)
• Now we know that for winning probability is 0.16 and for losing the
probability is 0.84
Expected Value
• We can compute the expected value as: (150*0.16) + (-10*0.84), which is
equal to +15.6 dollars
• That is, a player would expect to win $15.6 dollars by playing, and therefore,
the game is not profitable for the gambling house
• If the house makes money, it needs to ensure that the expected value won by
the player is negative
• What to do: decrease the prize money, increase the penalty, decrease the
players chances of winning, etc.
Expected Value: Example
• An insurance company estimates the probability that an accident will occur
within the next year is 0.00071. Basis this information, what premium should
the insurance company charge to break even on a $400,000 1-year term
policy?
• On average, the company needs to pay about 284 dollars to settle a policy
• This can also be computed using the probability distribution of the random
variable
• Consider a simple set-up with 2 red and 1 blue balls, what is the probability of
drawing 1, 2, and 3 red balls
• The probability of getting a red ball in one trial is 2/3 and the probability of getting a
blue ball is 1/3
• If you are drawing 4 balls, what is the probability of drawing 4 balls from the bag
Binomial Distribution
• As per the multiplication rule, the probability of events 1 (E1) and 2 (E2) happening
is P(E1)*P(E2)
• Here, we have four events, getting 4 red balls, the probability of the event is: P(Event
1)*P(Event 2)*P(Event 3)*P(Event 4)
• the probability of getting 4 red balls after 4 trials = Probability of getting a red ball in
the first trial * Probability of getting a red ball in the 2nd trial * Probability of getting a
red ball in the 3rd trial * Probability of getting a red ball in the 4th trial = = ⅔*⅔*⅔*⅔
= 0.197
Binomial Distribution
• Remember that we are replacing the red ball again after drawing it
• Consider a case of drawing a blue (B) ball and 3 red (R) balls ‘BRRR’
• 1/3 is the probability of getting a blue ball in 1 trial and 2/3 is the probability of
getting a red ball in 1 trial
• With multiplication rule: P(drawing a blue ball) which would be 1/3 times P(drawing a
red ball) which would be 2/3 times P(drawing a red ball) which would again be 2/3
times P(drawing a red ball) : (1/3)*(2/3)*(2/3)*(2/3) = 0.0987
Binomial Distribution
• There are various possible combinations in which we can get 3 red balls and 1 blue
ball
• There are 4 such sequences in which we get 3 red balls and 1 blue ball- RBRR,
RRBR and RRRB
• This is rule of addition, that is, for independent events E1 and E2, the probability of
E1 or E2 is P(E1)+P(E2)
Binomial Distribution: II
Binomial Distribution
• Consider an event X=1, one red ball and three blue balls
• Let us generalize this case, that is, probability of getting a red ball is p
• For X=2, it is 6 p^2 (1-p)^2; and X=3 it is 4p^3*(1-p); and X=4, it is p^4
• Now let us extend this to a more generic case of making n draws with a sucess
probability of p and r favorable outcomes
• For P(X=r), e.g., r red balls and n-r blue balls, the probability of getting one such
combination is p^r*(1-p)^(n-r)
• Also, there nCr combinations of getting r red balls out of total n balls
nC (p)1(1-p)n-1
1
• The third and final condition is that the probability of success is the same in all trials,
denoted by p
• When all of these conditions are satisfied, then the random variable will follow a
binomial distribution and the probability for X = r, that is, getting r successes in n
trials, can be calculated as nCr(p)r(1-p)n-r, as given by the binomial distribution
Binomial Probability Distribution –
Expected Value and Standard Deviation
Expected Value and Standard Deviation
• In our earlier experiment, where in we got 2.64 red balls per game as the
expected value
• The expected value is nothing but the average value that we would ‘expect’ to get
for a random variable
• E(X)= x1*P(X=x1)+x2*P(X=x2)+x3*P(X=x3)+….+xn*P(X=xn);
• The participant can draw 4 balls, so how many red balls on average, the
participant can draw
• Here n=4 and p=2/3, so E(X)= n*p=4*2/3 =8/3=2.67, i.e., the expected value of
getting a red ball in a game
• What if we wanted to calculate the probability of getting less than equal to 3 red
balls, i.e., P(X=0)+P(X=1)+P(X=2)+P(X=3)= 0.01235 + 0.09877 + 0.2963 + 0.39506
= 0.80247
• That is an 80.2% chance that any randomly selected participant will have selected
maximum 3 red balls while drawing 4 balls
Cumulative Probability
• Any probability where we have to determine the likelihood of X being less than a certain number is
called a cumulative probability
• For example, 0.802 is the cumulative probability for X <= 3. And instead of saying P(X<=3) is 0.802,
we can use F(X = 3) is 0.802, where F represents the cumulative probability
X P(X=x) F(X=x)
• For X=0, F(X=0) is also same 0 =4𝐶0 ∗ 2 0 1 4
=0.01235 0.01
∗
3 3
as P(X<=0), since X<=0 takes 1 =4𝐶1 ∗ 2 1
∗
1 3
=0.09877 0.11
3 3
only one value, X=0, then this 2 =4𝐶2 ∗ 2 2
∗
1 2
=0.2963 0.41
3 3
2 3 1 1
is also same as P(X=0) 3 =4𝐶3 ∗
3
∗
3
=0.3951 0.80
2 4 1 0
4 =4𝐶4 ∗ ∗ =0.1975 1.00
3 3
Cumulative Probability
• Next, let’s calculate F(X = 1). Here, F(X = 1) means P(X<=1), which will be P(X = 0) + P(X = 1), which
will come to 0.11111
• Similarly, we have F(X=1), which will be P(X=0) + P(X=1) and F(X=2), which will be P(X=0) + P(X=1) +
P(X=2)
X P(X=x) F(X=x)
2 0 1 4
• Thus, we can write F(X=2) as F(X=1) + 0 =4𝐶0 ∗ ∗ =0.01235 0.01
3 3
2 1 1 3
P(X=2) 1 =4𝐶1 ∗
3
∗
3
=0.09877 0.11
2 2 1 2
• Hence, F(X=2) will be 0.111 + 0.296= 2 =4𝐶2 ∗
3
∗
3
=0.2963 0.41
2 3 1 1
0.41 3 =4𝐶3 ∗
3
∗
3
=0.3951 0.80
2 4 1 0
4 =4𝐶4 ∗ ∗ =0.1975 1.00
• And F(X=4) will be F(X=3) + P(X = 4) 3 3
• The random variable used to define the continuous probability distribution is called
the continuous random variable
• If you are a pizza delivery manager, you are concerned with the average time it takes
for a pizza delivery to reach a customer. Let us call this delivery time as random
variable ‘X’
• This variable can take on various values like 5-min, 30-min, 15.13-min
• Here defining the variable exact time, is defined as continuous random variable, i.e.,
it can be defined even up to last milliseconds
Continuous random variables
• The ‘amount of water present in a bottle’ or ‘the exact stock price at the end of a
trading session’, or anything that is generally exact, this is always going to be a
continuous random variable
• The values that random variable may take are infinite. For example, even between 20
to 20.1 there can be infinite values taken by a continuous random variable
• Also, if we assume that all the possible delivery times range in between 0 to 1 hour,
the X will range from 0 to 1-hour, and the area of the curve in this range will be 1
Continuous random variables
• First, a continuous random variable is such that it can
have an infinite number of outcomes
• Second, we represent a continuous random variable
using what we call a probability density function which
is essentially a continuous line drawn for all the range
of values that X can take
• Thirdly, the area under the curve, represents the probability that the random
variable lies in that interval
• Finally, the total area under the curve will always be equal to 1
Continuous Probability Distributions: Cumulative
probability for continuous Random Variables (RV)
Cumulative probability for continuous RV
• Consider the following example. The probability
density function of our continuous random variable
is plotted on x-y axis. The x-axis goes from - ∞ to +
∞.
• What is the probability of observing X from -1 to +1, that is area under the cure
from X=-1 to X=+1 (colored region)
• When we talk about cumulative probabilities, we are always dealing with ranges
Cumulative probability for continuous RV
• If you are given that the cumulative probability for
X=1 and X=-1, as 0.6 and 0.4 respectively
• There are many places where the normal distribution appears naturally
• Normal distribution has very convenient and interesting properties that makes it easy
to work with them
• Remember the 30-minute discount scheme to deliver the pizza. If the time is more
than 30-minutes, then pizza is free or available with large discounts
• As a manager at one of the pizza outlets and given this 30-minute guarantee, you
want to ensure that most of the pizzas are delivered well before these 30 minutes
• Let us consider only those commutes from the pizza outlet to the customer’s location
and not include the deliveries where the delivery boy has to visit multiple locations as
he is delivering multiple orders
Normal Distribution
• Let us assume that it takes a maximum of 10
minutes to make the pizza. That leaves us with only
20 minutes to deliver the pizza from the outlet to the
customer’s location
• Just by looking at this distribution, we can make several observations about the
characteristics of this distribution
Normal Distribution
• The mean of this distribution, which in our case is 10
minutes, lies in the exact center of this distribution
or 0.997
Probabilities for a Normal Distribution
• Let us revisit the Pizza delivery time example, assume a normal distribution with μ
• Using our empirical rule 95% of the population lies between μ - 2σ and μ + 2σ, i.e.,
6-18 minutes; in other words 95% probability of observing a delivery time between 6-
18 mins
• Due to symmetry of the normal distribution, we can say that 95%/2=47.5% of the
population lies from μ - 2σ to μ
• Adding the two, we get 97.35% of the population lie between 6-21 mins , i.e., μ -
2σ to μ + 3σ
Probabilities for a Normal Distribution
• What is the probability of observing a delivery in less than 15-mins, i.e., P(X<=15) or
• We can divide the desired area in two parts: (a) 50% of area up to μ=12, and (b) 12-
15 mins (μ to μ + 1σ)
• As noted earlier 68% of the area lies from (μ - 1σ) to (μ + 1σ) or 34% area lies from
μ to μ + 1σ
• Therefore, 50%+34% =84% of pizza deliveries are taking place in less than 15-
minutes
Standard Normal Distribution: Part I
Standard Normal Distribution
• What if you want to find, ‘what is the percentage of deliveries where the commute time
is between 6 and 17 minutes or between 6 and 16.95 minutes
• These values are not as easily identifiable as some of the earlier examples
• Here we can see that μ is 12 and X is 17, and the difference between these two values
should be 5
• if we have to represent X in the form of (μ + some multiple times σ), then we can find
this multiplication factor by using (X-μ) divided by σ
• In our case, (X-μ) / σ comes out to be (17 - 12) / 3, which equates to 5 by 3, which is
around 1.67
Standard Normal Distribution
• The value (X-μ) / σ is denoted by z and z is called the standard normal variable
• Now, the probability of finding X lying between (μ - 2σ) and (μ + 2σ) is same as the
probability of finding the new standard normal random variable z between -2 to +2,
which is 95%
• Similarly, the probability of X lying between (μ - 1σ) and (μ + 1σ) is the same as the
probability of Z lying between -1 and +1, which is 68%
• For a random variable X, we can find the probability of X within a certain range, in the
form of a standard normal variable z= (X-μ) / σ
• (μ + 1σ) will mark to 1, (μ + 2σ) will mark to 2 and, similarly, on the left hand side as
• First, let us convert 17 into the Z value; Remember that the μ and σ for this distribution
was 12 and 3 minutes
• Thus, calculating (X-μ) / σ, we get (17 - 12) / 3, which equates to 5 by 3, which is 1.67
• P(X < = 17) is the same as the probability, P(Z <= 1.67); we are essentially calculating
the cumulative probability for Z = 1.67
Thanks!