Lesson 3-Discreet Probability Distributions
Lesson 3-Discreet Probability Distributions
3
Discreet Probability
Distributions
Random Variables
In sum:
Discrete variables are counted quantities expressed in whole
numbers.
Continuous variables are measured quantities having to do with
length, weight, volume, velocity, etc.
Probability Distributions
Corresponding to the two types of random variables there are two types of
probability distributions:
With a discrete probability distribution, each possible value of the discrete random variable can
be associated with a non-zero probability. Thus, a discrete probability distribution is often
presented in tabular form.
Random Variables
In probability and statistics, a random variable is a variable whose value is subject to variations
due to chance (i.e. randomness, in a mathematical sense). As opposed to other mathematical
variables, a random variable conceptually does not have a single, fixed value (even if unknown);
rather, it can take on a set of possible different values, each with an associated probability.
A random variable’s possible values might represent the possible outcomes of a yet-to-be-
performed experiment, or the possible outcomes of a past experiment whose already-existing
value is uncertain (for example, as a result of incomplete information or imprecise
measurements). They may also conceptually represent either the results of an “objectively”
random process (such as rolling a die), or the “subjective” randomness that results from
incomplete knowledge of a quantity.
Random variables can be classified as either discrete (that is, taking any of a specified list of
exact values) or as continuous (taking any numerical value in an interval or collection of
intervals). The mathematical function describing the possible values of a random variable and
their associated probabilities is known as a probability distribution.
Discrete random variables can take on either a finite or at most a countably infinite set of
discrete values (for example, the integers). Their probability distribution is given by a probability
mass function which directly maps each value of the random variable to a probability. For
example, the value of x1 takes on the probability p1, the value of x2 takes on the probability p2,
and so on. The probabilities pi must satisfy two requirements: every probability pi is a number
between 0 and 1, and the sum of all the probabilities is 1: (p1+p2+⋯+pk=1).
Discrete Probability Distribution: This shows the probability mass function of a discrete probability distribution. The
probabilities of the singletons {1}, {3}, and {7} are respectively 0.2, 0.5, 0.3. A set not containing any of these points
has probability zero.
Examples of discrete random variables include the values obtained from rolling a die and the
grades received on a test out of 100.
Probability distributions for discrete random variables can be displayed as a formula, in a table,
or in a graph.
A discrete random variable x has a countable number of possible values. The probability
distribution of a discrete random variable x lists the values and their probabilities, where
value x1 has probability p1, value x2 has probability x2, and so on. Every probability pi is a
number between 0 and 1, and the sum of all the probabilities is equal to 1.
The number of eggs that a hen lays in a given day (it can’t be 2.3)
The number of people going to a given soccer match
The number of students that come to class on a given day
The number of people in line at McDonald’s on a given day and time
The formula, table, and probability histogram satisfy the following necessary conditions of
discrete probability distributions:
1. 0≤f(x)≤1, i.e., the values of f(x) are probabilities, hence between 0 and 1.
2. ∑f(x)=1, i.e., adding the probabilities of all disjoint cases, we obtain the probability of the
sample space, 1.
Sometimes, the discrete probability distribution is referred to as the probability mass function
(pmf). The probability mass function has the same purpose as the probability histogram, and
displays specific probabilities for each discrete random variable. The only difference is how it
looks graphically.
Probability Mass Function: This shows the graph of a probability mass function. All the values
of this function must be non-negative and sum up to 1.
x f(x)
2 0.2
3 0.3
5 0.5
Discrete Probability Distribution: This table shows the valuesof the discrete random variable
can take on and their corresponding probabilities.
3.2 Cumulative Distribution Functions
You might recall that the cumulative distribution function is defined for discrete random variables
as:
❑
F ( x )=P ( X ≤ x )=∑ f (t)
t≤x
Again, F(x) accumulates all of the probability less than or equal to x. The cumulative distribution
function for continuous random variables is just a straightforward extension of that of the
discrete case. All we need to do is replace the summation with an integral.
x
F ( x )= ∫ f ( t ) dt
−∞
For -∞<x<∞
The expected value of a random variable is the weighted average of all possible values that this
random variable can take on.
A discrete random variable X has a countable number of possible values. The probability
distribution of a discrete random variable X lists the values and their probabilities, such
that xi has a probability of pi. The probabilities pi must satisfy two requirements:
In probability theory, the expected value (or expectation, mathematical expectation, EV, mean,
or first moment) of a random variable is the weighted average of all possible values that this
random variable can take on. The weights used in computing this average are probabilities in
the case of a discrete random variable.
The expected value may be intuitively understood by the law of large numbers: the expected
value, when it exists, is almost surely the limit of the sample mean as sample size grows to
infinity. More informally, it can be interpreted as the long-run average of the results of many
independent repetitions of an experiment (e.g. a dice roll). The value may not be expected in the
ordinary sense—the “expected value” itself may be unlikely or even impossible (such as having
2.5 children), as is also the case with the sample mean.
Suppose random variable X can take value x1 with probability p1, value x2 with probability p2,
and so on, up to value xi with probability pi. Then the expectation value of a random
variable XX is defined as E[X]=x1 p1+ x2 p2+⋯+xi pi, which can also be written as: E[X]=∑xi p1.
If all outcomes xi are equally likely (that is, p 1= p2 =⋯=pi), then the weighted average turns into
the simple average. This is intuitive: the expected value of a random variable is the average of
all values it can take; thus, the expected value is what one expects to happen on average. If the
outcomes xi are not equally probable, then the simple average must be replaced with the
weighted average, which takes into account the fact that some outcomes are more likely than
the others. The intuition, however, remains the same: the expected value of X is what one
expects to happen on average.
For example, let X represent the outcome of a roll of a six-sided die. The possible values
for X are 1, 2, 3, 4, 5, and 6, all equally likely (each having the probability of 1/6). The
expectation of X is:
In this case, since all outcomes are equally likely, we could have simply averaged the numbers
together:
(1 + 2 + 3 + 4 + 5 + 6)/6 = 3.5.
Average Dice Value Against Number of Rolls: An illustration of the convergence of sequence averages
of rolls of a die to the expected value of 3.5 as the number of rolls (trials) grows.
3.4 The Binomial Distribution
Binomial Experiment
Consider the following statistical experiment. You flip a coin 2 times and count the number of
times the coin lands on heads. This is a binomial experiment because:
Binomial Distribution
Suppose we flip a coin two times and count the number of heads (successes). The binomial
random variable is the number of heads, which can take on values of 0, 1, or 2. The binomial
distribution is presented below.
Number of
Probability
Heads
0 0.25
1 0.50
2 0.25
The binomial probability refers to the probability that a binomial experiment results
in exactly x successes. For example, in the above table, we see that the binomial probability of
getting exactly one head in two-coin flips is 0.50.
Given x, n, and P, we can compute the binomial probability based on the binomial formula.
b (x; n, P) = nCx * Px * (1 - P) n - x
or
b (x; n, P) = {n! / [ x! (n - x)!]} * Px * (1 - P) n - x
Example 1
Suppose a die is tossed 5 times. What is the probability of getting exactly 2 fours?
Solution: This is a binomial experiment in which the number of trials is equal to 5, the number of
successes is equal to 2, and the probability of success on a single trial is 1/6 or about 0.167.
Therefore, the binomial probability is:
A cumulative binomial probability refers to the probability that the binomial random variable
falls within a specified range (e.g., is greater than or equal to a stated lower limit and less than
or equal to a stated upper limit).
Example 2
Solution: To solve this problem, we compute 46 individual probabilities, using the binomial
formula. The sum of all these probabilities is the answer we seek. Thus,
b (x < 45; 100, 0.5) = b (x = 0; 100, 0.5) + b (x = 1; 100, 0.5) + . . . + b (x = 45; 100, 0.5)
b (x < 45; 100, 0.5) = 0.184
Example 3
The probability that a student is accepted to a prestigious college is 0.3. If 5 students from the
same school apply, what is the probability that at most 2 are accepted?
Solution: To solve this problem, we compute 3 individual probabilities, using the binomial
formula. The sum of all these probabilities is the answer we seek. Thus,
A Poisson distribution is the probability distribution that results from a Poisson experiment.
Note that the specified region could take many forms. For instance, it could be a length, an
area, a volume, a period of time, etc.
Notation
The following notation is helpful, when we talk about the Poisson distribution.
Poisson Distribution
A Poisson random variable is the number of successes that result from a Poisson experiment.
The probability distribution of a Poisson random variable is called a Poisson distribution.
Given the mean number of successes (μ) that occur in a specified region, we can compute the
Poisson probability based on the following Poisson formula.
Poisson Formula. Suppose we conduct a Poisson experiment, in which the average number of
successes within a given region is μ. Then, the Poisson probability is:
where x is the actual number of successes that result from the experiment, and e is
approximately equal to 2.71828.
The average number of homes sold by the Acme Realty company is 2 homes per day. What is
the probability that exactly 3 homes will be sold tomorrow?
Solution: This is a Poisson experiment in which we know the following:
P (3; 2) = 0.180
A cumulative Poisson probability refers to the probability that the Poisson random variable is
greater than some specified lower limit and less than some specified upper limit.
Suppose the average number of lions seen on a 1-day safari is 5. What is the probability that
tourists will see fewer than four lions on the next 1-day safari?
To solve this problem, we need to find the probability that tourists will see 0, 1, 2, or 3 lions.
Thus, we need to calculate the sum of four probabilities: P (0; 5) + P (1; 5) + P (2; 5) + P (3; 5).
To compute this sum, we use the Poisson formula:
P (x < 3, 5) = [ (e-5) (50) / 0!] + [ (e-5) (51) / 1!] + [ (e-5) (52) / 2!] + [ (e-5) (53) / 3!]
P (x < 3, 5) = 0.2650