Topic 5 Discrete Distributions
Topic 5 Discrete Distributions
Distributions
1
f (x; k) = , for x = x1 , x2 , ..., xk .
k
Note: We have used the notation f (x; k) instead of f (x) to indicate that the
uniform discrete distribution depends on the parameter k.
Example: When a die is tossed, each element of the sample space S =
{1, 2, 3, 4, 5, 6} occurs with probability 16 . Therefore, we have a discrete uni-
form distribution with:
1
f (x; 6) = , for x = 1, 2, 3, 4, 5, 6.
6
Example: Suppose that a student is selected at random from a class of 25 to
1
participate in a certain seminar. Each student has the same probability of 25 of
being selected. If we assume that the students have been numbered in some
way from 1 to 25, the distribution is uniform with:
1
f (x; 25) = , for x = 1, 2, ..., 25.
25
1
2 Elements of Probability & Statistics– MTH1202
That is, the expectation of a Bernoulli random variable is the probability that the
random variable assumes a value of 1 (success). Also,
E(X2 ) = 12 · p + 02 · q = p
and
for x = 0, 1, 2,
! ..., n.
n n!
where = =n Cx with n being the number of trials, x the
x x!(n − x)!
number of successes in the n trials and p the probability of success on a single
trial.
If X is a random variable that is binomially distributed with parameters
n and p, we write X ∼ binomial(n, p). Note that, by the binomial theorem, the
probabilities sum to 1, that is,
n n !
X X n
P(X = x) = px qn−x = (p + q)n = 1
x
x=0 x=0
Example 1 A coin is tossed 5 times. Find the probability that exactly 3 heads will
appear.
Soln: Let X be the random variable representing the number of heads that will appear.
X ∼ binomial(n, p), where n = 5 and p = 12 .
!
n
P(X = x) = px (1 − p)n−x , x = 0, 1, 2, 3, 4, 5
x
Thus
! 3 2
5 1 1 5
P(exactly three heads) = P(X = 3) = =
3 2 2 16
a−1
X
2. P(X < a) = P(X = x)
x=0
Example 2 The probability that a patient recovers from a rare blood disease is 0.4. If
15 people are known to have contracted the disease, what is the probability that:
Soln:
Let X be the number of people that survive. X ∼ binomial(n, p), where n = 15 and
p = 0.4 and
!
n
P(X = x) = px (1 − p)n−x , x = 0, 1, . . . , 15
x
We can use tables to answer this question as well, look at the Cumulative Binomial
table at the back.
5 4
(i) P(X = 5) = b(x; 15, 0.4) = 0.4032 − 0.2173 = 0.1859
P P
b(x; 15, 0.4) −
x=0 x=0
9
(ii) P(X ≥ 10) = 1 − P(X < 10) = 1 − b(x; 15, 0.4) = 1 − 0.9662 = 0.0338
P
x=0
Example 3 A fair coin is tossed 14 times. What is the probability that the head will
appear:
Example 4 A new car dealer knows from past experience that on the average, she will
make sale to about 20% of her customers. What is the probability that in five randomly
selected presentations, she makes a sale to:
E(X) = np
Proof: The first method is by using the meaning of a Bernoulli trial: If the
random variables X1 , X2 , . . . , Xn form n Bernoulli trials with parameter p and if
X = X1 + X2 + · · · + Xn , then X has a binomial distribution with parameters n
and p.
Recall that E(Xi ) = p for i = 1, 2, . . . , n. Thus
E(X) =E(X1 + X2 + · · · + Xn )
=E(X1 ) + E(X2 ) + · · · + E(Xn )
= p + p + · · · + p = np
| {z }
n times
6 Elements of Probability & Statistics– MTH1202
n
X
E(X) = xP(X = x)
x=0
n !
X n
= x px qn−x
x
x=0
n
X n!
= x px qn−x
x=0
x!(n − x)!
n
X n!
= px qn−x
(x − 1)!(n − x)!
x=1
n
X (n − 1)!
=np px−1 qn−x
(x − 1)!(n − x)!
x=1
n !
X n−1
=np px−1 qn−x
x−1
x=1
=np(p + q)n−1 = np
This can be proved in two different ways. First from the definition of a Bernoulli
trial. Recall that for a Bernoulli random variable Xi , Var(Xi ) = pq. Since a
binomial random variable X = X1 + X2 + · · · + Xn , then
Var(X) =Var(X1 + X2 + · · · + Xn )
=Var(X1 ) + Var(X2 ) + · · · + Var(Xn )
= pq + pq + · · · + pq = npq
| {z }
n times
√
Thus Var(X) = npq and the standard deviation of X is npq.
In the second method, you use the definition of variance. Recall that E(X) = np
and from Var(X) = E(X2 ) − [E(X)]2 , we seek to find E(X2 ). Also recall from the
Binomial expansion that
n !
X n
(r + s)n = rx sn−x
x
x=0
c
B.K. Nannyonga, H.W. Kayondo, N. Muyinda, & B.N. Kirenga
7
Now
n
X
E(X(X − 1)) = x(x − 1)P(X = x)
x=0
n !
X n
= x(x − 1) px qn−x
x
x=0
n
X n!
= x(x − 1) px qn−x
x=0
x!(n − x)!
n
X n!
= px qn−x
x=2
(x − 2)!(n − x)!
n
X(n − 2)!
=n(n − 1)p2 px−2 qn−x
x=2
(x − 2)!(n − x)!
n !
X n−2
=n(n − 1)p2 px−2 qn−x
x−2
x=2
=n(n − 1)p (p + q)n−2
2
=n(n − 1)p2
But
E(X(X − 1)) = E(X2 − X) = E(X2 ) − E(X)
Thus
E(X2 ) = E(X(X − 1)) + E(X) = n(n − 1)p2 + np
Therefore, 7
Example 5 In a family the probability of having a boy is 0.8. If there are seven children
in the family determine the:
(a) expected number of boys,
(b) expected number of girls,
(c) probability that at least four children are girls,
(d) probability that they are all boys.
Solution:
8 Elements of Probability & Statistics– MTH1202
(b) let Ybe the random variable representing the number of girls.
Then Y ∼ binomial(n, p) with n = 7 and p = 0.2. E(Y) = 7 ∗ 0.2 = 1.4
Example 6 A multiple choice quiz has ten questions each with four alternative answers
of which one is correct. Determine the expected number of correct answers and the
variance, if someone does the quiz by sheer guesswork.
Solution:
Let X be the random variable representing the number of correct answers. Then
X ∼ binomial(n, p) with n = 10 and p = 14 .
1
E(X) = np = 10 ∗ = 2.5
4
and
Var(X) = npq = 10 ∗ 1/4 ∗ 3/4 = 1.875
Example 7 A multiple-choice exam consist of twenty questions each with five alter-
native answers of which only one is correct. Five marks are awarded for each answer
and one mark is subtracted for each incorrect or un attempted question. If the exam is
answered by sheer guesswork, find the:
Solution:
Let X be the random variable representing the number of correct answers and Y the
number of incorrect answers. Then, with n = 20, p = 51 = 0.2, q = 45 = 0.8
(ii) The expected mark for correct answers =4∗5 = 20. The expected mark for incorrect
answers =16 ∗ 1 = 16. Therefore, the expected overall mark is 20 − 16 = 4 marks.
Example 8 If X is a binomially distributed with p = 0.45 and n = 10, find the most
likely value of X to occur.
Solution:
First compute the expected value of X, that is E(X) = np = 10 ∗ 0.45 = 4.5. When this
c
B.K. Nannyonga, H.W. Kayondo, N. Muyinda, & B.N. Kirenga
9
is done, try to find the probabilities of values around the mean, in this case 3, 4, 5, 6 as
below.
!
10
P(X = 3) = (0.45)3 (0.55)7 = 0.1664
3
!
10
P(X = 4) = (0.45)4 (0.55)6 = 0.2383
4
!
10
P(X = 5) = (0.45)5 (0.55)5 = 0.2340
5
!
10
P(X = 6) = (0.45)6 (0.55)4 = 0.1595
6
From this you note that X = 4 has the highest probability and as such the most occuring
value, or the mode is 4.
Thus the most likely number of people with eye disorder is 10 since it has the highest
probability.
Example 11 30% of college students own video cassette recorders. The Telektronic
Company produced a videotape and sent pilot copies to 20 college students. What is the
mean and standard deviation of this sample?
Solution:
Once again -this is binomial, so µ = 20 ∗ 0.3 = 6, σ2 = 20 ∗ 0.3 ∗ 0.7 = 4.2, and
√
σ = 4.2 = 2.04939.
Example 12 In the testing of a product from a production line, the probability that the
product survives the test is 34 . What is the probability that exactly two of the next four
products tested will also survive the test?
Solution
Assuming that the tests are independent, then X ∼ binomial(4, 43 ). Therefore
! 2 2
4 3 1 27
P(X = 2) = = .
2 4 4 128
e−λ λx
f (x, λ) = P(X = x) = , x = 0, 1, 2, . . . (5.3)
x!
The symbol e stands for a constant approximately equal to 2.7183. It is a famous
constant in mathematics, named after the Swiss mathematician L. Euler, and it
is also the base of the so-called natural logarithm.
It is clear that f (x) ≥ 0 for each value of x. In order to verify that the function
f (x, λ) defined in equation (5.3) satisfies the requirements of every p.m.f , it must
∞
X
be shown that f (x, λ) = 1 . It is known from elementary calculus that for any
x=0
real number λ,
∞
X λx
eλ =
x=0
x!
Therefore,
∞ ∞
X X λx
f (x, λ) = e −λ
= e−λ eλ = 1
x=0 x=0
x!
since the term corresponding to x = 0 in this series is 0, we can omit this term
and can begin the summation with the term x = 1. Therefore,
∞ ∞ ∞
X X e−λ λx X e−λ λx−1
E(X) = x f (x, λ) = x =λ
x! (x − 1)!
x=1 x=1 x=1
If we let y = x − 2, we obtain
∞
X e−λ λ y
E[X(X − 1)] = λ2 = λ2 (5.4)
y=0
y!
Since E[X(X − 1)] = E(X2 ) − E(X) = E(X2 ) − λ, it follows from Eq. (5.4) that
E(X2 ) = λ2 + λ. Therefore
Var(X) = E(X2 ) − [E(X)]2 = λ2 + λ − λ2 = λ
12 Elements of Probability & Statistics– MTH1202
Thus, for a Poisson distribution for which the p.m.f is defined by Eq. (5.3), we
have established the fact that both the mean and variance are equal to λ.
Example 13 Suppose that the average number of accidents occurring weekly on a
particular stretch of a highway equals 3. Calculate the probability that there is at least
one accident this week.
Soln: Let X denote the number of accidents occurring on the stretch of highway in
question during this week. Now, X ∼ Poisson(λ), where λ = 3.
e−3 30
P(X ≥ 1) = 1 − P(X = 0) = 1 − = 1 − e−3 = 0.9502
0!
Example 14 Consider an experiment that consists of counting the number of α particles
given off in a one-second interval by one gram of radioactive material. If we know from
past experience that, on the average, 3.2 such α-particles are given off, what is a good
approximation to the probability that no more than 2 α-particles will appear?
Soln: The number of α-particles given off will be a Poisson random variable with
parameter λ = 3.2. Hence the desired probability is
(3.2)2 −3.2
P(X ≤ 2) = e−3.2 + 3.2e−3.2 + e = 0.380
2
Example 15 If the average number of claims handled daily by an insurance company
is 5, what proportion of days have less than 3 claims?
Soln: Suppose that the number of claims handled daily, call it X, is a Poisson random
variable. Since E(X ) = 5, the probability that there will be fewer than 3 claims on any
given day is
51 52
P(X < 3) = P(X = 0) + P(X = 1) + P(X = 2) = e−5 + e−5 + e−5 = 0.1247
1! 2!
Example 16 The average number of field mice per acre of a field is estimated to be 10.
Find the probability that a given acre will contain more than 15 mice.
Soln: Let X be the number of field mice, X ∼ Poisson(10).
P(X > 15) = 1 − P(X ≤ 15)
Note: We can use cumulative Poisson table to get the answer easily. Look at the
table at end of this Topic.
P(X > 15) = 1 − P(X ≤ 15) = 1 − 0.9513 = 0.0487.
Example 17 Suppose that radioactive particles strike a certain target in accordance
with a Poisson process at an average rate of 3 particles per minute. Determine the
probability that 10 or more particles will strike the target in a particular 2-minute
period.
Soln: In a Poisson process, the number of particles striking the target in any particular
one-minute period has a Poisson distribution with mean λ. since the mean number of
strikes in any one-minute period is 3, it follows that λ = 3. Therefore, the number of
strikes X in any 2-minute period will have a Poisson distribution with mean 6. It can
be found from the table of the Poisson distribution that P(X ≥ 10) = 0.0838.
c
B.K. Nannyonga, H.W. Kayondo, N. Muyinda, & B.N. Kirenga
13
Example 18 Telephone calls enter a school switchboard on average of two every three
minutes. What is the probability of at least 5 calls arriving in a 9-minute period?
Soln: Let X be the number of calls arriving in a 9-minute period. X ∼ Poisson(λ),
where λ = 6.
P(X ≥ 5) = 1 − P(X ≤ 4) = 1 − 0.285 = 0.715
e−λ1 λx1
Proof 1 If X ∼ Poisson(λ1 ), then P(X = x) = so that
x!
P(X = 0) =e−λ1
P(X = 1) =e−λ1 · λ1
e−λ1 · λ21
P(X = 2) =
2!
y
e−λ2 λ2
Also if Y ∼ Poisson(λ2 ), then P(Y = y) = so that
y!
P(Y = 0) =e−λ2
P(Y = 1) =e−λ2 · λ2
e−λ2 · λ22
P(Y = 2) =
2!
If X and Y are two independent Poisson random variables, then you will have
x+y 0 1 2 ···
−(λ1 +λ2 ) −(λ1 +λ2 ) −(λ1 +λ2 ) (λ1 +λ2 )2
P(X + Y = x + y) e e · (λ1 + λ2 ) e · 2! ···
Example 19 A company has 2 machines A and B. On average there are 0.8 breakdowns
per week on machine A and 1.2 breakdowns on machine B. What is the probability of
14 Elements of Probability & Statistics– MTH1202
Thus the smallest integer n such that the probability that there are more than n
misprints on the center page is less than 0.1 is n = 7.
Example 23 A certain kind of sheet metal has on average five defects per 10 square feet.
If we assume a Poisson distribution, what is the probability that a 15 square foot sheet
of metal will have at least six defects? Solution: Let X denote the number of defects
in a 15 square foot sheet of the metal. Then, since the unit of area is 10 square feet,
t=15/10=1.5 this implies that λ0 = λt = 5 ∗ 1.5 = 7.5. Thus P(X ≥ 6) = 1 − P(X ≤
5) = 1 − 0.2414 = 0.7586.
We shall now show that when the value of n is large and the value of p is close
to 0, the binomial distribution with parameters n and p can be approximated by
a Poisson distribution with mean np. This approximation is most adequate for
practical purposes and quite accurate if n ≥ 20 and p ≤ 0.05 and it is very good
if n ≥ 100 and np ≤ 10.
Suppose that a random variable X has a binomial distribution with param-
eters n and p, and let P(X = x) = f (x, n, p), for any given value of x. Then for
x = 1, 2, . . . , n,
n(n − 1) · · · (n − x + 1) x
!
n
f (x, n, p) = px (1 − p)n−x = p (1 − p)n−x
x x!
λx n n − 1 n−x+1 λ n λ −x
f (x, n, p) = · ··· 1− 1−
x! n n n n n
We shall now let n → ∞ and p → 0 in such a way that the value of the product
np remains equal to the fixed value λ throughout this limiting process. Since
the values of λ and x are held as fixed as n → ∞, then
n n−1 n−x+1 λ −x
lim · ··· 1− =1
n→∞ n n n n
Furthermore, it is known from elementary calculus that
λ n
lim 1 − = e−λ
n→∞ n
It now follows that for any nonnegative integer x,
e−λ λx
f (x, n, p) →
x!
The expression on the right is the p.m.f f (x, λ) of the Poisson distribution with
mean λ. Therefore when n is large and p is close to zero, the value of the p.m.f
f (x, n, p) of the binomial distribution can be approximated, x = 0, 1, 2, . . ., by the
value of the p.m.f f (x, λ) of the Poisson distribution for which λ = np.
16 Elements of Probability & Statistics– MTH1202
Example 24 A factory packs bolts in boxes of 500. The probability that a bolt is
defective is 0.002. Find the probability that a box contains 2 defective bolts. Solution
Let X be the number of defective bolts in a box. Then X is binomially distributed with
n = 500 and p = 0.002. This problem solved as a pure binomial distribution has the
following solution.
!
500
P(X = 2) = (0.002)2 (0.988)498 = 0.184
2
By Poisson approximation, note that n = 500 is large and p = 0.002 is small. Also
λ = np = 500 ∗ 0.002 = 1. This
e−λ λx e−1
P(X = x) = ⇒ P(X = 2) = = 0.184
x! 2!
Both answers agree to 3 decimal places, but may seem much easier in Poisson approxi-
mation.
Example 25 A manufacturer of Christmas tree light bulbs knows that 2% of his are
defective. Find the probability that a box of 100 bulbs contains at most 3 defective bulbs.
Solution Since n = 100 (large) and p = 2% = 0.02 (small) then you approximate with
Poisson distribution, with λ = np = 100 ∗ 0.02 = 2. Thus
3
X
P(X ≤ 3) = p(x; 2) = 0.857
x=0
*from tables.
Using the binomial distribution we have
3
X
P(X ≤ 3) = b(x; 100, 0.02)
x=0
compute!!
Example 26 Suppose that on average 1 person in 1000 is HIV positive. Find the
probability that a random sample of 8000 people will yield less than 7 HIV-positive
cases. Solution This is a binomial distribution problem where n = 8000 and p = 0.001.
Since n is quite large and p is close to zero, you approximate it with a Poisson,
λ = np = 8000 ∗ 0.001 = 8 and so
6
X 6
X
P(X < 7) = b(x; 8000, 0.001) = p(x; 8) = 0.3134
x=0 x=0
from tables.
c
B.K. Nannyonga, H.W. Kayondo, N. Muyinda, & B.N. Kirenga
17
1. the outcome of each trial can be classified into one of the two categories,
success or failure,
Now
∞
X ∞
X ∞
X
g(x, p) = pqx−1 = p qx−1
x=1 x=1 x=1
1. P(X = r) = pqr−1
1
E(X) =
p
c
B.K. Nannyonga, H.W. Kayondo, N. Muyinda, & B.N. Kirenga
19
Proof:
X
E(X) = xP(X = x)
allx
X∞
= xpqx−1
x=1
X∞
=p xqx−1
x=1
∞ ∞
X d x d X x
=p (q ) = p q
dq dq
x=1 x=1
!
d q p p 1
=p = 2
= 2 =
dq 1 − q (1 − q) p p
Thus,
2q 2q 1 2 1
E[X(X − 1)] = E(X2 ) − E(X) = 2
⇒ E(X2 ) = 2 + = 2 −
p p p p p
20 Elements of Probability & Statistics– MTH1202
Example 27 Each time a player plays a certain card game called "patience", there is a
probability of 0.15 that the game will work out successfully. The probability of success
in any game is independent of the outcome of any other game. A player plays games of
patience until a game works out successfully.
Solution:
The number of games played, X, has a geometric distribution.
X ∼ Geo(0.15).
a) If she plays 4 games altogether she must lose the first 3 games and be successful
on the 4th game. So
1 1
c) E(X) = = = 6.67
p 0.15
Example 28 If 40% of people in a certain population have blood type A, find the
probability that in a sequential testing of randomly selected people, the first person with
type A blood is found on the fourth test.
Solution:
Let X be the number of tests up to the first person of type A blood. Then X ∼ Ge0(0.4).
We wish to find P(X = 4) = (0.6)3 (0.4)
c
B.K. Nannyonga, H.W. Kayondo, N. Muyinda, & B.N. Kirenga
21
Example 31 In a particular game, a player can get out of check only if she gets two
heads when she tosses two coins.
(a) Find the probability that more than 6 attempts are needed to get out of check.
(b) What is the smallest value of n if there is to be at least a 90% chance of getting
out of check on or before the n − th attempt.
Solution:
P(2heads when two coins are tossed)= 14 . So P(success) = 1
4 and q = 34 . If X= the
number of attempts required to get out check, then
(a) P(X > 6) = q6 = ( 34 )6 = 0.17
log10 0.1
n log10 0.75 ≥ log10 0.1 ⇒ n ≥ = 8.0039
log10 0.75
Example 32 The probability that a marksman hits a target eye is 0.4 for each shot,and
each shot is independent of the other. Find the:
22 Elements of Probability & Statistics– MTH1202
(a) probability that he hits the target for the first time on his fourth attempt,
(b) mean number of shots needed to hit the target and standard deviation,
Solution: Let X be the number of attempts up to and including the first target. Then
P(X = x) = pqx−1 , x = 1, 2, 3, ...; with p = 0.4 and q = 0.6 .
Therefore, the most likely number of shots is 1, since the probabilities are decreasing and it
is the one with the highest probability. Generally it is noted that p > qp > q2 p > q3 p > ...
so the mode of the geometric distribution (or the most likely value) occurs at X = 1 with
probability P(X = 1) = p.
k+y−1
!
P(Y = y) = pk q y , y = 0, 1, . . .
y
Example: Find the probability that a person tossing 3 coins will get either
all heads or all tails for the second time on the fifth toss.
Solution
Using the Negative Binomial with x = 5, k = 2 and p = 14 , we have:
c
B.K. Nannyonga, H.W. Kayondo, N. Muyinda, & B.N. Kirenga
23
!
5−1
b∗ (5; 2, 0.25) = 0.252 0.753 = 0.1445
2−1
Example: An oil company conducts a geological study that indicates that
an exploratory oil well should have a 20% chance of striking oil. What is the
probability that the first strike comes on the third well drilled?
Solution:
1. a smaller batch:
Example 33 A batch of 100 piston rings is known to contain 10 defective rings. If two
piston rings are drawn from the batch, write down the probabilities that:
2. the second ring is defective given that the first one is defective.
Soln:
24 Elements of Probability & Statistics– MTH1202
10 1
1. The probability that the first ring is defective is clearly = .
100 10
2. Assuming that the first ring selected is defective and we do not replace it, the
9
probability that the second ring is defective is equally clearly = 1
99 11
The Hypergeometric distribution may be thought of as arising from sam-
pling from a batch of items where the number of defective items contained in
the batch is known.
Essentially the number of defectives contained in the batch is not a random variable, it
is fixed.
The calculations involved when using the Hypergeometric distribution are
usually more complex than their Binomial counterparts. If we sample without
replacement we may proceed in general as follows:
Example 35 A committee of size 5 is to be selected at random from three men and five
women. Find the probability distribution for the number of:
Soln:
(i) Let the variable X be the number of men on the committee. X has a hypergeometric
distribution in which N = 8, n = 5, M = 3 and r = 0, 1, 2, 3. Thus,
3
C0 ×5 C5 1
P(X = 0) = 8C
=
5 56
3
C1 ×5 C4 15
P(X = 1) = 8 =
C5 56
3
C2 ×5 C3 30
P(X = 2) = 8 =
C5 56
3
C3 ×5 C2 10
P(X = 3) = 8 =
C5 56
x 0 1 2 3
1 15 30 10
P(X=x) 56 56 56 56
(ii) If we let Y to be the number of women on the committee, the distribution can
easily be seen from (i) as:
y 2 3 4 5
10 30 15 1
P(Y=y) 56 56 56 56
Note: The committee cannot have one or no woman as this will require to have more
than three men given that the committee must be composed of five individuals.
Example 36 Boxes of 40 light bulbs are called acceptable if they contain no more than
3 defectives. The procedure of sampling the box is to select 5 bulbs at random and reject
a box if a defective bulb is found. What is the probability that exactly one defective bulb
will be found in a sample if there are three defective bulbs in the entire box?
M
Cr ×N−M Cn−r
Soln: Using P(X = r) = NC
, we know that N = 40, M = 3, n = 5 and
n
r = 1.
3
C1 ×37 C4
Hence P(X = 1) = 40 C
= 0.3011
5
It is possible to derive formulae for the mean and variance of the Hyper-
geometric distribution. However, the calculations are more difficult than their
Binomial counterparts.
26 Elements of Probability & Statistics– MTH1202
M
Cx ×N−M Cn−x
P(X = x) = NC
, x = 0, 1, . . . , n
n
N−n M
E(X) = np and Var(X) = np(1 − p) where p =
N−1 N
Proof: From the definition of expectation of a discrete random variable
n M
X X Cx ×N−M Cn−x
E(X) = xP(X = x) = x NC
x=0 n
allx
n N−M
X M! Cn−x
= x NC
x!(M − x)! n
x=1
n N−M
X (M − 1)! Cn−x
=M N
(x − 1)!(M − x)! Cn
x=1
n M−1 N−M
X C x−1 × Cn−x
=M NC
x=1 n
Letting y = x − 1, we have
n−1 M−1
X C y ×N−M Cn−1−y
E(X) = M NC
y=0 n
Writing
N N−1
N−M
Cn−1−y =(N−1)−(M−1) Cn−1−y and N Cn = × Cn−1
n
we have
n−1 M−1
nM X C y ×(N−1)−(M−1) Cn−1−y
E(X) = N−1 C
N y=0 n−1
The expression inside the summation represents the total of all probabilities of a
hypergeometric experiment when a sample of n − 1 items is selected at random
from N − 1 items, of which M − 1 are defective, and hence equals to 1. Therefore,
the expectation is
nM
E(X) =
N
c
B.K. Nannyonga, H.W. Kayondo, N. Muyinda, & B.N. Kirenga
27
To compute the variance, we first derive the expression for E(X(X − 1)). Thus
from the definition of the expectation, we have
n M
X X Cx ×N−M Cn−x
E[X(X − 1)] = x(x − 1)P(X = x) = x(x − 1) NC
x=0 n
allx
n N−M
X (M − 2)! Cn−x
=M(M − 1)
x=2
(x − 2)!(M − x)! N Cn
n−2
n(n − 1)M(M − 1) X M−2 C y ×(N−2)−(M−2) Cn−2−y
= N−2 C
N(N − 1) y=0 n−2
The expression inside the summation is the sum of (n − 2) terms of the hyperge-
ometric distribution and is equal to 1. Hence
M(M − 1)
E[X(X − 1)] = E(X2 ) − E(X) = n(n − 1)
N(N − 1)
Thus
M(M − 1) nM
E(X2 ) = n(n − 1) +
N(N − 1) N
5.7.2 Applications
The hypergeometric distribution is used for calculating probabilities for samples
drawn from relatively small populations and without replication, which means
that an item’s choice of being selected increases on each trial.
For sampling with replication, in which case an item has an equal chance of
being selected, the binomial distribution should be used. If N >> n, in which
case the population size of interest is too large compared to the sample size,
f ≈ 1 (the finite population correct can be neglected). The probabilities from
a hypergeometric distribution and those of a binomial distributions are no
different.
Example 37 In the manufacture of car tyres, a particular production process is known
to yield 10 tyres with defective walls in every batch of 100 tyres produced. From a
production batch of 100 tyres, a sample of 4 is selected for testing to destruction. Find:
!
r
Table A.1 (continued) Binomial Probability Sums b(x; n, p)
x=0
p
n r 0.10 0.20 0.25 0.30 0.40 0.50 0.60 0.70 0.80 0.90
15 0 0.2059 0.0352 0.0134 0.0047 0.0005 0.0000
1 0.5490 0.1671 0.0802 0.0353 0.0052 0.0005 0.0000
2 0.8159 0.3980 0.2361 0.1268 0.0271 0.0037 0.0003 0.0000
3 0.9444 0.6482 0.4613 0.2969 0.0905 0.0176 0.0019 0.0001
4 0.9873 0.8358 0.6865 0.5155 0.2173 0.0592 0.0093 0.0007 0.0000
5 0.9978 0.9389 0.8516 0.7216 0.4032 0.1509 0.0338 0.0037 0.0001
6 0.9997 0.9819 0.9434 0.8689 0.6098 0.3036 0.0950 0.0152 0.0008
7 1.0000 0.9958 0.9827 0.9500 0.7869 0.5000 0.2131 0.0500 0.0042 0.0000
8 0.9992 0.9958 0.9848 0.9050 0.6964 0.3902 0.1311 0.0181 0.0003
9 0.9999 0.9992 0.9963 0.9662 0.8491 0.5968 0.2784 0.0611 0.0022
10 1.0000 0.9999 0.9993 0.9907 0.9408 0.7827 0.4845 0.1642 0.0127
11 1.0000 0.9999 0.9981 0.9824 0.9095 0.7031 0.3518 0.0556
12 1.0000 0.9997 0.9963 0.9729 0.8732 0.6020 0.1841
13 1.0000 0.9995 0.9948 0.9647 0.8329 0.4510
14 1.0000 0.9995 0.9953 0.9648 0.7941
15 1.0000 1.0000 1.0000 1.0000
16 0 0.1853 0.0281 0.0100 0.0033 0.0003 0.0000
1 0.5147 0.1407 0.0635 0.0261 0.0033 0.0003 0.0000
2 0.7892 0.3518 0.1971 0.0994 0.0183 0.0021 0.0001
3 0.9316 0.5981 0.4050 0.2459 0.0651 0.0106 0.0009 0.0000
4 0.9830 0.7982 0.6302 0.4499 0.1666 0.0384 0.0049 0.0003
5 0.9967 0.9183 0.8103 0.6598 0.3288 0.1051 0.0191 0.0016 0.0000
6 0.9995 0.9733 0.9204 0.8247 0.5272 0.2272 0.0583 0.0071 0.0002
7 0.9999 0.9930 0.9729 0.9256 0.7161 0.4018 0.1423 0.0257 0.0015 0.0000
8 1.0000 0.9985 0.9925 0.9743 0.8577 0.5982 0.2839 0.0744 0.0070 0.0001
9 0.9998 0.9984 0.9929 0.9417 0.7728 0.4728 0.1753 0.0267 0.0005
10 1.0000 0.9997 0.9984 0.9809 0.8949 0.6712 0.3402 0.0817 0.0033
11 1.0000 0.9997 0.9951 0.9616 0.8334 0.5501 0.2018 0.0170
12 1.0000 0.9991 0.9894 0.9349 0.7541 0.4019 0.0684
13 0.9999 0.9979 0.9817 0.9006 0.6482 0.2108
14 1.0000 0.9997 0.9967 0.9739 0.8593 0.4853
15 1.0000 0.9997 0.9967 0.9719 0.8147
16 1.0000 1.0000 1.0000 1.0000
30 Elements of Probability & Statistics– MTH1202
!
r
Table A.2 (continued) Poisson Probability Sums p(x; µ)
x=0
µ
r 10.0 11.0 12.0 13.0 14.0 15.0 16.0 17.0 18.0
0 0.0000 0.0000 0.0000
1 0.0005 0.0002 0.0001 0.0000 0.0000
2 0.0028 0.0012 0.0005 0.0002 0.0001 0.0000 0.0000
3 0.0103 0.0049 0.0023 0.0011 0.0005 0.0002 0.0001 0.0000 0.0000
4 0.0293 0.0151 0.0076 0.0037 0.0018 0.0009 0.0004 0.0002 0.0001
5 0.0671 0.0375 0.0203 0.0107 0.0055 0.0028 0.0014 0.0007 0.0003
6 0.1301 0.0786 0.0458 0.0259 0.0142 0.0076 0.0040 0.0021 0.0010
7 0.2202 0.1432 0.0895 0.0540 0.0316 0.0180 0.0100 0.0054 0.0029
8 0.3328 0.2320 0.1550 0.0998 0.0621 0.0374 0.0220 0.0126 0.0071
9 0.4579 0.3405 0.2424 0.1658 0.1094 0.0699 0.0433 0.0261 0.0154
10 0.5830 0.4599 0.3472 0.2517 0.1757 0.1185 0.0774 0.0491 0.0304
11 0.6968 0.5793 0.4616 0.3532 0.2600 0.1848 0.1270 0.0847 0.0549
12 0.7916 0.6887 0.5760 0.4631 0.3585 0.2676 0.1931 0.1350 0.0917
13 0.8645 0.7813 0.6815 0.5730 0.4644 0.3632 0.2745 0.2009 0.1426
14 0.9165 0.8540 0.7720 0.6751 0.5704 0.4657 0.3675 0.2808 0.2081
15 0.9513 0.9074 0.8444 0.7636 0.6694 0.5681 0.4667 0.3715 0.2867
16 0.9730 0.9441 0.8987 0.8355 0.7559 0.6641 0.5660 0.4677 0.3751
17 0.9857 0.9678 0.9370 0.8905 0.8272 0.7489 0.6593 0.5640 0.4686
18 0.9928 0.9823 0.9626 0.9302 0.8826 0.8195 0.7423 0.6550 0.5622
19 0.9965 0.9907 0.9787 0.9573 0.9235 0.8752 0.8122 0.7363 0.6509
20 0.9984 0.9953 0.9884 0.9750 0.9521 0.9170 0.8682 0.8055 0.7307
21 0.9993 0.9977 0.9939 0.9859 0.9712 0.9469 0.9108 0.8615 0.7991
22 0.9997 0.9990 0.9970 0.9924 0.9833 0.9673 0.9418 0.9047 0.8551
23 0.9999 0.9995 0.9985 0.9960 0.9907 0.9805 0.9633 0.9367 0.8989
24 1.0000 0.9998 0.9993 0.9980 0.9950 0.9888 0.9777 0.9594 0.9317
25 0.9999 0.9997 0.9990 0.9974 0.9938 0.9869 0.9748 0.9554
26 1.0000 0.9999 0.9995 0.9987 0.9967 0.9925 0.9848 0.9718
27 0.9999 0.9998 0.9994 0.9983 0.9959 0.9912 0.9827
28 1.0000 0.9999 0.9997 0.9991 0.9978 0.9950 0.9897
29 1.0000 0.9999 0.9996 0.9989 0.9973 0.9941
30 0.9999 0.9998 0.9994 0.9986 0.9967
31 1.0000 0.9999 0.9997 0.9993 0.9982
32 1.0000 0.9999 0.9996 0.9990
33 0.9999 0.9998 0.9995
34 1.0000 0.9999 0.9998
35 1.0000 0.9999
36 0.9999
37 1.0000