Stat
Stat
1. Flipping A Coin Twice. Let random variable X be the number of heads that
come up.
(e) True / False If heads come up twice as often as tails, the distribution of
X is then
x 0 1 2
1 1 1 1 2 4 2 2 4
P (X = x) 3
× 3
= 9
2× × 3 3
= 9 3
× 3
= 9
Fall, 2002.
58 Chapter 4. Random Variables
S
6 x x x x x x
5 x x x x x x
4 x x x x x x
Die 2
3 x x x x x x
2 x x x x x x
1 x x x x x x
1 2 3 4 5 6
Die 1
x 8 9 10 11 12
5 4 3 2 1
P (X = x) 36 36 36 36 36
3. Flipping Until a Head Comes Up. A (weighted) coin has a probability of p = 0.7
of coming up heads (and so a probability of 1 − p = 0.3 of coming up tails).
This coin is flipped until a head comes up or until a total of 4 flips are made.
Let X be the number of flips.
Section 1. Random Variables 59
0 3 6 9 12 15 18 21 24 27 30 33 36
2 5 8 11 14 17 20 23 26 29 32 35
00
1 4 7 10 13 16 19 22 25 28 31 34
first section second section third section
(a) The sample space consists of 38 outcomes: {00, 0, 1, . . . , 35, 36}. The
event “an even comes up” (numbers 2, 4, 6, . . . , 36, but not 0 or 00)
consists of (circle one) 18 / 20 / 22 numbers.
The chance an even comes up is then (circle one) 18/38 / 20/38 / 22/38
The event “a number in the second section comes up” (12 numbers: 13,
16, 19, 14, 17, 20, 15, 18 and 21) consists of
(circle one) 12 / 20 / 22 numbers.
The chance a second section comes up is then
(circle one) 12
38
/ 20
38
/ 22
38
(b) Let random variable X be the winnings from a $1 bet placed on an even
coming up. If an even number does come up, the gambler keeps his dollar
and receives another dollar (+$1). If an odd number comes up, the gambler
loses the dollar he bets (−$1). In other words, an even pays “1 to 1”. And
so
P {X = $1} = (circle one) 1838
/ 20
38
/ 22
38
P {X = −$1} = (circle one) 18 38
/ 20
38
/ 22
38
True / False The distribution of X is then
x -$1 $1
22 20
P (X = x) 38 38
60 Chapter 4. Random Variables
(c) Let random variable Y be the winnings from a $1 bet placed on the second
section coming up. If a second section number does come up, the gambler
keeps his dollar and receives another two dollars (+$2). If a first or third
section number comes up, the gambler loses the dollar he bets (−$1). In
other words, an second section bet pays “2 to 1”.
P {Y = $2} = (circle one) 12
38
/ 20
38
/ 26
38
P {Y = −$1} = (circle one) 1238
/ 20
38
/ 26
38
True / False The distribution of Y is then
y -$1 $2
P (Y = y) 26
38
12
38
20 26
By the way, P {Y = $1} = (circle one) 0 / 38
/ 38
5. Random Variables And Urns. Two marbles are taken, one at a time, without
replacement, from an urn which has 6 red and 10 blue marbles. We win $2 for
each red marble chosen and lose $1 for each blue marble chosen. Let X be the
winnings.
0 1marbles
(a) The chance both 0 1 are0red 1is0 1 0 10 1
@ A@
6 10 A @ 9 A@ 8 A @ A@
8 11 A
2
(circle one) 0 10 / 10 12 /
1
0 2
1
@ 16 A @ 16 A @ 16 A
2 3 3
(b) Since the winnings are X = $4, if both marbles are red, then
P {X = $4} = (circle one) 0.025 / 0.125 / 0.225
Use your calculator to work out the combinations.
(c) Choose the correct distribution below.
i. Distribution A.
x -$2 $1 $4
P (X = x) 0.500 0.375 0.125
ii. Distribution B.
x -$2 $1 $4
P (X = x) 0.375 0.500 0.125
x 0 1 2
1 2 1
P {X = x} 4 4 2
or
Consequently,
or
0.25, x<0
F (x) = 0.75, 0≤x<1
1, 1≤x
0.75
0.50
0.25
0 1 2 3
x 8 9 10 11 12
5 4 3 2 1
P (X = x) 36 36 36 36 36
62 Chapter 4. Random Variables
Consequently,
1 2 3
F (2)P {X ≤ 2} = (circle one) 36 / 36 / 36
1 2 3
F (3) = P {X ≤ 3} = (circle one) 36 / 36 / 36
F (11) = (circle one) 34
36
/ 35
36
/1
34 35
F (12) = (circle one) 36 / 36 / 1
True / False
0, x < 2
1
, 2≤x<3
36
3, 3≤x<4
36
F (x) = .. ..
. .
35
, 11 ≤ x < 12
1,36
12 ≤ x
0 1
P {X < 2} = (circle one) 36 / 36 /1
0
P {X > 2} = (circle one) 36 / 35
36
/1
P {2 ≤ X < 4} = (circle none, one or more) F (4) − F (2) / F (3) − F (1)
/ F (3) − F (1)
(b) Let X be the number of 4’s rolled. Recall,
25 10 1
P {X = 0} = , P {X = 1} = , P {X = 2} = .
36 36 36
Consequently,
25 35
F (0) = (circle one) 36
/ 36
/1
25 35
F (1) = (circle one) 36
/ 36
/1
25 35
F (2) = (circle one) 36
/ 36
/1
True / False
0, x<0
25
, 0≤x<1
F (x) = 36
35
, 1≤x<2
36
1, 2≤x
1 1 1
(c) P {X = 2} = F (2) − F (1) = (circle one) 6
/ 3
/ 2
(a) True / False If a < b, then F (a) ≤ F (b); that is, F is nondecreasing.
(b) True / False limb→∞ F (b) = 1
(c) True / False limb→−∞ F (b) = 0
(d) True / False limn→∞ F (bn ) = F (b); that is, F is right continuous (which
determines where the solid and empty endpoints are on the graph of a
distribution function)
p(a) = P {X = a}
X 0 2 4 6 8 10
p(x) 0.17 0.21 0.18 0.11 0.16 0.17
0 2 4 6 8 10 0 2 4 6 8 10 0 2 4 6 8 10
x 0 1 2 3
1 3 3 1
P (X = x) 8 8 8 8
Section 3. Discrete Random Variables 65
1 3 4
(a) At exactly x = 0, P (X = 0) = (circle one) 0 / 8
/ 8
/ 8
1 3 4
(b) whereas at x = −2, P (X = −2) = (circle one) 0 / 8
/ 8
/ 8
1 3 4
(c) and, indeed, at any x < 0, P (X = x) = (circle one) 0 / 8
/ 8
/ 8
1
(d) Since, also, at x = 14 , P (X = 14 ) = (circle one) 0 / 8
/ 38 / 4
8
1 3
(e) and at x = 12 , P (X = 12 ) = (circle one) 0 / 8
/ 8
/ 48
1 3 4
(f) and, indeed, at any 0 < x < 1, P (X = x) = (circle one) 0 / 8
/ 8
/ 8
1 3 4
(g) But, at exactly x = 1, P (X = 1) = (circle one) 0 / 8 / 8 / 8
(h) whereas, at x = 1 14 , P (X = 1 14 ) = (circle one) 0 / 18 / 38 / 48
4. Flipping a Coin. The number of heads, X, in one flip of a coin, is given by the
following probability distribution.
i. Distribution A.
X 0 1
p(x) 0.25 0.75
ii. Distribution B.
X 0 1
p(x) 0.75 0.25
iii. Distribution C.
X 0 1
p(x) 0.50 0.50
(e) The number of different ways of describing a distribution include (check
none, one or more)
i. function
ii. tree diagram
iii. table
iv. graph
(f) True / False F (a) = allx≤a p(x)
5. Rock, Scissors and Paper. Rock, scissors and paper (RSP) involves two players
in which both can either show either a “rock” (clenched fist), “scissors” (V–
sign) or “paper” (open hand) simultaneously, where either does not know what
the other is going to show in advance. Rock beats scissors (crushes it), scissors
beats paper (cuts it) and paper bets rock (covers it). Whoever wins, receives
a dollar ($1). The payoff matrix RSP is given below. Each element represents
the
amount player C (column) pays player R (row).
(a) According to the payoff matrix, if both players C and R show “rock”, then
player C pays player R
(circle one) -$1 / $0 / $$1
(In other words, no one wins–one player does not pay the other player.)
(b) If player C shows “rock” and player R shows “paper”, then player C pays
player R
(circle one) -$1 / $0 / $$1
(c) To say player C pays player R negative one dollar, -$1, means (circle one)
Section 3. Discrete Random Variables 67
X 0 2 4 6 8 10
p(x) 0.17 0.21 0.18 0.11 0.16 0.17
(a) A First Look: Expected Value Is Like The Fulcrum Point of Balance.
0 2 4 6 8 10 0 2 4 6 8 10 0 2 4 6 8 10
(c) General Formula For The Expected Value. True / False The general
formula for the expected value (mean) is given by
n
E[X] = xi p(xi )
i=1
x 0 1 2 3
1 3 3 1
p(x) 8 8 8 8
x 0 1 2 3
4 2 1 1
p(x) 8 8 8 8
the mean is
4 2 1 1
E[X] = ×0 + ×1 + ×2 + ×3 =
8 8 8 8
which is equal to (circle one) 1.500 / 0.875 / 1.375 / 0.625
4. Rolling a Pair of Dice. If the dice are fair, the distribution of X (the sum of
two rolls of a pair of dice) is then
x 2 3 4 5 6 7
1 2 3 4 5 6
P (X = x) 36 36 36 36 36 36
x 8 9 10 11 12
5 4 3 2 1
P (X = x) 36 36 36 36 36
The mean (expected) sum of the roll of a pair of fair dice is then
1 2 2 1
E[X] = µX = ×2 + ×3 + · · · + ×11 + ×12
36 36 36 36
which is equal to (circle one) 5 / 6 / 7 / 8.
(Think about it: this is a symmetric distribution balanced on what number?)
70 Chapter 4. Random Variables
6. Roulette. The roulette table has 38 numbers: the numbers are 1 to 36, 0 and
00. A ball is spun on a corresponding roulette wheel which, after a time, settles
down and the ball drops into one of 38 slots which correspond to the 38 numbers
on the roulette table.
(a) Let random variable X be the winnings from a $1 bet placed on an even
coming up, where this bet pays 1 to 1. Recall,
x -$1 $1
20 18
p(x) 38 38
X 0 1 2 3 4
p(x) 0.32 0.42 0.21 0.05 0.004
X 0 2 4 6 8 10
p(x) 0.17 0.21 0.18 0.11 0.16 0.17
(a) If the medical costs for each seizure, X, is $200, g(x) = 200x, the new
distribution for g(x) becomes,
72 Chapter 4. Random Variables
X 0 2 4 6 8 10
g(X) = 200x 200(0) = 0 200(2) = 400 800 1200 1600 2000
p(g(x)) 0.17 0.21 0.18 0.11 0.16 0.17
The expected value (mean) cost of seizures is then given by
X 0 2 4 6 8 10
g(X) = 200x + 1500 200(0) + 1500 = 1500 1900 2300 2700 3100 3500
p(g(x)) 0.17 0.21 0.18 0.11 0.16 0.17
The expected value (mean) cost of seizures is then given by
2. Flipping Until a Head Comes Up. A (weighted) coin has a probability of p = 0.7
of coming up heads (and so a probability of 1 − p = 0.3 of coming up tails).
This coin is flipped until a head comes up or until a total of 4 flips are made.
Let X be the number of flips. Then, recall,
X 1 2 3 4
2 3
p(x) 0.7 0.3(0.7) = 0.21 0.3 (0.7) = 0.063 0.3 = 0.027
4.6 Variance
We will now look at the variance, V (X),
X 0 2 4 6 8 10
P(X = x) 0.17 0.21 0.18 0.11 0.16 0.17
and the expected value (mean) number of seizures is given by µ = E(X) = 4.78,
(a) A First Look: Variance Measures How “Dispersed” The Distribution Is.
0 2 4 6 8 10 0 2 4 6 8 10 0 2 4 6 8 10
(a) seizure distribution (b) another distribution (c) and another distribution
x 0 1 2 3
1 3 3 1
P (X = x) 8 8 8 8
76 Chapter 4. Random Variables
(a) One Way To Calculate Variance. Since the mean (expected) number of
smokers is µ = 1.5, then the variance is given by,
1 3 3 1
Var(X) = (0 − 1.5)2 + (1 − 1.5)2 + (2 − 1.5)2 + (3 − 1.5)2
8 8 8 8
which is equal to (circle one) 0.02 / 0.41 / 0.59 / 0.75.
The standard
√ deviation is given by
SD(X) = 0.75 (circle one) 0.47 / 0.86 / 1.07 / 2.25.
(b) Another Way To Calculate Variance. Since the mean (expected) number
of smokers is E[X] = µ = 1.5 and the second moment is given by
1 3 3 1
E[2 ] = (0)2 + (1)2 + (2)2 + (3)2 = 3
8 8 8 8
then the variance is given by,
3. Rolling a Pair of Dice. If the dice are fair, the distribution of X (the sum of
two rolls of a pair of dice) is
x 2 3 4 5 6 7
1 2 3 4 5 6
P (X = x) 36 36 36 36 36 36
x 8 9 10 11 12
5 4 3 2 1
P (X = x) 36 36 36 36 36
and µ = 43 , then
2 2
4 3−1 4 3−2
Var(X) = 1 − + 2− ,
3 3 3 3
5. Roulette.
(a) Let random variable X be the winnings from a $1 bet placed on an even
coming up, where this bet pays 1 to 1. Recall,
x -$1 $1
20 18
p(x) 38 38
2
where the mean is µ = − 38 and so
2 2
2 20 2 18
Var(X) = −1 − − + 1− − ,
38 38 38 38
5.1 Introduction
For all real x ∈ (−∞, ∞),
P {X ∈ B} = f (x) dx
B
where f (x) is called the probability density function and so
b
P {x ≤ X ≤ B} = f (x) dx
a
a
P {X = a} = f (x) dx = 0
a
a
P {X < a} P {X ≤ a} = f (x) dx
−∞
area is 1
x x 49 x
0 49 51 0 49 50 51 0 51
49.3 50.7
(a) f(x) = 0.5 on [49, 51] (b) f(x) = 0.5 on [49, 51] (c) f(x) = 0.5 on [49, 51]
107
108 Chapter 5. Continuous Random Variables
(a) If all of the filled bags must fall between 49 and 51 pounds, then there
is (circle one) a little / no chance that one filled bag, chosen at random
from all filled bags, will weigh 48.5 pounds.
(b) There is (circle one) a little / no chance that one filled bag, chosen at
random, will weigh 51.5 pounds.
(c) There is a (circle one) 100% / 50% / 0% chance that one randomly chosen
filled bag chosen will weigh 53.5 pounds.
(d) One randomly chosen filled bag will weigh 36 pounds with probability
(circle one) 1 / 0.5 / 0.
(e) One randomly chosen filled bag will weigh (strictly) less than 49 pounds
with probability P (x < 49) = (circle one) 1 / 0.5 / 0.
(f) One randomly chosen filled bag will weigh (strictly) more than 51 pounds
with probability P (x > 51) = (circle one) 1 / 0.5 / 0.
(g) Figure (a). One randomly chosen filled bag will weigh between 49 and 51
pounds (inclusive) with probability P (49 ≤ x ≤ 51) =
(circle one) 1 / 0.5 / 0.
(h) More Figure (a). The probability P (49 ≤ x ≤ 51) is represented by or
equal to the (circle none, one or more)
i. rectangular area equal to 1.
ii. rectangular area equal to the width (51 − 49 = 2) times the height
(0.5).
iii. definite integral of f (x) = 0.5 over the interval [49, 51].
51
iv. 49 0.5 dx = [0.5x]5149 = 0.5(51) − 0.5(49) = 1.
(i) True / False The probability density function is given by the piecewise
function,
0 if x < 49
f (x) = 0.5 if 49 ≤ x ≤ 51
0 if x > 51
This is an example of a uniform probability density function (pdf).
(j) Figure (b). One randomly chosen filled bag will weigh between 49 and 50
(not 51!) pounds (inclusive) with probability
P (49 ≤ x ≤ 50) = (50 − 49)(0.5) = (circle one) 0 / 0.5 / 1.
(k) More Figure (b). One randomly chosen filled bag will weigh between 49
and 50 pounds (inclusive)
50 with probability
P (49 ≤ x ≤ 50) = 49 0.5 dx = [0.5x]50
49 = 0.5(50) − 0.5(49) =
(circle one) 0 / 0.5 / 1.
Section 1. Introduction 109
(l) Figure (c). One randomly chosen filled bag will weigh between 49.3 and
50.7 pounds (inclusive) with probability
P (49.3 ≤ x ≤ 50.7) = (50.7 − 49.3)(0.5) = (circle one) 0 / 0.5 / 0.7.
(m) More Figure (c). One randomly chosen filled bag will weigh between 49.3
and 50.7 pounds (inclusive)
50.7 with probability
P (49.3 ≤ x ≤ 50.7) = 49.3 0.5 dx = [0.5x]50.7
49.3 = 0.5(50.7) − 0.5(49.3) =
(circle one) 0 / 0.5 / 0.7.
50.9
(n) P (49.1 ≤ x ≤ 50.9) = 49.1 0.5 dx = [0.5x]50.9
49.1 = 0.5(50.9) − 0.5(49.1) =
(circle one) 0 / 0.5 / 0.9.
(o) Another example.
50.9
P (x ≤ 50.9) = f (x) dx
−∞
49 50.9
= f (x) dx + f (x) dx
−∞ 49
49 50.9
= 0 dx + 0.5 dx
−∞ 49
= 0+ [0.5x]50.9
49
= 0.5(50.9) − 0.5(49) =
2 area = 1 area = 1
area = 1 0
x
x x
0.751 1.69 2.09 2.515 2.930
(a) f(x) = -(9/20)x + 1.5 (b) f(x) = x2 - 2/x (c) f(x) = x2- 5
(a) Find C such that f (x) = Cx is a probability density function over the
interval [2, 4]. In other words, find C such that
4
P (2 ≤ x ≤ 4) = Cx dx
2
4
C 2
= x
2 2
C 2 C 2
= (4) − (2)
2 2
C 2
= (4) − (2)2
2
C
= (12)
2
= 6C
= 1
1 1
and so C = (circle one) / / 16 .
4 5 4 1
2 6
x dx = 1.
(b) Find k such that f (x) = kx is a probability density function over the
Section 1. Introduction 113
and the standard deviation is defined as the square root of the variance. Some prop-
erties include,
∞
E[g(X)] = g(x)f (x) dx
−∞
E[aX + b] = aE[X] + b
Var(X) = a2 Var(X)
1/3
1/3.9
x
x 49.5 50.5 x
0 49 52 0 0 48.4 52.3
50.5 50 50.35
expected value expected value expected value
(a) f(x) = 1/3 on [49, 52] (b) f(x) = 1 on [49.5, 50.5] (c) f(x) = 1/3.9 on [48.4, 52.3]
(a) Figure (a). Since the weight of potatoes are uniformly spread over the
interval [49, 52], we would expect the weight of a potato chosen at random
from all these potatoes to be
49+52
2
= (circle one) 50 / 50.5 / 51.
(b) Figure (a) Again. The expected weight of a potato chosen at random can
also be calculated in the following way:
∞
E[X] = xf (x) dx
−∞
49 52 ∞
= xf (x) dx + xf (x) dx + xf (x) dx
−∞ 49 52
49 52 ∞
1
= x(0) dx + x dx + x(0) dx
−∞ 49 3 52
52
1 2
= 0+ x +0
6 49
1 1
= (52)2 − (49)2 =
6 6
50 / 50.5 / 51.
116 Chapter 5. Continuous Random Variables
(c) Figure (b). Since the weight of potatoes are uniformly spread over the
interval [49.5, 50.5], we would expect the weight of a potato chosen at ran-
dom from all these potatoes to be
49.5+50.5
2
= (circle one) 50 / 50.5 / 51.
(d) Figure (b) Again. The expected weight of a potato chosen at random can
also be calculated in the following way:
∞
E[X] = xf (x) dx
−∞
49.5 50.5 ∞
= xf (x) dx + xf (x) dx + xf (x) dx
−∞ 49.5 50.5
49.5 50.5 ∞
= x(0) dx + x(1) dx + x(0) dx
−∞ 49.5 50.5
50.5
1 2
= 0+ x +0
2 49.5
1 1
= (50.5)2 − (49.5)2 =
2 2
50 / 50.5 / 51.
(e) Figure (c). The expected weight of a potato chosen at random can also be
calculated in the following way:
∞
E(x) = xf (x) dx
−∞
52.3
1
= x dx
48.4 3.9
52.3
1 2
= x
7.8 48.4
1 1
= (52.3)2 − (48.4)2 =
2(7.8) 2(7.8)
2. Expected Values For Other Probability Density Functions Consider the following
probability density functions, as shown in the three graphs below.
Section 2. Expectation and Variance of Continuous Random Variables 117
0
x
0.751 x 1.69 2.09 x
2.515 2.930
0.359 1.93 2.77
expected value expected value expected value
(a) f(x) = -(9/20)x + 1.5 (b) f(x) = x2 - 2/x (c) f(x) = x2- 5
(a) Figure (a). Since there is “more” probability on the “left” of the interval
[0, 0.751], we would expect the expected (or mean) weight of value chosen
from this distribution to be (circle one) smaller than / equal to / larger
than the middle value, 0+0.751
2
= 0.3755.
(b) Figure (a) Again. The expected value is
∞
E[X] = xf (x) dx
−∞
0 0.751 ∞
= xf (x) dx + xf (x) dx + xf (x) dx
−∞ 0 0.751
0 0.751 ∞
9
= x(0) dx + x − x + 1.5 dx + x(0) dx
−∞ 0 20 0.751
0.751
9 2
= − x + 1.5x dx
0 20
0.751
−9 3 1.5 2
= x + x =
60 2 0
2
(b) The variance of the probability density f (x) = x2 − x
on the interval
[1.69, 2.09], is
Var(X) = E(x2 ) − [E(x)]2
∞ ∞
2
2
= x f (x) dx − xf (x) dx
−∞ −∞
2.09 2.09
2
2 2 2 2 2
= x x − dx − x x − dx
1.69 x 1.69 x
2.09 2.09
2
4 3
= x − 2x dx − x − 2 dx =
1.69 1.69
1
area is 1 area is 1
1/3 area is 1 1/3.9
x
x x
0 49 52 0 49.5 50.5 0 48.4 52.3
(a) f(x) = 1/3 on [49, 52] (b) f(x) = 1 on [49.5, 50.5] (c) f(x) = 1/3.9 on [48.4, 52.3]
P (x ≤ 50.2) = F (50.2)
50.2 − α
=
β−α
50.2 − 49
= =
52 − 49
(circle one) 1
3
/ 13:2 / 13:4 .
v. P (x ≥ 50.2) = (circle one) 03:8 / 13:2 / 13:8 .
vi. Expectation and Variance.
E[X] = β+α2
= 52+49
2
= (circle one) 50 / 50.5 / 51.
(β−α)2 (52−49)2
Var(X) = 12 = 12 (circle one) 0.75 / 1 / 1.25.
(b) Figure (b).
i. The probability P (49.5 ≤ x ≤ 50.5) is represented by or equal to the
(circle none, one or more)
A. rectangular area equal to 1.
B. rectangular area equal to the width (50.5 − 49.5 = 3) times the
height (1).
C. definite integral of f (x) = 1 over the interval [49.5, 50.5].
50.5
D. 49.5 1 dx = [x]50.5
49.5 = 51.5 − 49.5 = 1.
ii. True / False The probability density function is given by the piece-
wise function,
1 if 49.5 ≤ x ≤ 50.5
f (x) =
0 elsewhere
Section 3. The Uniform Random Variable 123
P (x ≤ 50.2) = F (50.2)
50.2 − α
=
β−α
50.2 − 49.5
= =
52 − 49.5
(circle one) 0.5 / 0.7 / 0.9.
v. P (x ≥ 50.1) = 1 − F (50.2) = (circle one) 0.2 / 0.3 / 0.4.
vi. Expectation and Variance.
E[X] = β+α2
= 50.5+49.5
2
= (circle one) 50 / 50.5 / 51.
(β−α)2 (50.5−49.5)2
Var(X) = 12 = 12
(circle one) 0.075 / 0.083 / 0.093.
(c) Figure (c).
i. The probability P ([48.4, 52.3]) = P ([48.4 ≤ x ≤ 52.3]) is represented
by or equal to the (circle none, one or more)
A. rectangular area equal to 1.
B. rectangular area equal to the width (52.3 − 48.4 = 3.9) times the
1
height ( 3.9 ).
1
C. definite integral of f (x) = 3.9 over the interval [48.4, 52.3].
52.3 1 1 52.3 1 1
D. 48.4 3.9 dx = 3.9 x 48.4 = 3.9 (52.3) − 3.9 (48.4) = 1.
ii. True / False The probability density function is given by the piece-
wise function,
1
3.9
if 48.4 ≤ x ≤ 52.3
f (x) =
0 elsewhere
124 Chapter 5. Continuous Random Variables
P (x ≤ 50.2) = F (50.2)
50.2 − α
=
β−α
50.2 − 48.4
= =
52.3 − 48.4
E[X] = µ
Var(X) = σ 2
f(x) f(x)
P(X < 1.42) = ?
0 x x
1.42 -2.11 0
(a) (b)
f(x)
f(x)
P(X > 0.54) = ? P(-1.73<X<1.62) = ?
(c) (d)
(a) The standard normal distribution, in (a) of the figure above, say, is (circle
one) skewed right / symmetric / skewed left.
(b) Since the standard normal is a probability density function, the total area
under this curve is (circle one) 50% / 75% / 100% / 150%.
(c) The shape of this distribution is (circle one)
triangular / bell–shaped / rectangular.
(d) This distribution has an expected value at (circle one) µ = 0o / µ = 1o .
(e) Since this distribution is symmetric, (circle one) 25% / 50% / 75% of the
temperatures are above (to the right) of 0o .
(f) The probability of the temperature being less than 1.42o is (circle one)
greater than / about the same as / smaller than 0.50. Use (a) in
the figure above.
(g) The probability the temperature is less than 1.42o,
P {X ≤ 1.42} = F (1.42)
= Φ(1.42)
1.42
1 2
= √ e−y /2 dy =
2π −∞
0.9222 / 0.0174 / 0.2946 / 0.9056.
(It is not possible to determine this integral in an analytical way (“by
hand”) and so you must use your calculator to perform a numerical ap-
proximation for this integration: 2nd DISTR 2:normalcdf(− 2nd EE 99,
1.42); look at graph (a) of the figure above to better visualize the proba-
bility that is being determined.)
Section 4. Normal Random Variables 127
2. Nonstandard Normal, A First Look: IQ Scores. It has been found that IQ scores
can be distributed by a nonstandard normal distribution. The following figure
compares the two normal distributions for the 16 year olds and 20 year olds.
f(x)
σ = 16
f(x)
20 year old IQs
σ = 20
µ = 100 µ = 120 x
(f) The normal distribution for the 20 year old IQ scores is (circle one) shorter
than / as tall as / taller than than the normal distribution for the 16
year old IQ scores.
(g) The total area (probability) under the normal distribution for the 20 year
old IQ scores is (circle one) smaller than / the same as / larger than
the area under the normal distribution for the 16 year old IQ scores.
(h) True / False Neither the normal distribution for the IQ scores for the
20 year old IQ scores nor the 16 year old IQ scores is a standard normal
because neither have mean zero, µ = 0, and standard deviation 1, σ = 1.
Both, however, have the same general “bell–shaped” distribution.
(i) There is (circle one) one / two / many / an infinity of nonstandard
normal distributions. The standard normal is one special case of the family
of (nonstandard) normal distributions where µ = 0 and σ = 1.
SD 16 SD 16
84 x 100 120 x
100 96
(a) (b)
SD 20 SD 20
x x
84 120 96 120
(c) (d)
(a) The upper two (of the four) normal curves above represent the IQ scores
for sixteen year olds. Both are nonstandard normal curves because the
(circle none, one or more)
i. the average is 100 and the SD is 16.
ii. neither the average is 0, nor is the SD equal to 1.
iii. the average is 16 and the SD is 100.
iv. the average is 0 and the SD is 1.
Section 4. Normal Random Variables 129
The lower two normal curves above represent the IQ scores for twenty year
olds (µ = 120, σ = 20).
(b) Since the sixteen year old distribution is symmetric, (circle one) 25% /
50% / 75% of the IQ scores are above (to the right) of 100.
(c) The probability of the IQ scores being less than 84, P {X < 84}, for the
sixteen year old distribution is (circle one) greater than / about the
same as / smaller than 0.50.
(d) P {X < 84} =
84 − µ
P {X > 84} = 1 − Φ
σ
84 − 100
= 1−Φ
16
84
1 2
= 1− √ e−(1/2)[(84−100)/16] dy =
σ 2π −∞
(a) The IQ scores for the 16 year olds are normal with µ = 100 and σ = 16.
The standardized value of the nonstandard IQ score of 110 for the 16 year
olds, then, is
Z = X−µσ
= 110−100
16
= (circle one) 0.625 / 1.255 / 3.455
and so P {X > 110} = P {Z > 0.625}.
(Compare 2nd DISTR 2:normalcdf(110, 2nd EE 99, 100, 16) with 2nd
DISTR 2:normalcdf(0.625, 2nd EE 99, 0, 1).)
(b) The IQ scores for the 20 year olds are normal with µ = 120 and σ = 20.
The standardized value of the nonstandard IQ score of 110 for the 20 year
olds, then, is
Z = X−µσ
= 110−120
20
= (circle one) 0.5 / -0.5 / 0.25.
and so P {X > 110} = P {Z > −0.5}.
(Compare 2nd DISTR 2:normalcdf(110, 2nd EE 99, 120, 20) with 2nd
DISTR 2:normalcdf(−0.5, 2nd EE 99, 0, 1).)
(c) If both a 16 year old and 20 year old score 110 on an IQ test, (check none,
one or more)
i. the 16 year old is brighter relative to his age group than the 20 year
old is relative to his age group
ii. the z–score is higher for the 16 year old than it is for the 20 year old
iii. the z–score allows us to compare the IQ score for a 16 year old with
the IQ score for a 20 year old
(d) If µ = 100 and σ =
16, then
P {X > 130} = P Z > 130−100
16
= (circle one) 0.03 / 0.31
(e) If µ = 120 and σ =
20, then
P {X > 130} = P Z > 130−120
20
= (circle one) 0.03 / 0.31
(f) If µ = 25 and σ = 5, then
P {27 < X < 32} = P 27−25 5
<Z< 32−25
5
=
(circle one) 0.03 / 0.26 / 0.31
ii. standard deviation is given by σ = np(1 − p) (circle one) 4 / 2.4 /
3.0 / 1.55.
iii. P {X ≥ 5} = (circle one) 0.367 / 0.289 / 0.577.
(Use your calculator; subtract 2nd DISTR A:binomcdf(10, 0.4, 4) from
one (1).)
(b) Normal Approximation. Consider a graph of the binomial and a normal
approximation to this distribution below.
f(x)
0.30
0.30
σ = 1.55 2
0.25
P(X >
_ 5) 0.25 Ν(4,1.55 )
0.20 0.20
P(X = x)
P(X >
_ 4.5)
0.15 0.15
0.10 0.10
0.05 0.05
0.00 0.00
0 1 2 3 4 5 6 7 8 9 10 µ=4
number of cases won, x number of cases won, x
F (a) = P {X ≤ a} = 1 − e−λa , a ≥ 0
(1)
(2) (2) (1)
(3) (3)
0 3 x 0 3 x
1.1
P {X ≤ 1.1} = F (1.1)
= 1 − e−λ(1.1)
1
= 1 − e− 2 (1.1) =
and also
P {X > 10, X > 0}
P {X > 10|X > 0} =
P {X > 0}
P {X > 10}
=
P {X > 0}
1 − F (10)
=
1 − F (0)
1 − (1 − e−3(10) )
=
1 − (1 − e−3(0) )
1 − (1 − e−3(10) )
= =
1
(circle one) e;10 / e;20 / e;30
or,
or, in other words the chance a battery lasts at least 10 hours or more,
is the same as the chance a battery lasts at least 10 hours more, given
that it has already lasted 0 hours or more (which is not too surprising).
ii. t = 5, s = 10. Suppose the batteries are 5 hours old when t = 5 and
they are 10 hours old when s = 10. Then, once again,
or, in other words, the chance a battery lasts at least 10 hours or more,
is the same as the chance a battery lasts at least 15 hours more, given
that it has already lasted 5 hours or more. This is kind of surprising,
because it seems to imply the battery’s life starts “fresh” after 5 hours,
as though the battery “forgot” about the first five hours of its life.
iii. What Is Not Being Said. True / False Although
(2)
(4)
(1)
(3)
0 x
5 10 15
Figure 5.11 (Memoryless Property of Exponential)
v. An Implication Of The Memoryless Property of Exponential Distribu-
tions? True / False If
p(x, y) = P {X = x, Y = y}
p(x, y, . . . , xn ) = P {X = x, Y = y, . . . , Xn = xn }
FX (a) = P {X ≤ a} = F (a, ∞)
FY (b) = P {Y ≤ b} = F (∞, b)
165
166 Chapter 6. Jointly Distributed Random Variables
and density1
∂2
f (a, b) = F (a, b)
∂a∂b
and which has a joint density function in n variables generalization,
F (a1 , a2 , . . . , an ) = P {X1 ≤ a1 , X2 ≤ a2 , . . . , Xn ≤ an }
1. Discrete Joint Density: Waiting Times To Catch Fish. The joint density,
P {X, Y }, of the number of minutes waiting to catch the first fish, X, and
the number of minutes waiting to catch the second fish, Y , is given below.
P {X = i, Y = j} j row sum
1 2 3 P {X = i}
1 0.01 0.02 0.08 0.11
i 2 0.01 0.02 0.08 0.11
3 0.07 0.08 0.63 0.78
column sum P {Y = j} 0.09 0.12 0.79
(a) The (joint) chance of waiting three minutes to catch the first fish and three
minutes to catch the second fish is
P {X = 3, Y = 3} = (circle one) 0.09 / 0.11 / 0.63 / 0.78.
(b) The (joint) chance of waiting three minutes to catch the first fish and one
minute to catch the second fish is
P {X = 3, Y = 1} = (circle one) 0.07 / 0.11 / 0.63 / 0.78.
(c) The (joint) chance of waiting one minute to catch the first and three min-
utes to catch the second fish is
P {X = 1, Y = 3} = (circle one) 0.08 / 0.11 / 0.63 / 0.78.
(d) The (marginal) chance of waiting three minutes to catch the first fish is
P {X = 3} = (circle one) 0.09 / 0.11 / 0.12 / 0.78.
(e) The (marginal) chance of waiting three minutes to catch the second fish is
P {Y = 3} = (circle one) 0.09 / 0.11 / 0.12 / 0.79.
(f) The (marginal) chance of waiting three minutes to catch the second fish is
(circle none, one or more)
i. P {Y = 3} = 0.79
1
Notice that the differentiation is with respect to a and b, rather than X and Y !
Section 1. Joint Distribution Functions 167
ii. P {X = 1, Y = 3} + P {X = 2, Y = 3} + P {X = 3, Y = 3} =
0.08 + 0.08 + 0.63 = 0.79
iii. pY (3) = p(1, 3) + p(2, 3) + p(3, 3) = 0.08 + 0.08 + 0.63 = 0.79
iv. pY (3) = y:p(x,y)>0 p(x, y) = p(1, 3) + p(2, 3) + p(3, 3) = 0.08 + 0.08 +
0.63 = 0.79
(g) The (marginal) chance of waiting two minutes to catch the first fish is
(circle none, one or more)
i. P {X = 2} = 0.11
ii. P {X = 2, Y = 1} + P {X = 2, Y = 2} + P {X = 2, Y = 3} =
0.01 + 0.02 + 0.08 = 0.11
iii. pX (2) = p(2, 1) + p(2, 2) + p(2, 3) = 0.01 + 0.02 + 0.08 = 0.11
iv. pX (2) = y:p(2,y)>0 p(2, y) = p(2, 1) + p(2, 2) + p(2, 3) = 0.01 + 0.02 +
0.08 = 0.11
(h) The (marginal) chance of waiting two minutes to catch the second fish is
(circle none, one or more)
i. P {Y = 2} = 0.12
ii. P {X = 2, Y = 1} + P {X = 2, Y = 2} + P {X = 2, Y = 3} =
0.01 + 0.02 + 0.08 = 0.11
iii. pY (2) = p(1, 2) + p(2, 2) + p(3, 2) = 0.02 + 0.02 + 0.08 = 0.12
iv. pY (2) = y:p(x,2)>0 p(x, 2) = p(1, 2) + p(2, 2) + p(3, 2) = 0.02 + 0.02 +
0.08 = 0.12
(i) The chance of waiting at least two minutes to catch the first fish is (circle
none, one or more)
i. P {X ≥ 2} = 0.11 + 0.78 = 0.89
ii. P {X = 2, Y = 1} + P {X = 2, Y = 2} + P {X = 2, Y = 3} + P {X =
3, Y = 1} + P {X = 3, Y = 2} + P {X = 3, Y = 3} = 0.01 + 0.02 +
0.08 + 0.07 + 0.08 + 0.63 = 0.89
iii. pX (2) = p(2, 1) + p(2, 2) + p(2, 3) = 0.01 + 0.02 + 0.08 = 0.11
iv. pX (2) = y:p(2,y)>0 p(2, y) = p(2, 1) + p(2, 2) + p(2, 3) = 0.01 + 0.02 +
0.08 = 0.11
(j) The chance of waiting at most two minutes to catch the first fish is (circle
none, one or more)
i. P {X ≤ 2} = 0.11 + 0.11 = 0.22
ii. P {X = 1, Y = 1} + P {X = 1, Y = 2} + P {X = 1, Y = 3} + P {X =
2, Y = 1} + P {X = 2, Y = 2} + P {X = 2, Y = 3} = 0.01 + 0.02 +
0.08 + 0.01 + 0.02 + 0.08 = 0.22
iii. F (2, 3) = P {X ≤ 2, Y ≤ 3} = 0.22
168 Chapter 6. Jointly Distributed Random Variables
2. Discrete Joint Density: Coin and Dice. A fair coin, marked “1” on one side
and “2” on the other, is flipped once and, independent of this, one fair die is
rolled once. Let X be the value of the coin (either 1 or 2) flipped and let Y be
the sum of the coin flip and die roll (for example, a flip of 2 and a roll of 1 gives
Y = 3).
(a) The chance of flipping a “1” and the sum of coin and die is equal to 2 is
P {X = 1, Y = 2} = P {coin is 1, sum is 2}
= P {sum is 2|coin is 1}P {coin is 1}
= P {die is 1}P {coin is 1}
1 1
= ·
6 2
1 1 1 1
which equals (circle one) 10
/ 11
/ 12
/ 13
.
Section 1. Joint Distribution Functions 169
(b) The chance of flipping a “1” and the sum of coin and die is equal to 3 is
P {X = 1, Y = 3} = P {coin is 1, sum is 3}
= P {sum is 3|coin is 1}P {coin is 1}
= P {die is 2}P {coin is 1}
1 1
= ·
6 2
1 1 1 1
which equals (circle one) 10
/ 11
/ 12
/ 13
.
(c) The chance of flipping a “2” and the sum of coin and die is equal to 2 is
P {X = 2, Y = 2} = P {coin is 2, sum is 2}
= P {sum is 2|coin is 2}P {coin is 2}
= P {die is 0}P {coin is 2}
1
= 0·
2
1 1 1
which equals (circle one) 0 / 11
/ 12
/ 13
.
(d) True / False. The following table is the joint density of P {X, Y }.
P {X = i, Y = j} j row sum
2 3 4 5 6 7 8 P {X = i}
1 1/12 1/12 1/12 1/12 1/12 1/12 0 6/12
i 2 0 1/12 1/12 1/12 1/12 1/12 1/12 6/12
column sum P {Y = j} 1/12 2/12 2/12 2/12 2/12 2/12 1/12
P{X = i, Y = j} P{X = i, Y = j}
y y
8 8
7 7
6 6
1/12 5 1/12 5
4 4
3 3
2 2
1 2 x 1 2 x
3. Discrete Joint Density: Marbles In An Urn. Three marbles are chosen at ran-
dom without replacement from an urn consisting of 6 black and 8 blue marbles.
Let Xi equal 1 if the ith marble selected is black and let it equal 0 otherwise.
P {X = i, Y = j} j row sum
20 40 P {X = i}
i 1 0.2 0.3 0.5
2 0.4 0.1 0.5
P {Y = j} 0.6 0.4
column sum
Determine the density of the total amount of money spent on speeding tickets
in a year, Z = XY , and use this density to calculate P {XY > 20}.
P {X = i, Y = j} j P {X = i}
Z = XY 20 40
i 1 0.2 0.3 0.5
1(20) = 20 1(40) = 40 4
2 0.4 0.1 0.5
2(20) = 40 2(40) = 80 5
P {Y = j} 0.6 0.4
The chance that the total amount paid for speeding tickets in a year is $20
is given by
(circle one) 0.1 / 0.2 / 0.4 / 0.5.
(b) The chance that the total amount paid for speeding tickets in a year is
$40, z = xy = 40, occurs in two possible ways, (2,1) and (1,2), with
probabilities (circle one)
i. 0.1 and 0.3, respectively.
ii. 0.2 and 0.3, respectively.
iii. 0.3 and 0.3, respectively.
iv. 0.4 and 0.3, respectively.
(c) Thus, the chance that the total amount paid for speeding tickets in a year
is $40, z = xy = 40, is
P{XY = 40} = 0.4 + 0.3 = (circle one) 0.4 / 0.6 / 0.7.
(d) The product, z = xy = 40, also occurs in two possible ways (circle one)
i. (2,2) and (1,2).
ii. (1,2) and (2,2).
iii. (1,1) and (2,2).
iv. (2,1) and (1,2).
(e) Complete the probability distribution of the total amount paid for speeding
tickets in a year, Z = XY ,
z = XY 20 40 80
P {XY } 0.2 0.1
5. Continuous Joint Density: Weight and Amount of Salt in Potato Chips. Three
machines fills potato chip bags. Although each bag should weigh 50 grams
each and contain 5 milligrams of salt, in fact, because of differing machines,
the weight and amount of salt placed in each bag varies according to the three
graphs below.
Section 1. Joint Distribution Functions 173
volume = 1 3x + y = 155
volume = 1
f(x,y) f(x,y)
f(x,y) 0.25 y y
y 8 8
8 7 1/6 7
6 6
7 5 5
6 4 4
1/12 5 3 3
4 2 2
3
2 volume = 1
49 51 x 49 51 x
49 51 x
(a) Machine A; Figure (a). One randomly chosen filled bag will weigh between
49 and 51 grams and contain between 2 and 8 milligrams of salt with
probability P {49 ≤ X ≤ 51, 2 ≤ Y ≤ 8} = (circle one) 1 / 0.5 / 0.
(b) More Machine A. The probability P {49 ≤ X ≤ 51, 2 ≤ Y ≤ 8} is repre-
sented by or equal to the (circle none, one or more)
i. rectangular box volume equal to 1.
ii. rectangular box volume equal to the width (51 − 49 = 2) times the
1
depth (8 − 2 = 6) times the height ( 12 ).
1
iii. definite integral of f (x) = 12 over the region (49, 51) × (2, 8).
iv. the integral,
8 51 8 51
1 x
dx dy = dy
2 49 12 2 12 49
8
51 − 49
= dy
2 12
8
2y
=
12 2
= 1
(c) More Machine A. True / False The joint probability density function is
given by,
1
12
49 ≤ x ≤ 51, 2 ≤ y ≤ 8
f (x, y) =
0 elsewhere
(d) More Machine A. The chance a potato chip bag, chosen at random, weighs
at most 50.5 grams and contains at most 4 grams of salt is (circle none,
one or more)
1 3
i. P {X ≤ 50.5, Y ≤ 4} = (1.5)(2) 12 = 12
= 0.25
174 Chapter 6. Jointly Distributed Random Variables
ii. F (50.5, 4) =
4 50.5 4 50.5
1 x
dx dy = dy
2 49 12 2 12 49
4
50.5 − 49
= dy
2 12
4
1.5y
=
12 2
3
=
12
iii. P {X = 1, Y = 1} + P {X = 2, Y = 1} = 0.01 + 0.02 = 0.03
iv. P {X ≤ 2, Y ≤ 1} = F (2, 1) = 0.11
(e) More Machine A. The chance a potato chip bag, chosen at random, weighs
at most 50.5 grams is (circle none, one or more)
1 9
i. P {X ≤ 50.5} = (1.5)(8 − 2) 12 = 12 = 0.75
ii. FX (50.5) = F (50.5, ∞)
∞ 50.5 8 50.5
1 x
dx dy = dy
2 49 12 2 12 49
8
50.5 − 49
= dy
2 12
8
1.5x
=
12 2
9
=
12
3
iii. FY (3) = F (∞, 3) = F (2, 3) = 12
iv. F (2, 5) = P {X ≤ 2, Y ≤ 5}
(f) More Machine A. The chance a potato chip bag, chosen at random, con-
tains at most 4 grams is (circle none, one or more)
1 4
i. P {Y ≤ 4} = (51 − 49)(2) 12 = 12 = 0.33
ii. FY (4) = F (∞, 4)
4 ∞ 4 51
1 x
dx dy = dy
2 49 12 2 12 49
4
51 − 49
= dy
2 12
4
2x
=
12 2
4
=
12
Section 1. Joint Distribution Functions 175
3
iii. FY (3) = F (∞, 3) = F (2, 3) = 12
iv. F (2, 5) = P {X ≤ 2, Y ≤ 5}
(g) More Machine A. The chance a potato chip bag, chosen at random, weighs
at least 50.5 grams and contains at least 4 grams of salt is (circle none, one
or more)
i. P {X ≥ 50.5, Y ≥ 4} =
8 51 x 51
8
1
dx dy = dy
4 50.5 12 4 12 50.5
8
51 − 50.5
= dy
4 12
8
0.5x
=
12 4
2
=
12
ii. P {X ≥ 50.5, Y ≥ 4} =
9 4 3
1 − FX (50.5) − FY (4) + F (50.5, 4) = 1 − − +
12 12 12
2
=
12
3
iii. FY (3) = F (∞, 3) = F (2, 3) = 12
iv. F (2, 5) = P {X ≤ 2, Y ≤ 5}
Notice that P {X ≥ 50.5, Y ≥ 4} = 1 − P {X < 50.5, Y < 4} because
P {X ≥ 50.5, Y ≥ 4} is the “right–back” portion of the distribution,
whereas P {X < 50.5, Y < 4} is the “left–front” portion of the distri-
bution.
(h) Machine B; Figure (b). One randomly chosen filled bag will weigh between
49 and 51 grams and contain between 4 and 6 milligrams of salt with
probability P {49 ≤ X ≤ 51, 4 ≤ Y ≤ 6} = (circle one) 1 / 0.5 / 0.
(i) More Machine B. The probability P {49 ≤ X ≤ 51, 4 ≤ Y ≤ 6} is repre-
sented by or equal to the (circle none, one or more)
i. rectangular box volume equal to 1.
ii. rectangular box volume equal to the width (51 − 49 = 2) times the
depth (6 − 4 = 2) times the height ( 14 ).
iii. definite integral of f (x) = 14 over the region (49, 51) × (4, 6).
176 Chapter 6. Jointly Distributed Random Variables
(j) More Machine B. True / False The joint probability density function is
given by,
1
4
49 ≤ x ≤ 51, 4 ≤ y ≤ 6
f (x, y) =
0 elsewhere
(k) More Machine B. The chance a potato chip bag, chosen at random, weighs
at most 50.5 grams and contains at most 5 grams of salt is (circle none,
one or more)
i. P {X ≤ 50.5, Y ≤ 5} = (1.5)(1) 14 = 1.5
4
ii. F (50.5, 5) =
5 50.5 5 50.5
1 x
dx dy = dy
4 49 4 4 4 49
5
50.5 − 49
= dy
4 4
5
1.5x
=
4 4
1.5
=
4
iii. P {X = 1, Y = 1} + P {X = 2, Y = 1} = 0.01 + 0.02 = 0.03
iv. P {X ≤ 2, Y ≤ 1} = F (2, 1) = 0.11
(l) More Machine B. The chance of potato chip bag, chosen at random, weighs
at most 50.5 grams is (circle none, one or more)
i. P {X ≤ 50.5} = (1.5)(6 − 4) 14 = 3
4
= 0.75
Section 1. Joint Distribution Functions 177
ii. P {X ≥ 50.5, Y ≥ 5} =
3 2 1.5
1 − FX (50.5) − FY (5) + F (50.5, 5) = 1 − − +
4 4 4
0.5
=
4
3
iii. FY (3) = F (∞, 3) = F (2, 3) = 12
iv. F (2, 5) = P {X ≤ 2, Y ≤ 5}
(o) Machine C; Figure (c). One randomly chosen filled bag will weigh, X
between 49 and 51 grams and contain between 2 and 8 milligrams of salt, Y ,
and also the weight and amount of salt obeys the constraint 3X +Y < 155,
with probability P {49 ≤ X ≤ 51, 2 ≤ Y ≤ 8} = (circle one) 1 / 0.5 / 0.
(p) More Machine C. The probability P {49 ≤ X ≤ 51.2 ≤ Y ≤ 8} is repre-
sented by or equal to the (circle none, one or more)
i. pie slice volume equal to 1.
ii. pie slice volume equal to the width (51 − 49 = 2) times one–half the
depth ( 12 (8 − 2) = 3) times the height ( 16 ).
iii. definite integral of f (x) = 16 over the region 49 < X < 51, 2 < Y < 8,
3X + Y < 155.
iv. the integral,
8 8
1 1
dx dy = dx dy
2 3x+y<155 6 2 x<155/3−(1/3)y 6
8 155/3−(1/3)y
x
= dy
2 6 49
8
(155/3 − (1/3)y) − 49
= dy
2 6
8
(8/3)y (1/6)y 2
= −
6 6 2
= 1
(q) More Machine C. True / False The joint probability density function is
given by,
1
6
49 ≤ x ≤ 51, 2 ≤ y ≤ 8, 3x + y < 155
f (x, y) =
0 elsewhere
(r) More Machine C. The chance a potato chip bag, chosen at random, weighs
at most 50.5 grams and contains at most 4 grams of salt is (circle none,
one or more)
Section 1. Joint Distribution Functions 179
∂2
i. ∂x∂y
F (x, y)
∂ 2 2 2
ii. ∂x∂y
(1 − e−(1) )(1 − e−(2) )
−x2 2 2 +y 2 )
iii. [−(−2x)e ] × [−(2y)e−y ] = 2xye−(x
and P {1 < x < 1.25, 1.5 < y < 2} = (circle none, one or more)
i. F (1, 1.5) + F (1.25, 2) − F (1, 2) − F (1.25, 1.5)
2 2 2 2 2
ii. (1 − e−(1) )(1 − e−(1.5) ) + (1 − e−(1.25) )(1 − e−(2) ) − (1 − e−(1) )(1 −
2 2 2
e−(2) ) − (1 − e−(1.25) )(1 − e−(1.5) )
iii. 0.56549 + 0.775912 − 0.620542 − 0.707082 = 0.013778
(Hint: Draw a picture of the rectangular region of integration to convince
yourself that adding and subtracting the joint distributions as given above
is appropriate.)
(b) Determine c so that
cx(3x − y) 0 ≤ x ≤ 2, 0 ≤ y ≤ 1, x + y < 1
f (x, y) =
0 elsewhere
1 4
P {X = 1, Y = 2, Z = 2} = (circle none, one or more) 54 (1)(2)(2) / 54
/
5 6
31
/ 31 .
1 5 8
P {X = 2, Y = 2, Z = 2} = (circle one) 54 / 54 / 54 / 12
54
.
1 5 8 12
P {X ≤ 2, Y = 2, Z = 2} = (circle one) 54 / 54 / 54 / 54 .
(b) Discrete Multinomial Joint Distribution. The multinomial joint distribu-
tion is given by
n!
P {X1 = x1 , X2 = x2 , . . . , Xr = xr } = pn1 pn2 · · · pnr r
n1 !n2 ! · · · nr ! 1 2
where ri=1 ni = n.
Suppose a fair die is rolled 8 times. The chance that 1 appears 3 times,
2 appears once, 3 appears once, 4 appears 3 times, and 5 or 6 does not
appear is P {X = 1, Y = 2, Z = 2} = (circle none, one or more)
i. P {X1 = 3, X2 = 1, X3 = 1, X4 = 3, X5 = 0, X6 = 0}
8!
ii. 3!1!1!3!0!0! (1/6)3 (1/6)1(1/6)1 (1/6)3 (1/6)1(1/6)0
8!
iii. 3!1!1!3!0!0! (1/6)8
iv. 0.0006668
Are the waiting times on the two days independent of one another? To
demonstrate independence, we must show that P {X, Y } = P {X}P {Y } for
X, Y = 1, 2, 3.
(a) X = 3, Y = 3.
The chance of waiting three minutes to catch one fish on the first day is
P {X = 3} = (circle one) 0.09 / 0.11 / 0.12 / 0.78.
The chance of waiting three minutes to catch one fish on the second day is
P {Y = 3} = (circle one) 0.09 / 0.11 / 0.12 / 0.79.
The chance of waiting three minutes to catch one fish on the first day and
waiting three minutes to catch one fish on the second day is
P {X = 3, Y = 3} = (circle one) 0.09 / 0.11 / 0.63 / 0.78.
Since the chance of waiting three minutes to catch one fish on the first day
and waiting three minutes to catch one fish on the second day,
P {X = 3, Y = 3} = 0.63,
(circle one) does / does not equal
P {X = 3}P {Y = 3} = (0.78)(0.79) = 0.6162,
the waiting three minutes on the second day depends on the waiting three
minutes on the first day.
(b) X = 2, Y = 3.
The chance of waiting two minutes to catch a fish on the first day is
P {X = 2} = (circle one) 0.09 / 0.11 / 0.12 / 0.78.
The chance of waiting three minutes to catch a fish on the second day is
P {Y = 3} = (circle one) 0.09 / 0.11 / 0.12 / 0.79.
The chance of waiting two minutes to catch a fish on the first day and
waiting three minutes to catch a fish on the second day is
P {X = 3, Y = 3} = (circle one) 0.08 / 0.11 / 0.63 / 0.78.
Since the chance of waiting two minutes to catch a fish on the first day
and waiting three minutes to catch a fish on the second day,
P {X = 2, Y = 3} = 0.08,
(circle one) does / does not equal
P {X = 2}P {Y = 3} = (0.11)(0.79) = 0.0869,
the waiting three minutes on the second day depends on the waiting two
minutes on the first day.
184 Chapter 6. Jointly Distributed Random Variables
(c) True / False In order for the waiting time on the second day to be in-
dependent of the waiting time on the first day, it must be shown that the
waiting times of one, two or three minutes on the second day must all be
shown to be independent of the waiting times of one, two or three minutes
on the first day. If any of the waiting times on the second day are shown
to be independent of any of the waiting times on the first day, this would
demonstrate the waiting time on the second day depends on the waiting
time of the first day.
2. More Discrete Distribution and Independence: Waiting Time To Fish Again.
The joint density of the number of minutes waiting to catch a fish on the first
and second day, P {X, Y }, is given below.
Are the waiting times on the two days independent of one another? To
demonstrate independence, we must show that P {X, Y } = P {X}P {Y } for
X, Y = 1, 2, 3.
(a) X = 2, Y = 3.
The chance of waiting two minutes to catch a fish on the first day is
P {X = 2} = (circle one) 0.09 / 0.10 / 0.12 / 0.78.
The chance of waiting three minutes to catch a fish on the second day is
P {Y = 3} = (circle one) 0.09 / 0.11 / 0.12 / 0.80.
The chance of waiting two minutes to catch a fish on the first day and
waiting three minutes to catch a fish on the second day is
P {X = 3, Y = 3} = (circle one) 0.08 / 0.11 / 0.64 / 0.80.
Since the chance of waiting two minutes to catch a fish on the first day
and waiting three minutes to catch a fish on the second day,
P {X = 2, Y = 3} = 0.08,
(circle one) does / does not equal
P {X = 2}P {Y = 3} = (0.10)(0.80) = 0.08,
the waiting three minutes on the second day is
(circle one) independent / dependent
on the waiting two minutes on the first day.
(b) True / False In fact, since P {X, Y } = P {X}P {Y } for X, Y = 1, 2, 3, the
waiting time on the second day is independent of the waiting time on the
first day.
Section 2. Independent Random Variables 185
X 1 2 3
P {X = x} 0.1 0.1 0.8
Basketball player B also has a 45% chance of making a free throw, and so the
chance s/he makes a second basket on the fourth throw is (using the negative
binomial distribution)
i−1 r i−r 4−1
P {Y = 4} = p (1 − p) = 0.452 (1 − 0.45)4−2
r−1 2−1
Assume basketball player A’s free throws are independent of basketball player
B’s free throws,
(a) The chance that A makes a basket on the fourth throw and B makes a
second basket on the fourth throw is
186 Chapter 6. Jointly Distributed Random Variables
P {X = 4, Y = 4} =
P {X = 4}P
{Y = 4} = (circle one)
4−1
0.45(1 − 0.45)1 × 0.452 (1 − 0.45)4;2
2−1
2 4−1
0.45(1 − 0.45) × 0.452 (1 − 0.45)4;2
2 − 1
3 4−1
0.45(1 − 0.45) × 0.452 (1 − 0.45)4;2
2−1
(b) The chance that A makes a basket on the third throw and B makes a
second basket on the fifth throw is
P {X = 3, Y = 5} =
P {X = 3}P
{Y = 5} = (circle one)
5−1
0.45(1 − 0.45)1 × 0.452 (1 − 0.45)5;2
2 − 1
2 5−1
0.45(1 − 0.45) × 0.452 (1 − 0.45)5;2
2 − 1
3 5−1
0.45(1 − 0.45) × 0.452 (1 − 0.45)5;2
2−1
(c) Suppose A makes a basket on the third throw, starts again and, indepen-
dent of the first round of throws, makes a basket on the fifth throw on the
second round of throws and then, independent of this, on a third round of
throws, makes a basket on the first attempt. The chance of this happening
is
P {X1 = 3, X2 = 5, X3 = 1} = P {X1 = 3}P {X2 = 5}P {X3 = 1} =
(circle one)
0.45(1 − 0.45)2 × 0.45(1 − 0.45)2 × 0.45
0.45(1 − 0.45)2 × 0.45(1 − 0.45)3 × 0.45
0.45(1 − 0.45)2 × 0.45(1 − 0.45)4 × 0.45
(a) fX (x) =
1
1
4xy dy = 2xy 2 y=0
0
= 2x(1) − 2x(0)2
2
=
(b) fY (y) =
1
1
4xy dx = 2x2 y x=0
0
= 2(1)2y − 2(0)2 y =
(a) fX (x) =
24xy dy = 24xy dy
x+y<1 y<1−x
1−x
= 24xy dy
0
2 1−x
= 12xy y=0
= 12x(1 − x)2 − 12x(0)2 =
(b) FY (y) =
F (∞, y) = lim F (x, y)
x→∞
2 2
= lim (1 − e−x )(1 − e−y )
x→∞
2
= (1 − 0)(1 − e−y ) =
(circle one) 2 / 2x / 2y / 1 − e;y .
2
2 2
(c) Since f (x, y) = (1 − e−x )(1 − e−y )
(circle one) does / does not equal
2 2
fX (x)fY (y) = (1 − e−x )(1 − e−y ),
random variable X is independent of Y .
8. Another Continuous Joint Distribution and Independence. If X and Y are
independent, what is the density of X/Y , if X and Y are both exponential
random variables with parameters λ and µ, respectively?
(a) General Density.
FZ (a) = P {X/Y < a}
= P {X < aY }
∞ ay
= fX (x)fY (y) dx dy
0 0
∞
= FX (ay)fY (y) dy
0
Section 3. Sums of Independent Random Variables 189
and so
∞
d
fZ (a) = FX (ay)fY (y) dy
da 0
(circle
1 one)
01 fX (ay)fY (y) dy
0 1X
f (ay)yfY (y) dy
− 0 fX (ay)yfY (y) dy.
(b) Density When X and Y Are Exponential Random Variables. If X is ex-
ponential with parameter λ and Y is exponential with parameter µ, fZ (a)
(circle
1 one);ay ;y
01 λe;ay µe ;ydy
0 1
λe yµe dy
;ay ;y
− 0 λe yµe dy.
and so
∞ ∞
d
fX+Y (a) = FX (a − y)fY (y) dy = fX (a − y)fY (y) dy
da −∞ −∞
x 1 2 3
P {X = x} 0.4 0.4 0.2
(a) One fish is caught on the first trip and three fish are caught on the second
trip. The sum of the number of fish caught over these two trips is
x1 + x2 = 1 + 3 = (circle one) 0.3 / 1.5 / 4.
(b) Two fish are caught on the first trip and three fish are caught on the second
trip. The sum of the number of fish caught over these two trips is
x1 + x2 = 2 + 3 = (circle one) 0.3 / 1.5 / 5.
(c) The joint distribution probabilities as well as the sum of the number of
fish caught on two trips to the lake are combined in the one table below.
P{x1 , x2 } x2
x1 + x2 1 2 3
1 0.16 0.16 0.08
2 3 4
x1 2 0.16 0.16 0.08
3 4 5
3 0.08 0.08 0.04
4 5 6
The sum for when three fish are caught on the first trip and two fish are
caught on the second trip, (3, 2), is 5 with chance given by
(circle one) 0.04 / 0.08 / 0.16.
(d) The sum, x1 + x2 = 4, occurs in three possible ways: (3,1), (2,2) and (1,3),
with probabilities (circle one)
i. 0.08, 0.08 and 0.08, respectively.
ii. 0.08, 0.16 and 0.16, respectively.
iii. 0.08, 0.16 and 0.08, respectively.
iv. 0.16, 0.16 and 0.08, respectively.
(e) Thus, the chance that the sum of the number of fish caught on two trips
to the lake is four is
P{X1 + X2 = 4} = 0.08 + 0.16 + 0.08 = (circle one) 0.04 / 0.16 / 0.32.
(f) The sum, x1 + x2 = 5, occurs in two possible ways (circle one)
i. (2,2) and (1,3).
ii. (2,3) and (2,3).
Section 3. Sums of Independent Random Variables 191
x 1 2 3
P {X = x} 0.1 0.1 0.8
(a) If two minutes are spent waiting for one fish and two minutes are spent
waiting for another fish, (2,2), the sum of time spent waiting is
x1 + x2 = 2 + 2 = (circle one) 0.3 / 1.5 / 4.
(b) Complete the following table of joint distribution probabilities as well as
the average times spent waiting to catch two fish.
P{x1 , x2 } x1
X1 + x2 1 2 3
1 0.01 0.01 0.08
2 3 4
x2 2 0.01 0.01 0.08
4 5
3 0.08 0.08 0.64
5 6
(c) Complete the probability distribution of the sum of the waiting time, X1 +
X2 , is
x1 + x2 2 3 4 5 6
P {X1 + X2 } 0.02 0.17 0.64
192 Chapter 6. Jointly Distributed Random Variables
3. Discrete: Poisson sum. The Poisson random variable has distribution given by
λi
p(i) = P {X = i} = e−λ , i = 0, 1, . . . , λ > 0
i!
with parameter λ and so the distribution of the sum X + Y is
n
P {X + Y = n} = P {X = k, Y = n − k}
k=0
n
= P {X = k}P {Y = n − k}
k=0
n
λ1 k −λ2 λ2 n−k
= e−λ1 e
k=0
k! (n − k)!
n
λ1 k λ2 n−k
−(λ1 +λ2 )
= e
k=0
k!(n − k)!
e−(λ1 +λ2 )
n
n!
= λ1 k λ2 n−k
n! k=0
k!(n − k)!
e−(λ1 +λ2 )
= (λ1 + λ2 )n binomial coefficient identity
n!
(a) Suppose the number of people who live to 100 years of age in Westville
per year has a Poisson distribution with parameter λ1 = 2.5; whereas,
independent of this, in Michigan City, it has a Poisson distribution with
parameter λ2 = 3. The chance that the sum of the number of people who
live to 100 years of age in Westville and Michigan City is 4 is
P {X1 + X2 = 4} = (circle none, one or more)
−(λ +λ ) e−(2.5+3)
i. e n!1 2
(λ1 + λ2 )n = 4!
(2.5 + 3)4
ii. 0.1558
(Hint: Poissonpdf(5.5,4))
(b) For λ1 = 2.5, λ2 = 3, P {X1 + X2 ≤ 4} =
(circle one) 0.311 / 0.358 / 0.543.
(Hint: Poissoncdf(5.5,4))
(c) For λ1 = 2.5, λ2 = 6, P {X1 + X2 ≤ 4} =
(circle one) 0.074 / 0.358 / 0.543.
Section 3. Sums of Independent Random Variables 193
and so the distribution of the sum X + Y , where Y has parameters (s, λ), is
∞
fX+Y (a) = fX (a − y)fY (y) dy
−∞
a
1
= λe−λ(a−y) (λ(a − y))s−1λe−λy (λy)t−1 dy
Γ(s)Γ(t) 0
a
−λa
= Ke (a − y)s−1y t−1 dy
0
1
−λa s+t−1 y
= Ke a (1 − x)s−1 xt−1 dx letting x =
0 a
−λa s+t−1
= Ce a
λe−λa (λa)s+t−1 1
= letting C =
Γ(s + t) Γ(s + t)
In general, ifXi is gamma with parameter (ti , λ), then ni=1 Xi is gamma with
parameter ( ni=1 t1 , λ).
(a) Suppose the number of people who live to 100 years of age in Westville per
year has a gamma distribution with parameter (t, λ) = (2, 2.5); whereas,
independent of this, in Michigan City, it has a gamma distribution with
parameter (s, λ) = (3, 2.5).
fX1 +X2 (4) = (circle none, one or more)
λe−λa (λa)s+t−1 2.5e−2.5 (2.5(4))2+3−1
i. Γ(s+t)
= Γ(2+3)
−(2.5)(4) (2.5(4))2+3−1
ii. 2.5e (5−1)!
iii. 0.0473
(b) For t = 2, s = 3, P {X1 + X2 ≤ 4} = (circle one) 0.189 / 0.358 / 0.543.
−(2.5)(4) (2.5(4))2+3−1
(Hint: fnInt( 2.5e (5−1)!
, X, 0, 4))
5. Continuous: normal. After some effort, it can be shown that the distribution
n
of the sum of independent random variables, i=1 Xi , where each Xi has a
normal distribution,
n and where
n each has the parameter (µi , σi2 ), is also normal
with parameter ( i=1 µi , i=1 σi2 ). Consequently, the distribution of X1 + X2 ,
where X1 is normal with parameter (2, 3) and X2 is normal with parameter
(1, 1), is normal with parameter (circle one) (2, 3) / (3, 4) / (4, 5).
Review Chapter
Properties of Expectation
7.1 Introduction
If
P {a ≤ X ≤ b} = 1
then
a ≤ E(X) ≤ b
1. Let X be the waiting time for a bus. The chance the waiting time is between
3 and 7 minutes is 100%, P {3 ≤ X ≤ 7} = 1. This means the waiting time is
expected to between (circle one) (3, 7) / (1, 6) / (4, 8) minutes.
2. Let X be the weight of a new–born child. The chance the weight is between
4 and 10 pounds is 100%, P {3 ≤ X ≤ 10} = 1. This means the weight is
expected to between (circle one) (3, 7) / (4, 10) / (4, 8) minutes.
• In general,
y g(x, y)p(x, y) if discrete
E[g(x, y)] = ∞x ∞
−∞ −∞
g(x, y)f (x, y) dxdy if continuous
207
208 Chapter 7. Properties of Expectation
P {X = i, Y = j} j
g(x, y) = x+y
2
1 2 3
1 0.16 0.16 0.08
3
1 2
2
i 2 0.16 0.16 0.08
3 5
2
2 2
3 0.08 0.08 0.04
5
2 2
3
(a) True / False This is a probability density because the probabilities sum
to one.
(b) One fish is caught on the first trip and three fish are caught on the second
trip with probability
(circle one) 0.04 / 0.08 / 16.
The average waiting time over these two trips is
g(1, 3) = 1+3
2
= (circle one) 0.3 / 1.5 / 2.
(c) Two fish is caught on the first trip and three fish are caught on the second
trip with probability
(circle one) 0.04 / 0.08 / 16.
The average waiting time over these two trips is
g(2, 3) = 2+3
2
= (circle one) 0.3 / 1.5 / 2.5.
(d) Three fish is caught on the first trip and two fish are caught on the second
trip with probability
(circle one) 0.04 / 0.08 / 16.
The average waiting time over these two trips is
g(3, 2) = 2+3
2
= (circle one) 0.3 / 1.5 / 2.5.
Section 2. Expectation of Sums of Random Variables 209
(e) The expected average waiting time over these two trips is
E[g(x, y)] = g(x, y)p(x, y)
x y
3
= (1)(0.16) + (0.16) + (2)(0.08)
2
3 3
+ (0.16) + (2)(0.16) + (0.08)
2 2
5
+ (2)(0.08) + (0.08) + (3)(0.04) =
2
(circle one) 1 / 1.5 / 1.8.
P {X = i, Y = j} j row sum
1 2 3 P {X = i}
1 0.01 0.02 0.08 0.11
i 2 0.01 0.02 0.08 0.11
3 0.07 0.08 0.63 0.78
P {Y = j} 0.09 0.12 0.79
column sum
g(x, y, z) = x + y + z
3 4 12 13 14
(a) pX (1) = y=1 z=1 p(1, y, z) = (circle one) 24
/ 24
/ 24
.
3 4 12 13 14
(b) pX (2) = y=1 z=1 p(2, y, z) = (circle one) 24
/ 24
/ 24
.
212 Chapter 7. Properties of Expectation
i. −∞ x 2x2 + 2x 3
dx = 0
2x + 3
dx
3
1
ii. 24 x4 + 2x9
0
2 2 13
iii. 4
+ 9
= 18
Section 2. Expectation of Sums of Random Variables 213
i. −∞ y 13 + y6 dy = 0 y3 + y6 dy
3 2
ii. 16 y 2 + y18
0
1 23 10
iii. 6
(2)2 + 18
= 9
13 10 11
(e) E(X + Y ) = E(X) + E(Y ) = 18
+ 9
= (circle one) 6
/ 3 / 4.
(f) E(X + Y ) = (circle none, one or more)
∞ ∞ 21 xy
i. −∞ −∞ (x + y)f (x, y) dx dy = 0 0 (x + y) x2 + 3
dx dy
21 2 2
2 2 1
iii. 0 x4 + 4x9 y + x 6y dy
2 2
0
iv. 0 14 + 4y 9
+ y6 dy
2
2
y3
v. y4 + 4y18
+ 18
= 116
0
where
1 if ith person selects own ticket
Xi =
0 if ith person does not select their own ticket
(a) Since each person will choose any ticket with equal chance, E(Xi ) =
P {Xi = 1} = (circle one) 0.1 / 0.2 / 0.3.
(b) And so E(X) = E(X1 ) + · · · + E(X10 ) = 10(0.1) = (circle one) 1 / 5 / 10.
In other words, we’d expect one of the ten individuals to choose their own
ticket.
(c) If, instead of ten individuals, n individuals played this game, then we would
expect E(X) = E(X1 ) + · · · + E(Xn ) = n n1 (circle one) 1 / 5 / 10.
• Covariance is defined by
Cov(X, Y ) = E[(X − E(X))(Y − E(Y ))] = E(XY ) − E(X)E(Y )
and has the following properties.
– Cov(X, Y ) = Cov(Y, X)
– Cov(X, X) = Var(X)
– Cov(aX, Y ) = aCov(X, Y )
– Cov( ni=1 Xi , ni=1 Yi ) = ni=1 nj=1 Cov(Xi , Yj )
• In general,
Var( ni=1 Xi ) = ni=1 Var(Xi ) +2 i<j Cov(Xi , Yj ),
P {X = i, Y = j} j row sum
1 2 3 P {X = i}
1 0.01 0.02 0.08 0.11
i 2 0.01 0.02 0.08 0.11
3 0.07 0.08 0.63 0.78
P {Y = j} 0.09 0.12 0.79
column sum
(h) Since Cov( ni=1 Xi , ni=1 Yi ) = ni=1 nj=1 Cov(Xi , Yj ),
Cov(X + Y, Z) = Cov(X, Z)+ Cov(Y, Z) = 0 + 0 =
(circle one) 0 / 0.335 / 0.545.
(i) Since Cov( ni=1 Xi , ni=1 Yi ) = ni=1 nj=1 Cov(Xi , Yj ),
Cov(X, Y + Z) = Cov(X, Y )+ Cov(X, Z) = 0 + 0 =
(circle one) 0 / 0.335 / 0.545.
(j) The expected value of g(x, y, z) = x2 is
E[X 2 ] = x2 pX (x)
x
2 12 2 12
= (1) + (2) =
24 24
60
(circle one) 24
/ 1.5 / 2.5.
60
(k) Var(X) = E(X 2 ) − [E(X)]2 = 24
− (1.5)2 = (circle one) 0 / 0.25 / 0.545.
(l) The expected value of g(x, y, z) = y 2 is
E[Y 2 ] = y 2 pY (y)
y
2 8 2 8 2 8
= (1) + (2) + (3) =
24 24 24
112
(circle one) 1 / 1.5 / 24
.
(m) Var(Y ) = E(Y 2 ) − [E(Y )]2 = 112
24
− (2)2 = (circle one) 0 / 0.5 / 23 .
(n) Since Var( ni=1 Xi ) = ni=1 Var(Xi ) +2 i<j
Cov(Xi , Yj ),
2
Var(X + Y ) = Var(X) + Var(Y ) +2 Cov(X, Y ) = 0.25 + 3
+ 2(0) =
9
(circle one) 12 / 10
12
/ 11
12
.
3. Variance of Rolling Dice. Calculate the expectation and variance of the sum of
15 rolls of a fair die.
6roll, Xi = 1, 2, 3, 4, 5, 6,
(a) For the ith
E[Xi ] = j=1 xp(x) = 1(1/6) + 2(1/6) + · · · + 6(1/6) =
(circle one) 32 / 52 / 72 .
(b) and so
15
10
E Xi = E[Xi ]
i=1 i=1
= 15E[Xi ]
7
= 15 =
2
75 90 105
(circle one) 2
/ 2
/ 2
.
Section 3. Covariance, Variance of Sums and Correlations 219
(c) Since E[Xi2 ] = 6j=1 x2 p(x) = 12 (1/6) + 22 (1/6) + · · · + 62 (1/6) = 91
6
(circle one) 88
6
/ 91
6
/ 95
6
.
and so Var(Xi ) = E[Xi ] − (E[Xi ])2 = 91/6 − (7/2)2 =,
2
(circle one) 30
12
/ 35
12
/ 40
12
.
(d) Since Xi are independent,
10
10
V Xi = V [Xi ]
i=1 i=1
= 15V [Xi ]
35
= 15 =
12
165 170 175
(circle one) 4
/ 4
/ 4
.
2 2
ii. yx2 + xy6
0
2 2x
iii. 2x + 3
(b) E(X) = (circle none, one or more)
∞ 1 3 2x2
i. −∞ x 2x2 + 2x 3
dx = 0
2x + 3
dx
3
1
ii. 24 x4 + 2x9
0
2 2 13
iii. 4
+ 9
= 18
(c) fY (y) = (circle none, one or more)
∞ 1 2 xy
i. −∞ x2 + xy 3
dx = 0
x + 3 dx
2
1
ii. 13 x3 + x6y
0
1 y
iii. 3
+ 6
(d) E(Y ) = (circle none, one or more)
∞ 1 y 2 y y2
i. −∞ y 3 + 6 dy = 0 3 + 6 dy
3 2
ii. 16 y 2 + y18
0
1 23 10
iii. 6
(2)2 + 18
= 9
220 Chapter 7. Properties of Expectation
(e) E(XY ) =
∞ ∞ 21 xy
i. −∞ −∞ (xy)f (x, y) dx dy = 0 0 (xy) x2 + 3
dx dy
21 2 2
ii. 0 0 x3 y + x 3y dx dy
2 4
3 2 1
iii. 0 x4y + x 9y dy
2 2
0
iv. 0 y4 + y9 dy
2
3 2
v. y8 + y27 =
0
43 44 45
(circle one) 54
/ 54
/ 54
.
43
13 10
(f) Cov(X, Y ) = E(XY ) − E(X)E(Y ) = 54
− 18 9
=
3
(circle one) −0.0062 / 0.0062 / 18 .
(g) E(X 2 ) =
∞ 2x
1 4 2x3
i. −∞ x2 2x2 + 3
dx = 0
2x + 3
dx
4
1
ii. 25 x5 + 2x
12
0
2 2
iii. 5
+ 12
=
16 17 18
(circle one) 30
/ 30
/ 30
.
17
13 2 73 17 18
(h) Var(X) = E(X 2 ) − [E(X)]2 = 30
− 18
= (circle one) 1620
/ 30
/ 30
.
(i) E(Y 2 ) =
∞ 2 2 y3
i. −∞ y 2 13 + y6 dy = 0 y3 + 6
dy
2
1 3 y4
ii. 9 y + 24
0
1 24
iii. 9
(2)3 + 24
=
13 14 15
(circle one) 9
/ 9
/ 9
.
14
10 2 26 17 18
(j) Var(Y ) = E(Y 2 ) − [E(Y )]2 = 9
− 9
= (circle one) 81
/ 30
/ 30
.
(k) Cov(5X, Y ) = 5 Cov(XY ) = 5(−0.0062) =
(circle one) −0.021 / −0.031 / −0.123.
(l) Cov(X, 4Y ) = 4 Cov(XY ) = 4(−0.0062) =
(circle one) −0.021 / −0.063 / −0.123.
Cov(X,Y )
(m) Correlation. ρ(x, y) = √ =√ −0.0062
=
Var(X)Var(Y ) (73/1620)(26/81)
(circle one) −0.024 / −0.052 / −0.084.
5. Covariance of Binomial. Let X be a binomial random variable with parameters
n and p and let
X = X1 + X2 + · · · + Xn
Section 4. Conditional Expectation 221
where
1 if ith trial is a success
Xi =
0 if ith trial is a failure
• computing expectation
by conditioning,
E[X] = E[E(X|Y )] = y E(X = x|Y = y)P (Y = y), discrete
∞
E[X] = −∞ E(X = x|Y = y)fY (y) dy, continuous
• conditional variance,
Var(X|Y = y) = E[(X − E(X|Y = y))2|Y = y] = E(X 2 |Y = y) − (E(X|Y =
y))2,
Var(X) = E[Var(X|Y )] + Var(E(X|Y ))
P {X = i, Y = j} j row sum
1 2 3 P {X = i}
1 0.01 0.02 0.08 0.11
i 2 0.01 0.02 0.08 0.11
3 0.07 0.08 0.63 0.78
P {Y = j} 0.09 0.12 0.79
column sum
(circle one) 16 / 26 / 41 / 68
6 6
41
29 2
Var(X|Y = 2) = E(X 2 |Y = 2) − (E(X|Y = 2))2 = 6
− 12
=
1
(circle one) 144 / 143
144
/ 247
144
/ 368
144
(j) Compute Var(X|Y = 3). 2
2 2 8 2 8
Since E[X
|Y = 3] = x x P {X = x|Y = 3} = (1) 79
+ (2) 79
+
(3)2 63
79
=
1 2
(circle one) 79 / 79 / 67 / 607
79 79 2
Var(X|Y = 3) = E(X 2 |Y = 3) − (E(X|Y = 3))2 = 607 79
− 213
79
=
1 232 247 2584
(circle one) 6241 / 6241 / 6241 / 6241
(k) Show Var(X) = E[Var(X|Y )] + Var(E(X|Y )).
Since E[Var(X|Y )] = (circle none, one or more)
i. y Var(X|Y = y)P (Y = y) = Var(X|Y = 1)P (Y = 1) + Var(X|Y =
2)P (Y = 2) + Var(X|Y = 3)P (Y = 3)
4 2584
ii. 9 (0.09) + 143
144
(0.12) + 6241
(0.79)
iii. 0.486255
and Var(E(X|Y )) = E[(E[X|Y ])2 ] − (E[E(X|Y )])2 = E[(E[X|Y ])2 ] −
(E[X])2
where E[(E[X|Y ])2 ] = (circle none, one or more)
2 2
i. y (E[X|Y = y]) P (Y = y) = (E[X|Y = 1]) P (Y = 1) + (E[X|Y =
2])2 P (Y = 2) + (E[X|Y = 3])2 P (Y = 3)
2 29 2 213 2
ii. 249
(0.09) + 12
(0.12) + 79
(0.79)
iii. 7.083744
and (E[X])2 = (circle none, one or more)
i. ( x xP (X = x))2 = ((1)P (X = 1) + (2)P (X = 2) + (3)P (X = 3))2
ii. ((1)(0.11) + (2)(0.11) + (3)(0.78))2
iii. 2.672 = 7.1289
And so Var(E(X|Y )) = E[(E[X|Y ])2 ] − (E[X])2 = 7.083744 − 7.1289 =
−0.045156
And so Var(X) = E[Var(X|Y )] + Var(E(X|Y )) = 0.486255 − 0.045156 =
0.441099.
2
xy 2
ii. yx2 + 6
0
x2
iii. 2x2 + 3
(b) fY (y) = (circle none, one or more)
∞ 1
i. −∞ x2 + xy 3
dx = 0 x2 + xy
3
dx
2 2 1
ii. 13 x3 + x 6y
0
1 y2
iii. 3
+ 6
(c) Compute E[X|Y = 1].
Since fX|Y (x|y) = (circle none, one or more)
f (x,y)
i. fY (y)
x2 + xy
3
ii. 1 y2
3
+ 6
6x2 +2xy
iii. 2+y 2
(circle one) 18 / 18 / 18 / 15
12 13 14
18
(d) Compute E[X|Y = 1.5].
6
4
+ 23 (1.5)
E[X|Y = 1.5] = 2+(1.5)2
=
(circle one) 10
17
/ 13
17
/ 14
17
/ 17
17
(e) Compare E[E[X|Y ]] to E[X].
Since E[E(X|Y )] = (circle none, one or more)
∞ 2 9+4y
1 y2
ii. 0 2x + 3 dx
226 Chapter 7. Properties of Expectation
1
1 4 2x3
iii. 2
x + 9
x=0
1 2 13
iv. 2
+ 9
= 18
In other words, E[X] (circle one) does / does not equal E[E(X|Y )].
(f) Compute E(X 2 |Y = 1).
E[X 2 |Y ] = (circle none, one or more)
∞ 2
1 4 +2x3 y
i. −∞ x2 6x2+y +2xy
2 dx = 0 6x2+y 2 dx
6 5 2 4
1
x + x y
ii. 5 2+y42
x=0
6 2
+ y
5 4 12+5y
iii. 2+y 2
= 20+10y 2
12+5(1)
and so E[X|Y = 1] = 20+10(1) 2 =
12 13 14 17
(circle one) 30 / 30 / 30 / 30
(g) Compute Var(X|Y = 1).
17
13 2
Var(X|Y = 1) = E(X 2 |Y = 1) − (E(X|Y = 1))2 = 30
− 18
=
1 3 4 73
(circle one) 1620 / 1620 / 1620 / 1620
(h) Determine Var(X|Y ).
Var(X|Y ) = (circle none, one or more)
i. E(X 2 |Y ) − (E(X|Y ))2
2
12+5y 9+4y
ii. 20+10y 2 − 12+6y 2
3. Prisoner’s Escape and Three Doors. A prisoner is faced with three doors. The
first door leads to a tunnel that leads to freedom in 4 hours. The second door
leads to a tunnel that returns the prisoner back to the prison in 5 hours. The
third door leads to a tunnel that returns the prisoner back to the prison in 10
hours. Assume the prisoner is equally likely to choose any door. Let X represent
the amount of time until the prisoner reaches freedom and let Y represent the
door (1, 2 or 3) he chooses. What is the expected length of time until the
prisoner reaches safety, E[X]?
Limit Theorems
8.1 Introduction
We look at various limit theorems used in probability theory, including some laws of
large numbers and some central limit theorems.
243
244 Chapter 8. Limit Theorems
σ2
P {|X − µ| ≥ k} ≤
k2
σ2 1
P {|X − µ| ≥ σk} ≤ 2 2
= 2
σ k k
1
P {|X − µ| ≤ σk} ≥ 1 − 2
k
1
P {µ − σk ≤ X ≤ µ + σk} ≥ 1 − 2
k
(b) True / False The expected Ph level and standard deviation in Ph level
are µ = 10.55 and σ = 3.01, respectively.
(Hint: Type the Ph levels into L1 , then STAT CALC 1:1–Var Stats.)
(c) The Ph level one standard deviation above the average is equal to µ + σ =
10.55 + 3.01 = 13.56. The Ph level two standard deviations below the
average is equal to µ − 2(3.01) = 4.53. Determine (a), (b), (c), (d) and (e)
in the figure and then fill in the table below.
246 Chapter 8. Limit Theorems
at least 89%
10.55 - 2(3.01) 10.55 + 2(3.01)
at least 75%
at least 0%
(d) The smallest Ph level, 4.3, is (circle one) inside / outside the interval
between 7.54 and 13.56. Also, the Ph level, 10.5, is (circle one) inside /
outside the interval (7.54, 13.56).
(e) Ph levels that are within one standard deviation of the average, refers to
Ph levels that are (circle one) inside / outside the interval (7.54, 13.56).
Ph levels that are within two standard deviations of the average, refers to
Ph levels that are (circle one) inside / outside the interval (4.53, 16.57).
(f) Instead of saying “Ph levels that are within one standard deviation of
the mean”, it is also possible to say “Ph levels are within k standard
deviations of the mean”, where k = 1. If the Ph levels are within two
standard deviations of the average,
then k = 1 / 2 / 3
If the Ph levels are within two and a half standard deviations of the average,
then k = 1 / 1.5 / 2.5
(g) If k = 1.5, then 1 − k12 = 1 − 1.51 2 ≈ 0.56 or 56%.
If k = 2, then 1 − k12 = 14 / 24 / 34
which is equal to 25% / 50% / 75%.
(h) Chebyshev’s inequality,
1
P {µ − σk ≤ X ≤ µ + σk} ≥ 1 −
k2
1 3
P {10.55 − 2σ ≤ X ≤ 10.55 + 2σ} ≥ 1 − 2 = ,
2 4
allows us to say, at least a 1 − k12 = 0.75 proportion or 75% of the 28 Ph
levels should be within two (k = 2) standard deviation of the average.
In fact, 27 of the 28 Ph levels (look at the data above and see for yourself),
Section 2. Chebyshev’s Inequality and the Weak Law of Large Numbers 247
or 27
28
= 0.964 or 96.4%, are in the interval (4.53, 16.57). Chebyshev’s
inequality (circle one) has / has not be violated in this case.
(i) Using Chebyshev’s inequality, what proportion should fall within k = 3
standard deviations of the average?
1 − 312 = (circle one) 34 / 67 / 89
In fact, what proportion of the Ph levels are actually in this interval (count
the number in the interval (1.52,19.58))?
(circle one) 26
28
/ 27
28
/ 28
28
(j) Using Chebyshev’s inequality, what proportion should fall within k = 2.5
standard deviations of the average?
1 − 2.51 2 = (circle one) 20
25
/ 21
25
/ 22
25
In fact, what proportion of the Ph levels are actually in this interval (count
the number in the interval (3.025,18.075))?
(circle one) 2628
/ 27
28
/ 28
28
(k) Since at least 75% or 21 Ph levels are inside the interval (4.534, 16.574),
then at most
(circle one) 25% / 35% / 45%
of the levels are outside the interval (4.534, 16.574).
(a) Using your calculators, P {1 < X < 9} = (circle one) 0.68 / 0.75 / 0.95
(Hint: normalcdf(1,9,5,2))
(b) Using Chebyshev’s inequality, P {1 < X < 9} =
1
P {µ − kσ ≤ X ≤ µ + kσ} ≥ 1 −
k2
1
P {5 − 2(2) ≤ X ≤ 5 + 2(2)} ≥ 1 − 2 =
2
(circle one) 0.68 / 0.75 / 0.95
(c) So, although Chebyshev’s inequality is correct, it (circle one) is / is not
a good approximation to the correct probability in this case.
σ2
then, from Chebyshev’s inequality, P {|X − µ| ≥ σk} ≤ σ2 k 2
,
X1 + · · · Xn σ2
P − µ ≥ ε ≤ 2
n nε
(circle one) 3 / 4 / 6
(d) Based on past experience, the mean test score is µ = 50 and the variance in
the test scores is σ 2 = 15. Determine the probability that the average score
of 40 students will between 39 and 61. The weak law of large numbers,
X1 + · · · Xn σ2
P − µ ≤ ε ≥ 1 − 2 ,
n nε
where µX =
(b) Another Sum. Suppose X has a (any!)
distribution −1.7 and
σX = 1.6. If n = 43, then determine P −76 < 43 i=1 X i < −71 .
i. µP Xi = nµ = 43(−1.7) = (circle one) −73.5 / −73.1 / −72.9.
√
ii. σP Xi = σ n = 1.6 43 = (circle one) 9.5 / 9.8 / 10.5.
√
iii. P −76 < 43 i=1 Xi < −71 ≈ (circle one) 0.09 / 0.11 / 0.19.
(2nd DISTR normalcdf(-76,-71,-73.1,10.5))
(c) And Yet Another Sum. Suppose X has a distribution where µX = 0.7 and
σX
= 1.1.
If n = 51, then
P 34.5 < 51 i=1 Xi < 35.1 ≈ (circle one) 0.01 / 0.02 / 0.03.
v. Normal
35 approximation,
P i=1 Xi > 9 ≈ (circle one) 0 / 0.11 / 0.43.
(2nd DISTR normalcdf(9,E99,8.75,1.48))
(g) Negative Binomial Sum. Suppose X has a negative binomial distribution
where the number of trials is i = 500, the required number of successes
are r = 4 and chance of success on each trial is p = 0.3. Determine
the chance of the sum of 35 independent identically
35 distributed random
variables, X, is greater than 450 successes, P i=1 X i > 450 , using the
normal approximation.
i. E[X] = µ = pr = (circle one) 12.3 / 13.3 / 14.3.
ii. µP Xi = nµ = 35(13.3) = (circle one) 465.5 / 470.5 / 495.5.
iii. Var[X] = σ 2 = r(1−p)
p2
= (circle one) 29.3 / 31.1 / 34.3.
√ √
iv. σP Xi = σ n = 31.1 35 = (circle one) 33 / 35 / 37.
√
v. Normal
35 approximation,
P i=1 Xi > 450 ≈ (circle one) 0 / 0.68 / 0.75.
(2nd DISTR normalcdf(450,E99,465.5,33))
(h) Dice Sum. What is the
chance,in 30 rolls of a fair die, that the sum is
between 100 and 105, P 100 < 30 i=1 Xi < 105 , using the normal approx-
imation.
i. E[X] = µ = 1 16 + · · · + 6 16 = (circle one) 2.3 / 3.5 / 4.3.
ii. µP Xi = nµ = 30(3.5) = (circle one) 100 / 105 / 110.
iii. E[X 2 ] = µ = 12 16 + · · · + 62 16 = (circle one) 2.3 / 3.5 / 15.2.
iv. Var[X] = E[X 2 ] − (E[X])2 = 15.2 − 3.52 (circle one) 2.9 / 3.1 / 3.3.
√ √
v. σP Xi = σ n = 2.9 30 = (circle one) 9.1 / 9.4 / 9.7.
√
vi. Normal
approximation,
P 100 < 30 i=1 Xi < 105 ≈ (circle one) 0 / 0.20 / 0.35.
(2nd DISTR normalcdf(100,105,105,9.4))
Section 3. The Central Limit Theorem 251
(a) The distributions of the average number of fish caught at a lake, X̄, where
n = 1, 2, 3 are given by
x, n = 1 1 2 3
P (X = x) 0.4 0.4 0.2
where µX = 1.8 and σX = 0.75,
3 5
x̄, n = 2 1 2
2 2
3
P (X̄ = x̄) 0.16 0.32 0.32 0.16 0.04
0.75
√
where µX̄ = 1.8 and σX̄ = 2
= 0.53,
4 5 7 8
x̄, n = 3 1 3 3
2 3 3
3
P (X̄ = x̄) 0.064 0.192 0.288 0.256 0.144 0.048 0.008
where µX̄ = 1.8 and σX̄ = 0.75
√ = 0.43. The probability histograms of these
3
three sampling distributions are given below.
0.35 0.30
0.40
P(X = x) = P(X = x)
0.30 0.25
_
0.35
P(X = x)
_ _
0.30 0.25
P(X = x)
0.20
_
_ _
0.25 0.20
0.15
0.20 0.15
0.15 0.10
0.10
0.10 0.05
0.05
0.05
0.00 0.00
0.00 X 1 4/3 5/3 2 7/3 8/3 3 X
1 2 3 X 1 3/2 2 5/2 3