MAT 271 Probability and Statistics Lecture 2: Sample Space and Probability
MAT 271 Probability and Statistics Lecture 2: Sample Space and Probability
1 Introduction
2 Sets
3 Probabilistic Models
4 Conditional Probability
6 Independence
7 Counting
8 Summary
Introduction
Introduction
∎ If I flip a coin lot of times, around 50% of the results would be heads.
∎ Illiad and Oddysey were written by the same person with probability 90%
▶ We will also learn some of the most popular and useful probability
theorems: Total Probability Theorem and Bayes’ Rule.
Sets
Sets
S = {x1 , x2 , . . . }
∎ Sets that can be written like this are called countable sets.
∎ Set of positive integers Z, set of even integers Z2n and set of rational
numbers Q are all countable.
S c = {x∣x ∉ S}
∎ Note that Ωc = ∅
▶ Union of two sets S and T , called S ∪ T is the set that contains all
of their elements
S ∪ T = {x∣x ∈ S or x ∈ T }
S ∩ T = {x∣x ∈ S and x ∈ T }
Set Operations
S∪T = T ∪S
S ∩ (T ∩ U ) = (S ∩ T ) ∩ U
S ∩ (T ∪ U ) = (S ∩ T ) ∪ (S ∩ U )
S ∪ (T ∩ U ) = (S ∪ T ) ∩ (S ∪ U )
(S c )c = S
S ∩ Sc = ∅
S∪Ω = Ω
S∩Ω = S
P (Ω) = 1.
Probability Laws
∎ You can think P (A) as the collective mass that was assigned to elements
of A. Then the additivity axiom becomes intuitive.
∎ Note that we are abusing the notation here by calling P ({si }) as P (si )
Discrete Models
▶ For the special case where each event is equally likely we have:
number of elements in A
P (A) = .
n
▶ Example: Romeo and Juliet have a date and each arrive with a delay
of 0 to 1 hour. The first to arrive waits for at most 15 minutes and
then leaves. What is the probability that they will meet?
Continuous Models
P (A∣B)
∎ How can we derive this probability law for givenP (A), P (B) etc?
Conditional Probability
P (A ∩ B)
P (A∣B) =
P (B)
▶ Example: Consider toss a fair coin three times. Let A be the event
more tails than heads come up and let B be the event 1st toss is a
head. Compute P (A∣B).
∎ In this case we can start with P (B) and P (A∣B), and then use the
following formula to compute P (A ∩ B)
P (A ∩ B) = P (A∣B)P (B)
▶ Example(Monty Hall): You are in a TV Show, and you are told that
the grand prize is equally likely to be found behind any of the closed
three doors. You point to one of the doors, and then host opens one
of the remaining two doors, after making sure the prize is not behind
it. Then the host gives you a chance to switch. Would you stick to
your initial choice, or switch to the unopened door? What is the best
strategy?
Total Probability Theorem and Bayes’ Rule
Total Probability Theorem
∎ There are number of causes (Ai ), which may explain a certain effect (B).
∎ Then by using Bayes’ Rule, we can compute P (Ai ∣B) (given that effect is
observed, what is the probability that this particular effect caused it?)
∎ Finally, by comparing different P (Ai ∣B), we can infer what was the most
probable cause of effect.
▶ Example: Return to the chess example. Given that you won the
match, what is the probability that your opponent was type I?
P (A∣B) = P (A)
P (A ∩ B) = P (A)P (B)
∎ If two physical processes are not interacting with each other, events
caused by them are usually independent.
∎ Note that A and Ac are disjoint, but knowing one of them has occurred,
determines the other exactly.
Independence
P (A∣B ∩ C) = P (A∣C)
▶ Consider two coins, a blue one and a red one. We choose one at
random and proceed with two independent tosses. Coins are biased,
probability of head is 0.99 with blue coin and 0.01 with red coin. Let
B be the event that blue coin is chosen and let Hi be the event that
ith toss results in a head. Show that H1 and H2 are dependent, but
they become conditionally independent conditioned on B.
Independence
Definition (Independence)
▶ Two events A and B are said to be independent if
P (A ∩ B) = P (A)P (B).
P (A∣B) = P (A)
P (A∣B ∩ C) = P (A∣C)
Independence of Multiple Events
pk (1 − p)n−k
Independent Trials and Binomial Probabilities
▶ Of course, Bernoulli trials are not only used to analyze coin tosses,
let’s look at a more realistic example.
▶ An Internet service provider has installed c modems to serve n
customers. It is estimated that each customer will need a connection
with probability p. What is the probability that more than c
customers will simultaneously need a connection?
▶ The answer is:
n n
n k
∑ p(k) = ∑ ( )p (1 − p)
k
k=c+1 k=c+1 k
▶ For the rest of the subsection, we will focus on three different types
of counting that involves selecting k objects out of n:
∎ If the order of the elements matters, it is called permutation.
∎ Otherwise, it is called combination.
∎ Finally, we will discuss a more general type of counting that involves
partition of a collection into multiple subsets.
n(n − 1) . . . (n − k)(n − k − 1) . . . 1 n!
n(n − 1) . . . (n − k) = =
(n − k − 1) . . . 1 (n − k)!
n(n − 1) . . . 1 = n!
k−permutations
▶ Example: You have n1 classical music CDs, n2 rock music CDs, and
n3 country music CDs. In how many different ways can you arrange
them so that CDs of the same type are next to each other?
Combination
∎ Can you see why this is also the formula for the binomial coefficient (nk)?
Combination
k=0 k
∎ Result is the number of subsets of an n element set. Can you see why?
▶ Example: Consider a group of n people. Consider clubs that consists
of a club leader and a number of (possibly zero) additional club
members. What is the number of possible clubs of this type?
∎ Note that solution leads to an interesting identity:
n
n
∑ k( ) = n2
n−1
k=1 k
Partitions
▶ This lecture:
▶ What is next?