Chapter 1 - Section 2
Chapter 1 - Section 2
Probabilty courses
2nd year of preparatory classes
Plan:
1. Probability Mass Function and Cumulative Distribution Function
2. Moments of a random variable
3. Generating functions
4. Transformations of random variables
5. Markov and Chebyshev inequalities
1. Probability Mass Function (PMF) 4
and Cumulative Distribution Function (CDF)
Introductory example
𝑥i 1 2 3 4 5 6 ∑
𝑃(𝑋=𝑥i) 1/21 2/21 3/21 4/21 5/21 6/21 1
Definition:
For a discrete random variable X, the probability mass
function is the function fX(𝑥i) that assigns to each realization
𝑥i ∊ 𝑋(Ω), a probability 𝑃(𝑋=𝑥i) :
Definition1
The cumulative distribution function or briefly the distribution
function 𝐹X(𝑥), for a random variable X is given by:
∀ 𝑥 ∈ ℝ, 𝐹𝑋 (𝑥) = 𝑃(𝑋 ≤ 𝑥)
9
Definition 2
For a discrete random variableX ,the distribution function de is
the mapping 𝐹𝑋 such as :
𝐹𝑋 (𝑥) : ℝ → [0,1]
𝑥 ↦ 𝑭𝑿 (𝒙) = 𝐏(𝐗 ≤ 𝒙) = 𝒙𝒊 ∊𝑿 𝜴 , 𝒙𝒊 ≤𝒙 𝑷(𝑿 = 𝒙𝒊 )
𝑭𝑿 𝒙 = 𝑷 𝑿=𝒕 with 𝒕 ∊ 𝑿 𝜴 ,
𝒕=−∞
10
Properties :
From this definition, several properties of the CDF can be
inferred :
𝐹𝑋 is non-decreasing
𝐹𝑋 is right-continuous
lim F𝑋 𝑥 = 0
𝑥→ −∞
lim F𝑋 (𝑥) = 1
𝑥→ +∞
Probabilities calculation 11
Probabilities of events for a variable X can be given in terms
of its distribution function 𝐹𝑋 (𝑥).
∀ (𝑥i , 𝑥j) ∊ℝ2 with 𝑥j > 𝑥i we have:
𝟏𝟓 𝟏 𝟏𝟒
𝑃(𝑥i ≤ 𝑋 ≤ 𝑥j ) = 𝐹𝑋(𝑥j ) −𝐹𝑋 (𝑥i-1) 𝑃( 2 ≤ 𝑋 ≤ 5 ) = 𝐹𝑋(5 ) −𝐹𝑋 (1) = − =
𝟐𝟏 𝟐𝟏 𝟐𝟏
𝟏𝟎 𝟏 𝟗
𝑃(𝑥i ≤ 𝑋 < 𝑥j ) = 𝐹𝑋(𝑥j -1 ) −𝐹𝑋 (𝑥i -1) 𝑃( 2 ≤ 𝑋 < 5) = 𝐹𝑋(4) −𝐹𝑋 (1) = − =
𝟐𝟏 𝟐𝟏 𝟐𝟏
𝟏𝟎 𝟑 𝟕
𝑃(𝑥i < 𝑋 < 𝑥j ) = 𝐹𝑋(𝑥j -1 ) −𝐹𝑋 (𝑥i) 𝑃( 2 < 𝑋 < 5) = 𝐹𝑋(4) −𝐹𝑋 (2) = − =
𝟐𝟏 𝟐𝟏 𝟐𝟏
𝟏𝟎 𝟏𝟏
𝑃(𝑋 > 𝑥i ) = 1−𝑃(𝑋 ≤ 𝑥i ) =1 −𝐹𝑋(𝑥i ) 𝑃(𝑋 > 4 ) = 1−𝑃(𝑋 ≤ 4 ) =1 −𝐹𝑋(4 ) = 𝟏 − =
𝟐𝟏 𝟐𝟏
𝟔 𝟏𝟓
𝑃(𝑋 ≥ 𝑥i ) = 1−𝑃(𝑋 < 𝑥i ) =1 − 𝐹𝑋 (𝑥i -1) 𝑃(𝑋 ≥ 4 ) = 1−𝑃(𝑋 < 4 ) =1 −𝐹𝑋(3 ) = 𝟏 − =
𝟐𝟏 𝟐𝟏
𝟏𝟎 𝟔 𝟒
𝑃(𝑋 = 𝑥i ) = 𝐹𝑋(𝑥i ) − 𝐹𝑋(𝑥 i-1 ) 𝑃(𝑋 = 4 ) = 𝐹𝑋(4 ) − 𝐹𝑋(3) = − =
𝟐𝟏 𝟐𝟏 𝟐𝟏
13
2. Moments of a random variable
(Moments d’une variable aléatoire)
Définition
Let 𝑋 be a discrete random variable defined on 𝑋(Ω) by
the mass function 𝑃(𝑋=𝑥) ,
The expected value of 𝑋, denoted E(𝑋) or µ is defined by:
𝒏
𝝁=𝐄 𝑿 = 𝒙𝒊 𝐏 𝑿 = 𝒙𝒊 𝒙𝒊 ∊ 𝑿(𝛀)
𝒊=𝟏
Example 15
𝑥i 1 2 3 4 5 6 ∑
𝑃(𝑋=𝑥i) 1/21 2/21 3/21 4/21 5/21 6/21 1
𝒏
𝐄 𝑿 = 𝒙𝒊 𝐏 𝑿 = 𝒙 𝒊
𝒊=𝟏
𝟏 𝟐 𝟑 𝟒 𝟓 𝟔 𝟏𝟑
𝐄 𝐗 =𝟏× + 𝟐× + 𝟑× + 𝟒× + 𝟓× + 𝟔× = = 𝟒, 𝟑𝟑
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟑
16
Expected value of a function
Let 𝑋 be a discrete random variable defined on 𝑋(Ω)
by the mass function 𝑃(𝑋=𝑥) ,
𝑥i 1 2 3 4 5 6 ∑
𝑃(𝑋=𝑥i) 1/21 2/21 3/21 4/21 5/21 6/21 1
𝐄 (𝒉(𝐗)) = 𝐄 𝑿𝟐 = 𝒙 𝒊 𝟐 𝐏 𝑿 = 𝒙𝒊
𝒊=𝟏
𝟏 𝟐 𝟑 𝟒 𝟓 𝟔
𝐄 𝑿𝟐 =𝟏× +𝟒× +𝟗× + 𝟏𝟔 × + 𝟐𝟓 × + 𝟑𝟔 × = 𝟐𝟏
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
18
Properties
E (𝑋+ a ) = E (𝑋)+ a ∀ a ∊ℝ
E (a 𝑋 ) = a E (𝑋)
𝐕 𝐗 = 𝑬 (𝒙𝒊 − 𝐄 𝐗 )𝟐
Then:
𝒏
𝑥i 1 2 3 4 5 6 ∑
𝑃(𝑋=𝑥i) 1/21 2/21 3/21 4/21 5/21 6/21 1
𝒏
𝟏
𝟐 𝟐
𝟐 𝟐
𝟑
𝐕 𝐗 = 𝟏 − 𝟒, 𝟑𝟑 × + 𝟐 − 𝟒, 𝟑𝟑 × + 𝟑 − 𝟒, 𝟑𝟑 ×
𝟐𝟏 𝟐𝟏 𝟐𝟏
𝟐
𝟒 𝟐
𝟓 𝟐
𝟔
+ 𝟒 − 𝟒, 𝟑𝟑 × + 𝟓 − 𝟒, 𝟑𝟑 × + 𝟔 − 𝟒, 𝟑𝟑 × = 𝟐, 𝟐𝟐
𝟐𝟏 𝟐𝟏 𝟐𝟏
* Standard deviation (L’écart-type) 21
𝜎(𝑋) = 𝐕(𝐗)
Example:
In the previous example: 𝑽(𝑿) = 𝟐, 𝟐𝟐
𝝈(𝑿) = 𝟐, 𝟐𝟐 = 𝟏, 𝟒𝟗
Properties 22
V (𝑋+ a ) = V (𝑋) ∀ a ∊ℝ
V (a 𝑋 ) = a2 V (𝑋)
𝐕 𝐗 = 𝐄 𝐗 𝟐 − (𝐄(𝐗))𝟐
Example:
In the previous example, we found 𝑽(𝑿) = 𝟐, 𝟐𝟐
We can find the same result using König-Huygens form:
𝟐
𝟐
𝟏𝟑 𝟐𝟎
𝑽 𝑿 =𝑬 𝑿 − (𝑬 𝑿 )𝟐 = 𝟐𝟏 − = = 𝟐, 𝟐𝟐
𝟑 𝟗
24
2.2. Other moments
(Autres moments)
𝒎𝒌 = 𝐄 𝐗 𝒌 = 𝒙𝒌𝒊 𝐏(𝐗 = 𝒙𝒊 )
𝒊=𝟏
Example
The 1st order raw moment in the rolling die example is given by :
𝟏 𝟐 𝟑 𝟒 𝟓 𝟔
𝒎𝟏 = 𝐄 𝐗𝟏 = 𝟏𝟏 × 𝟏
+ 𝟐 × 𝟏
+ 𝟑 × 𝟏
+ 𝟒 × 𝟏
+ 𝟓 × 𝟏
+ 𝟔 × = 𝟒, 𝟑𝟑
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
𝟐
𝟏 𝟐 𝟐
𝟐 𝟐
𝟑 𝟐
𝟒 𝟐
𝟓 𝟐
𝟔
𝒎𝟐 = 𝐄 𝐗 =𝟏 × + 𝟐 × + 𝟑 × + 𝟒 × + 𝟓 × + 𝟔 × = 𝟐𝟏
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
Central Moments (Moments centrés ) 26
The k th order central moment of the random variable 𝑋 ( k ∊ N) is defined by:
𝒏
𝐤
𝝁𝒌 = 𝐄 (𝐗 − 𝐄(𝐗)) = ( 𝒙𝒊 − 𝐄 𝐗 )𝐤 𝐏(𝐗 = 𝒙𝒊 )
𝒊=𝟏
Example
The 1st order central moment in the rolling die example is given by :
𝟏 𝟏
𝟏 𝟏
𝟐 𝟏
𝟑
𝝁𝟏 = 𝐄 (𝐗 − 𝐄(𝐗)) = 𝟏 − 𝟒, 𝟑𝟑 × + 𝟐 − 𝟒, 𝟑𝟑 × + 𝟑 − 𝟒, 𝟑𝟑 ×
𝟐𝟏 𝟐𝟏 𝟐𝟏
𝟏
𝟒 𝟏
𝟓 𝟏
𝟔
+ 𝟒 − 𝟒, 𝟑𝟑 × + 𝟓 − 𝟒, 𝟑𝟑 × + 𝟔 − 𝟒, 𝟑𝟑 × = 𝟎
𝟐𝟏 𝟐𝟏 𝟐𝟏
𝐄 𝑿 𝐗 − 𝟏 … (𝐗 − 𝐤 + 𝟏) = 𝑿 𝐗 − 𝟏 … (𝐗 − 𝐤 + 𝟏) 𝐏(𝐗 = 𝒙𝒊 )
𝒊=𝟏
Example
𝟏 𝟐 𝟑
𝐄 𝑿 𝐗−𝟏 = 𝟏 𝟏−𝟏 + 𝟐 𝟐−𝟏 + 𝟑 𝟑−𝟏
𝟐𝟏 𝟐𝟏 𝟐𝟏
𝟒 𝟓 𝟔
+ 𝟒 𝟒−𝟏 + 𝟓 𝟓−𝟏 + 𝟔 𝟔−𝟏 = 𝟏𝟔, 𝟔𝟕
𝟐𝟏 𝟐𝟏 𝟐𝟏
28
Remarks :
The expectation and the variance are special moments :
Definition:
Let 𝑋 be a discrete random variable defined on 𝑋(Ω) by the
mass function 𝑃(𝑋=𝑥)
The moment generating function (MGF) of X, denoted 𝑴𝑿 𝒕 ,
is defined for all values of t by:
𝝏𝟐 𝑴𝑿 𝟎
𝐄[𝑿𝟐 ] = 𝐌𝐗′′ 𝟎 =
(𝝏𝒕)𝟐
We notice that for k = 0 :
𝑴𝑿 𝟎 = 𝑬 𝒆𝟎.𝑿 = 𝐄[𝟏] = 𝟏
32
* The moment generating function of the sum of
independent random variables is the product of the
individual moment generating functions
𝟏 𝒕 𝟐 𝟐𝒕 𝟑 𝟑𝒕 𝟒 𝟒𝒕 𝟓 𝟓𝒕 𝟔 𝟔𝒕
𝑴𝑿 𝒕 = 𝒆 + 𝒆 + 𝒆 + 𝒆 + 𝒆 + 𝒆
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
The PGF is used to obtain the probabilities for each value 𝑥i ∈𝑋(Ω)
of a discrete random variable.
Définition
The Probability Generating Function denoted 𝐺𝑋(𝑡), of a discrete
random 𝑋 non negative (𝑥 ∈{0,1,2,…}) is defined by:
𝑮 𝑿 𝒕 = 𝑬 𝒕𝑿 = 𝒕𝒙𝒊 𝑷 𝑿 = 𝒙𝒊
𝒙𝒊 ∈𝑿(𝛀)
Example :
35
In the previous example of the die,
𝑥i 1 2 3 4 5 6 ∑
𝑃(𝑋=𝑥i) 1/21 2/21 3/21 4/21 5/21 6/21 1
𝑮𝑿 𝒕 = 𝑬 𝒕 𝑿 = 𝒕 𝒙𝒊 𝑷 𝑿 = 𝒙 𝒊
𝒙𝒊 ∈𝑿(𝛀)
𝟏 𝟐 𝟐 𝟑 𝟑 𝟒 𝟒 𝟓 𝟓 𝟔 𝟔
𝑮𝑿 𝒕 = 𝒕+ 𝒕 + 𝒕 + 𝒕 + 𝒕 + 𝒕
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
Property1: 36
Particularly: 𝑷 𝑿 = 𝟎 = 𝑮𝑿 𝟎
𝑷 𝑿 = 𝟏 = 𝑮′𝑿 𝟎
𝟏 ′′
𝑷 𝑿=𝟐 = 𝑮
𝟐 𝑿
𝟎
…
Example : 37
Let 𝐺𝑋(𝑡) be the Probability Generating Function defined by:
𝟏 𝟑 𝟏
𝑮𝑿 𝒕 = 𝒕 + 𝒕+𝟏 𝟐
𝟐 𝟖
𝟏 𝟏 𝟏 𝟏
𝑮𝑿 𝒕 = 𝟖
𝒕𝟎 + 𝟒
𝒕 + 𝟖
𝒕𝟐 + 𝟐
𝒕𝟑
On the other hand, we know that:
𝑮𝑿 𝒕 = 𝒕𝟎 𝑷 𝑿 = 𝟎 + 𝒕𝟏 𝑷 𝑿 = 𝟏 + 𝒕𝟐 𝑷 𝑿 = 𝟐 + 𝒕𝟑 𝑷 𝑿 = 𝟑 +…
𝑮𝑿 𝒕 =
𝟏 𝟎
𝒕 +
𝟏
𝒕 +
𝟏 𝟐
𝒕 +
𝟏 𝟑
𝒕 ①
38
𝟖 𝟒 𝟖 𝟐
𝑮𝑿 𝒕 = 𝒕𝟎 𝑷 𝑿 = 𝟎 + 𝒕𝟏 𝑷 𝑿 = 𝟏 + 𝒕𝟐 𝑷 𝑿 = 𝟐 + 𝒕𝟑 𝑷 𝑿 = 𝟑 +… ②
𝟏
𝑷 𝑿=𝟎 =
𝟖
𝟏
𝑷 𝑿=𝟏 =
𝟒
𝟏
𝑷 𝑿=𝟐 =
𝟖
𝟏
𝑷 𝑿=𝟑 =
𝟐
Second methode: 39
(𝒌)
𝐆𝐗 𝟎
By using the property: 𝑷 𝑿 = 𝒌 = 𝒌!
𝟏 𝟎 𝟏 𝟏 𝟐 𝟏 𝟑 𝟏
𝑮𝑿 𝒕 = 𝒕 + 𝒕 + 𝒕 + 𝒕 𝑷 𝑿=𝟎 = 𝑮𝑿 𝟎 =
𝟖 𝟒 𝟖 𝟐 𝟖
𝟏
𝟏 𝟏 𝟑 𝑷 𝑿=𝟏 = 𝑮′𝑿 𝟎 =
𝐆𝐗′ 𝒕 = 𝟒
+𝟒 𝒕 +𝟐 𝒕𝟐 𝟒
𝟏 ′′ 𝟏
𝐆𝐗′′ 𝒕 =
𝟏
+𝟑𝒕 𝑷 𝑿=𝟐 = 𝑮𝑿 𝟎 =
𝟒 𝟐 𝟖
(𝟑) 𝟏 (𝟑) 𝟏
𝐆𝐗 𝒕 =𝟑 𝑷 𝑿=𝟑 = 𝑮𝑿 𝟎 =
𝟔 𝟐
Property 2: 40
Let 𝐺𝑋(𝑡) be the Probability Generating Function of a
discrete random variable X defined on 𝑋(Ω),
𝑮𝑿 𝟏 = 𝑬(𝟏𝑿 ) = 𝑬(𝟏)=1
𝑮′𝑿 𝟏 = 𝑬(𝑿)
𝑮′′
𝑿 𝟏 = 𝑬 (𝑿) 𝑿 − 𝟏
41
Let’s take the second order factorial moment:
𝐄 (𝐗) 𝐗 − 𝟏 = 𝐆𝐗′′ 𝟏
⇒ 𝐄 𝐗𝟐 − 𝐗 = 𝐆𝐗′′ 𝟏
⇒ 𝐄 𝐗 𝟐 − 𝐄 𝐗 = 𝐆𝐗′′ 𝟏
𝐄 𝐗 𝟐 = 𝐆𝐗′′ 𝟏 + 𝐆𝐗′ 𝟏
Example: 42
By using the PGF of the previous example of the die,
𝟏 𝟐 𝟐 𝟑 𝟑 𝟒 𝟒 𝟓 𝟓 𝟔 𝟔
𝑮𝑿 𝒕 = 𝒕+ 𝒕 + 𝒕 + 𝒕 + 𝒕 + 𝒕
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
We can find the factorial moments and deduce the
expectation and the variance as follows:
The first order factorial moment (Expectation):
𝟏 𝟒 𝟗 𝟐 𝟏𝟔 𝟑 𝟐𝟓 𝟒 𝟑𝟔 𝟓
𝐆𝐗′ 𝒕 = + 𝒕+ 𝒕 + 𝒕 + 𝒕 + 𝒕
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
𝟏 𝟒 𝟗 𝟏𝟔 𝟐𝟓 𝟑𝟔
⇒ 𝐄(𝐗) = 𝐆𝐗′ 𝟏 = + + + + + = 𝟒, 𝟑𝟑
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
The second order factorial moment : 43
𝟒 𝟏𝟖 𝟒𝟖 𝟐 𝟏𝟎𝟎 𝟑 𝟏𝟖𝟎 𝟒
𝐆𝐗′′ 𝒕 = + 𝒕+ 𝒕 + 𝒕 + 𝒕
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
𝟒 𝟏𝟖 𝟒𝟖 𝟏𝟎𝟎 𝟏𝟖𝟎
𝐄 𝐗 −𝐗 𝟐
= 𝐆𝐗′′ 𝟏 = + + + + = 𝟏𝟔, 𝟔𝟕
𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏 𝟐𝟏
Property 3:
Definition :
Let 𝑋 be a discrete random variable on (Ω,𝒜,𝑃) and let
𝑌=g(𝑋) be a transformation of 𝑋.
𝑷(𝒀 = 𝒚𝒋 ) = 𝑷(𝑿 = 𝒙𝒊 )
𝒙𝒊 ∈𝑿 𝛀 :𝐠 𝒙𝒊 =𝒚𝒋
46
xi -3 -2 -1 0 1 2 3
P (X= xi) 0,1 0,15 0,2 0,25 0,05 0,1 0,15
For 𝑋= - 3 or 𝑋= 3 ; 𝑌= 11
For 𝑋= - 2 or 𝑋= 2 ; 𝑌= 6
For 𝑋= -1 or 𝑋= 1 ; 𝑌= 3
For 𝑋= 0 ; 𝑌= 2
Markov’s inequality
Theorem:
Let 𝑋 be a non negative random variable defined
on the probability space (Ω,𝒜,𝑃), then:
𝐄(𝐗)
∀𝐚 > 𝟎, 𝐏(𝐗 ≥ 𝐚) ≤
𝒂
50
Chebyshev’s inequality
Theorem:
Let 𝑋 be a non-negative random variable defined
on the probability space (Ω,𝒜,𝑃), then:
𝑽(𝑿)
∀𝐚 > 0, 𝐏( 𝐗 − 𝐄(𝐗) ≥ 𝐚) ≤
𝒂2
Proof 51
E[Y]
P(Y ≥ b) ≤ 𝑏
2
2 2
E[ X − E(X) ]
P( X − E(X) ≥ 𝑎 )≤
𝑎2
Then:
𝑉(𝑋)
P( X − E(X) ≥ 𝑎) ≤ 𝑎2
Example 52
Let X be a random variable the associates the number of
customers who come to a store in a day, with E(X) = 50 and
V(X)=36.
𝟓𝟎
𝐏(𝐗 ≥ 𝟖𝟎) ≤ Then : 𝐏(𝐗 ≥ 𝟖𝟎) ≤ 𝟎, 𝟔𝟐𝟓
𝟖𝟎
𝟑𝟔
𝐏( 𝐗 − 𝟓𝟎 ≥ 𝟑𝟎) ≤ Then:
𝟑𝟎𝟐
𝐏 𝐗 − 𝟓𝟎 ≥ 𝟑𝟎 ≤ 𝟎, 𝟎𝟒