2A2. Review of Probability
2A2. Review of Probability
FUNDAMENTALS OF PROBABILITY
A random variable (Y) is a variable whose value (y) is determined by the outcome of a
chance experiment. A discrete random variable takes on only a finite, or countable,
number of values. A continuous random variable can take any value in some interval of
values.
≤ 𝑏) for continuous Y.
𝑎 𝑌
fY(y) fY(y)
P(a ≤ Y ≤ 𝑏)
Y Y
The cumulative density function (cdf) of the random variable Y, F(y), gives the
probability that Y is less than or equal to a specific value x, that is, FY(y) = P(Y ≤ y ).
1
EXPECTED VALUE / MEAN [measure of central tendency] {first moment}
The expected value of a random variable is the average value that occurs in many
repeated trials of an experiment.
𝐸(𝑌) = 𝜇𝑌 = ∑𝑦 𝑦𝑓𝑌(𝑦)
This is a weighted average of all possible values of Y, with weights being the
probabilities that the values occurs.
�
The larger the variance of a random variable, the more “spread out” its values are.
fY(y)
Y
If Y is some random variable from a population, it is called the population variance.
The square root of the variance is the standard deviation, sd(Y) = Y.
Properties of variance
(a) Var(a) = 0, for any constant a
(b) Var(aY) = a2Var(Y), for any constant a
(c) Var(aY + b) = a2Var(Y), for any constants a, b
2
variance 𝜎2, a standardized random variable with mean 0 and variance 1 can be
▶ Standardizing a random variable: Given a random variable Y with mean and
defined as:
𝑍 = [𝑌 − 𝐸(𝑌)]/√𝑉𝑎𝑟(𝑌), with
𝐸(𝑍) 𝐸(𝑌)
= 0 and = �𝑌 =
2
𝜇 𝑉𝑎𝑟(
= − = − 𝑌)
𝜇𝑌 𝜇𝑌
𝑉𝑎𝑟(𝑍) = 1
𝑌
𝜎𝑌 𝜎 𝜎 𝜎
2
�𝑌 𝑌
�
𝑌 𝑌 𝑌
In econometrics, we consider the probability that 𝑢𝑖 and 𝑢𝑗 occurs at the same time.
3
of a certain factor Xk.
4
COVARIANCE [measure of linear association between 2 random variables]
𝐶𝑜𝑣( 𝑋, 𝑌) = 𝜎𝑋𝑌 = 𝐸[ 𝑋 − 𝐸( 𝑋)][𝑌 − 𝐸( 𝑌)] = 𝐸( 𝑋𝑌) − 𝐸(X) 𝐸( 𝑌)
This is positive (negative) if the two random variables move in the same direction
(opposite directions).
If X and Y are some random variables from a population, it is called the population
covariance.
Properties of covariance
(a) Cov(a, d) = 0, for any constants a, d
(b) Cov(aX , cY) = acCov(X,Y), for any constants a, c
(c) Cov(aX + b, cY + d) = acCov(X, Y), for any constants a, b, c, d
(d) Cov(X, Y) = 0 if X and Y are independent (but the reverse does not necessarily hold)
𝐶𝑜𝑣( 𝑋, 𝑌) 𝜎𝑋𝑌
eliminates the units of measurement, and defines the correlation between X and Y.
5
More properties of expected value and variance
1. E(aX + bY + c) = aE(X) + b E(Y) + c
1’. If {𝑎1, 𝑎2 , …, 𝑎𝑁} are constants and {𝑌1, 𝑌2 , …, 𝑌𝑁}
𝐸(∑𝑁 𝑎𝑖𝑌𝑖) = 𝐸(𝑎1𝑌1 + 𝑎2𝑌2 + ⋯ + 𝑎𝑁𝑌𝑁) = 𝑎1𝐸(𝑌1) + 𝑎2𝐸(𝑌2) +
are random variables,
⋯+
𝑖
then
CONDITIONAL EXPECTATION
𝐸(𝑌|𝑋 = 𝑥) = 𝜇𝑌|𝑋 = ∑ 𝑦𝑓𝑌(𝑦|𝑥)
𝑦
6
Standard distribution
NORMAL DISTRIBUTION
fY(y)
fY(Y)
7
t DISTRIBUTION
⬚
Let 𝑍~𝑁(0,1) X~ be two independent random variables.
N
and 2
𝑋2 / 𝑁1 ,
𝑁2 𝑁2
2𝑁2(𝑁1+𝑁2−2)
𝐸(𝐹) ==
Var(𝐹) 𝑁2/(𝑁2 − 2), 2
𝑁1(𝑁2−2)2(𝑁2−4)
𝑡2 = 𝐹1,𝑁
�
fY(Y)
For large df, the chi-square, t and F distributions converge to the normal distribution.