Chapter05 Moments 2023su
Chapter05 Moments 2023su
Department of Economics
National Taiwan University
2023.7.21.
Section 1
Moments
Moments
Expectation
Definition (Expectation)
Let S = supp(X). The expectation of a random variable X is defined
as
⎧
⎪∑x∈S x f (x)
⎪
⎪ discrete
E(X) = ⎨
⎪
⎪
⎪ x f (x)dx continuous
⎩∫x∈S
E(Y) = 0.9
That is, “on average, the investor expects to earn 0.9.” ⇐ What
does this mean?
(NTU Econ) Basic Quantitative Method 2023.7.21. 5 / 51
Moments
Simulation Results
If you invest in this stock N years. Let Yi denote the price change
for year i, and the average earning is thus ∑ i=1
N
Yi
N
∑N
i=1 Yi
We can see that N is very close to E(Y) = 0.9 when N is
large.
1.5
1.0
Average Earning
0.5
0.0
−0.5
Expectation
Example:
⎧
⎪
⎪ 2 with P(X = 2) = 1/3
⎪
⎪
X = ⎨ 1 with P(X = 1) = 1/3
⎪
⎪
⎩ −1 with P(X = −1) = 1/3
⎪
⎪
Definition (Variance/SD)
The variance of a discrete random variable X is defined by
⎧
⎪∑x (x − E(X))2 f (x)
⎪
⎪ 2
Var(X) = E [(X−E(X)) ] = ⎨
⎪
⎪
⎪ (x − E(X))2 f (x)dx
⎩∫ x
E(c) = c,
and
Var(c) = 0.
Therefore,
E(E(X)) = E(X)
E(aX + b) = aE(X) + b
Var(aX + b) = a 2 Var(X)
E(X − E(X)) = 0
Var(X) = E(X 2 ) − [E(X)]2
Definition
The fair price of a stock is defined by a price such that the expected
return equals the risk free rate.
More on Expectation
E(g(X)) ≠ g(E(X))
Jensen’s Inequality
Theorem
If X is a random variable and g(⋅) is a convex function, then
E(g(X)) ≥ g(E(X))
Proof.
X−µ
Z=
σ
Then Z ∼ (0, 1) is called a standardized random variable.
Moments
k-th Moments
E(X k )
k-th Central Moments
E[(X − E(X))k ]
Skewness
3
3rd standardized moment (Skewness): γ3 = E [( √X−E(X)
Var(X)
)]
Kurtosis
4
4th standardized moment (Kurtosis): γ4 = E [( √X−E(X)
Var(X)
)]
Section 2
Definition (MGF)
Let X be a discrete (continuous) random variable, and the pmf (pdf)
is f (x). Given h > 0 and for all −h < t < h, if the following function
M X (t) = E(e tX )
Properties
Proof: expand e tX as
Properties
Theorem (Uniqueness)
For all t ∈ (−h, h), if M X (t) = MY (t), then X and Y has exactly the
same distribution, FX (c) = FY (c) for all c ∈ R.
Proof: Beyond the scope of this course (via so-called the inverse
Fourier transform).
Properties
MY (t) = e bt M X (at).
Proof: By definition.
Section 3
Definition
Let X and Y be discrete (continuous) random variables with joint
pmf (pdf) f XY (x, y). Let g(X, Y) be a function of these two random
variables, then:
Covariance
Definition (Covariance)
Properties
Theorem
Given constants a, b, c, and d
E(aX + bY) = aE(X) + bE(Y)
Cov(X, X) = Var(X)
Cov(X, c) = 0
Correlation Coefficient
Cov(X, Y)
ρ XY = Corr(X, Y) = √ √
Var(X) Var(Y)
Correlation Coefficient
Theorem
The correlation coefficient lies between 1 and -1:
−1 ≤ ρ XY ≤ 1
Correlation Coefficient
Note:
ρ XY = 1 (perfect correlation)
ρ XY = −1 (perfect negative correlation)
ρ XY = 0 (zero correlation, no correlation, uncorrelated)
However, no correlation does not mean that there is no
relationship between X and Y
It just suggests that there is no linear relationship between X and
Y
Section 4
Theorem
Let X and Y are independent variables. Then
E[g(X)h(Y)] = E[g(X)]E[h(Y)]
Theorem
X and Y are independent random variables. Their MGFs are M X (t)
and MY (t), respectively. Let Z = X + Y, then
Theorem
Given that X and Y are independent:
E(XY) = E(X)E(Y)
Cov(X, Y) = 0
Example
Section 5
Conditional Expectation
Definition
The conditional expectation of Y given X = x is
E(Y∣X) = g(X)
Conditional Variance
Definition
The conditional variance of Y given X = x is
Var(Y∣X) = h(X)
Important Theorems
Theorem
Useful Rule
E[h(X)Y∣X] = h(X)E[Y∣X]
Simple Law of Iterated Expectation
E(E[Y∣X]) = E(Y)
E(E[XY∣X]) = E(XY)
Application:
Important Theorems
E(Y∣X) = α + βX
α = E(Y) − βE(X)
є ≡ Y − E(Y∣X) = Y − (α + βX)
Hence,
Y = α + βX + є
Section 6
E[g(X1 , X2 , . . . , X n )]
= ∑ ⋯ ∑ g(x1 , x2 , . . . , x n ) f X (x1 , x2 , . . . , x n )
x1 xn
E[g(X1 , X2 , . . . , X n )]
= ∫ ⋯ ∫ g(x1 , . . . , x n ) f X (x1 , . . . , x n )dx n ⋯dx1
x1 xn
Properties
n n
E (∑ X i ) = ∑ E(X i )
i=1 i=1
n n n i−1
Var (∑ X i ) = ∑ Var(X i ) + 2 ∑ ∑ Cov(X i , X j )
i=1 i=1 i=1 j=1
Theorem
Let X1 , X2 , . . . , X n are independent variables. Then
Theorem
X1 , X2 ,. . . , X n are independent with MGF: M X1 (t), M X2 (t),. . .,
M X n (t). Let Y = ∑ni=1 X i , then
n
MY (t) = M X1 (t)M X2 (t)⋯M X n (t) = ∏ M X i (t)
i=1
Theorem
Let {X i }ni=1 are i.i.d. random variables with E(X i ) = µ and
Var(X i ) = σ 2 , and let
n
Y = ∑ Xi .
i=1
Then
E(Y) = nµ,
Var(Y) = nσ 2 .
Example
Let
{X i }ni=1 ∼i.i.d. Bernoulli(p)
That is,
E(X i ) = p and Var(X i ) = p(1 − p)
Let n
Y = ∑ Xi
i=1