Mfin6201 Week1
Mfin6201 Week1
Mfin6201 Week1
An Overview of Econometrics
Raphael Park
May 30, 2021
1
Who’s talking?
2
Course Admins
3
Course Schedule
4
Assessment
The course outline is bit outdated due to the virus situation, please refer to
slides and materials on Moodle.
5
Overview of today’s lecture
6
What is Econometrics?
• Statistics + economics
• Standard assumptions in statistics
• Nature of economic data
• Financial econometrics = econometrics + finance
• Use econometrics techniques to study a variety of problems
from finance
• Focus on hypothesis testing, causal inference and forecasting
7
Types of data in finance
8
Brief overview of the first part of the course
10
How to carry out empirical analysis in finance?
11
An empirical example: class size and education output
12
Using data for empirical analysis
13
Initial look at the data
14
Initial look at the data
15
Eyeball econometrics is not rigorous enough!
We need to get some numerical evidence on whether districts with low STRs
have higher test scores but how?
• Compare average test scores in districts with low STRs to those with high
STRs (“estimation”).
Point estimation involves the use of sample data to calculate a single
value (known as a statistic), which is served as the “best estimate” of an
unknown (fixed or random) population parameter.
• Test the null hypothesis that the mean test scores in the two types of
districts are the same, against the “alternative” hypothesis that they
differ (“hypothesis testing”).
Testing a hypothesis on the basis of observing a process that is modeled
via a set of random variables.
• Estimate an interval for the difference in the mean test scores, high v.
low STR districts (“confidence interval”). A confidence interval is a
type of interval estimate of a population parameter.
16
What is next?
• Or, you can use (linear) regression to measure the relation (i.e., the
slope) between student-to-teacher ratios and average test scores.
• Why linear? - it is easy to implement, easy to interpret and it gives the
best linear approximation even the relationship is not linear.
• Why regression? - it went through a robust mathematical procedure to
get the estimate (hard to beat with alternatives like eyeballing)
• Before turning to regression, however, we will review some of the
underlying theory of estimation, hypothesis testing, and confidence
intervals:
• Why do these procedures work, and why use these rather than
others?
• We will review the intellectual foundations of statistics and
econometrics
17
Review of Statistical Theory
18
Probability framework for statistical inference
19
Population, random variable, and distribution
Population
• The group or collection of ALL possible entities of interest
(school districts)
• We will think of populations as infinitely large
• e.g. Population Average of height of human being...? (only
God knows)
Random variable
• Random variable is a variable whose value is subject to
variations due to chance
• As a result, it can take on a set of possible different values
each with an associated probability
• Numerical summary of a random outcome (district average
test score, district STR)
20
Discrete random variable
Example
1 with probability 0.2
2 with probability 0.3
X =
3 with probability 0.3
4 with probability 0.2
21
Discrete random variable: pdf
f (x) ≡ P(X = x)
22
Discrete random variable: cdf
f (x) ≡ P(X = x)
23
Discrete random variable: cdf
24
Relation between pdf and cdf
F (x) = f (x) + f (x − ∆) + · · ·
= f (x) + F (x − ∆)
f (x) = F (x) − F (x − ∆)
25
Continuous random variable
For example,
• Normal distribution (bell-shaped curve)
X ∼ N(µ, σ 2 )
: X is drawn from a normal distribution with mean µ and variance
σ2
• pdf of the normal distribution
26
Continuous random variable
X = P(X ≤ x)
27
pdf and cdf of continuous distribution
28
pdf and cdf
cdf
• Discrete: F (x) =
P
t≤x f (t)
Rx
• Continuous: F (x) = −∞ f (t)
29
Probability density of stock returns
30
Moments: mean
• Example
1 with probability 0.2
2 with probability 0.3
X =
3 with probability 0.3
4 with probability 0.2
31
Moments: variance
• Standard deviation
p
std(X ) = Var [X ]
• Example
• Expectation
E [a + bX ] = a + bE [X ]
E [X + Y ] = E [X ] + E [Y ]
Combined,
E [aX + bY ] = aE [X ] + bE [Y ]
• Variance
Var [a + bX ] = b 2 Var [X ]
Var [X + Y ] = Var [X ] + Var [Y ] + 2Cov [X , Y ]
Combined,
Var [aX + bY ] = a2 Var [X ] + b 2 Var [Y ] + 2abCov [X , Y ] 33
Moments: skewness, kurtosis
E [(X −µ)3 ]
• Skewness (3rd moment) = σ3
• measure of asymmetry of a distribution
• skewness = 0: distribution is symmetric
• skewness > (<) 0: distribution has long right (left) tail
E [(X −µ)4 ]
• Kurtosis (4th moment) = σ4
• measure of mass in tails
• measure of probability of large values
• kurtosis = 3: normal distribution
• kurtosis > 3: heavy tails (leptokurtotic)
34
Skewness and Kurtosis
35
Covariance
36
Correlation
• Correlation is defined as
Cov (X , Y ) σxy
Corr (X , Y ) = p = = ρxy
Var (X )Var (Y ) σx σy
• −1 ≤ corr (X , Y ) ≤ 1
• Cov(X,Y) > 0 means a positive relation between X and Y
• Corr(X,Y) = 1 mean perfect positive linear association
• Corr(X,Y) = -1 means perfect negative linear association
• Corr(X,Y) = 0 means no linear association
37
Correlation examples
• Mean
E [X ] = p · 1 + (1 − p) · 0 = p
• Variance
Y/X 1 0
1 p2 p(1 − p)
0 p(1 − p) (1 − p)2
• Covariance
39
Correlation examples
• Correlation
Cov (X , Y )
Corr (X , Y ) = p =0
Var (X )Var (Y )
40
Correlation examples
Y/X 1 0
1 p 0
0 0 1-p
• Covariance
• Correlation
Cov (X , Y ) p(1 − p)
Corr (X , Y ) = p = =1
Var (X )Var (Y ) p(1 − p)
42
Correlation
43
Correlation
• Corr(X, Y) ∈ (0, 1)
: When X is high, Y is likely, but not perfectly, to be high
• Corr(X, Y) ∈ (-1, 0)
: When X is high, Y is likely, but not perfectly, to be low
• Note: correlation does not imply causality
• Example: More education is related to more salary.
• But does it mean getting more education will definitely
increase the salary? No! Maybe you earn more salary then you
have the money to get education, or higher education means
better family background..!
44
Bayes’ theorem
45
Conditional probability: example
2
P(first is a son | at least one is a son) =
3
P(first is a son & at least one is a son)
=
P(at least one is a son)
1/2
=
3/4
2
=
3
46
Textbook example: Table 2.2
• Joint distribution
47
Textbook example: Table 2.2
• Marginal distribution
n
X
P(Y = y ) = P(X = xi , Y = y )
i=1
48
Textbook example: Table 2.2
• Conditional distribution
P(X = x and Y = yB)
P(Y = y |X = x) =
P(X = x)
0.15
P(long commute|rain) = = 0.50
0.30
0.15
P(rain|long commute) = = 0.68
0.22
0.63
P(short commute|no rain) = = 0.90
0.70
49
Conditional probability example: AIDS testing
• Question
The probability that a patient has HIV is 0.001 and the
diagnostic test for HIV can detect the virus with a probability
of 0.98. Given that the chance of a false positive is 6%, what
is the probability that a patient who has already tested
positive really has HIV?
50
Conditional probability example: AIDS testing
P(HIV ) = 0.001
P(positive | HIV ) = 0.98
P(positive | not HIV ) = 0.06
51
Conditional probability example: AIDS testing
52
Conditional probability example: AIDS testing
• Bayes’ theorem
P(positive|HIV ) · P(HIV )
P(HIV | positive) =
P(positive)
0.98 × 0.001
=
0.06092
= 0.0161
53
Conditional means
E (Y |X = x)
54
Conditional means
56