Pertemuan 3
Pertemuan 3
• The simple regression model can be used to study the relationship between two variables.
• The simple regression model has limitations as a general tool for empirical analysis.
• it is sometimes appropriate as an empirical tool.
• Learning how to interpret the simple regression model is good practice for studying multiple
regression, which we will do in subsequent chapters.
•Much of applied econometric analysis begins with the following premise: 𝑦 and 𝑥
are two variables, representing some population, and we are interested in “explaining
𝑦 in terms of 𝑥,” or in “studying how 𝑦 varies with changes in 𝑥.”
𝑬(𝒚|𝒙) = 𝜷𝟎 + 𝜷𝟏 𝒙
• We call this Population Regression Function , and is shown as a linear function (of x)
23/09/20 Source: Ekki Syamsulhakim, Gujarati & Wooldridge
Population Regression Function
error
error
•We generally don’t have population data; using sample data to obtain the parameter
• Statistical inference
•Note:
• We denote population parameter WITHOUT HAT (without ^)
• We denote sample parameter with HAT (^)
• Missing hat may lead to confusion (and also reduce marks)
23/09/20 Source: Ekki Syamsulhakim, Gujarati & Wooldridge
Important: Symbols
• Econometric Model
𝑦 = 𝛽! + 𝛽" 𝑥 + 𝑢
• PRF (𝐸(𝑢|𝑥) = 0)
𝐸(𝑦|𝑥) = 𝛽! + 𝛽" 𝑥
• SRF (by “OLS”)
𝑦0 = 𝛽1! + 𝛽1" 𝑥
• We want to minimize the residual (the distance between actual data and estimated value – of
our independent variable)
• Minimize 𝑢%" è minimize (𝑦" − 𝑦)
(
• Because we have 𝑛 observation, we will have as many as 𝑛 of 𝑢%" , i.e {,
𝑢! … 𝑢,# }
• A better idea is to take the absolute value of 𝑢%" , i.e |𝑢%" | , so the representative number
#
is ∑"$! |𝑢%" |
• ..but this is also a problem if we have more than one SRF that has the same ∑#"$! |𝑢 1" |
• To overcome the problem we need to square 𝑢%" , so that the representative number will be
# % %
∑"$! 𝑢%" , or simply ∑ 𝑢%" , named “sum of squares of residual” or “residual sum of
squares”
23/09/20 Source: Ekki Syamsulhakim, Gujarati & Wooldridge
Least Squared Method ^
SRF2:Y2= a1+a2X
Y
^1
SRF1:Y = b1+b2X
1 -2
1
1 -1/2
2 -11/2
0
-1
-1
X
SRF1: S|u| = |1| + |-1| + |-1| + |1| + |-1.5| = 5.5
Su2 =12 + 12 + 12 + 12 + 1.52 = 6.25 smaller
SRF2: S|u| = |2| + |0| + |-1/2| + |1| + |-2| = 5.5
Su2 = 22 + 02 + (-1/2)2 + 12+ (-2)2 = 9.25
23/09/20 Source: Ekki Syamsulhakim, Gujarati & Wooldridge
Ordinary Least Squares (OLS) Method
Yi = b1 + b2Xi + ui
u i = Y i - b1 - b2X i
¶f(.)
= - 2 S (Y i - b1 - b2Xi )
¶ b1
¶f(.)
= - 2 S Xi (Yi - b1 - b2Xi )
¶ b2
Set each of these two derivatives equal to zero and
solve these two equations for the two unknowns: b1 b2
23/09/20 Source: Ekki Syamsulhakim, Gujarati & Wooldridge
To minimize f(.), you set the two
derivatives equal to zero to get:
¶f(.)
= - 2 S (Y i – b1 – b2Xi ) = 0
¶ b1
¶f(.)
= - 2 S xi (Yi - b1 – b2Xi ) = 0
¶ b2
When these two terms are set to zero,
b1 and b2 become b1 and b2 because they no longer
represent just any value of b1 and b2 but the special
values that correspond to the minimum of f(.) .
23/09/20 Source: Ekki Syamsulhakim, Gujarati & Wooldridge
- 2 S (Y i - b1 – b2Xi ) = 0
-2S Xi (Y i – b1 – b2Xi ) = 0
S Yi - nb1 – b2 SXi = 0
2
S Xi Yi - b1 S X i - b2 S Xi = 0
nb1 + b2 S Xi = S Yi
2
b1 S Xi + b2S Xi = S Xi Yi
23/09/20 Source: Ekki Syamsulhakim, Gujarati & Wooldridge
n SXi b1
= S Yi
S i S i
X X 2
b2 = S Xi Yi
Solve the two unknowns
n SXi Yi - S Xi SYi
b2 =
n SX i - (SXi )
2 2
b1 = Y - b2 x
23/09/20 Source: Ekki Syamsulhakim, Gujarati & Wooldridge
Assumptions of Simple Regression
1. The linear regression Model:linear in
parameters
Y = b1+ b2X+ u
2. X values are fixed in repeated sampling, so that X is
not constant (X is nonstochastic).
3. Zero mean value of error terms (disturbance, ui),
E( ui | xi) = 0
4. Homoscedasticity or equal variance of ui, the
conditional variances of u i are identical, i.e.,
var(ui|xi) = s2
23/09/20 Source: Ekki Syamsulhakim, Gujarati & Wooldridge