qrm2 Session1 2
qrm2 Session1 2
qrm2 Session1 2
METHODS II
VERA E. TROEGER
DEPARTMENT OF ECONOMICS
E-MAIL: [email protected]
OFFICE HOURS: BY APPOINTMENT
COURSE OUTLINE
1. Introduction to the Course: the OLS model, Gauss-Markov
Assumptions and Violations
2. Heteroskedasticity, cross-sectional correlation,
multicollinearity, omitted variable bias: tests and common
solutions.
3. Specification issues in Linear Models: Non-Linearities and
Interaction Effects
4. Dynamics, serial correlation and dependence over time
5. Pooling Data 1: Heterogeneity: How to choose the right
model – how good are the available tests.
COURSE OUTLINE CONT.
6. Pooling Data 2: Slope and Parameter Heterogeneity:
Seemingly Unrelated Regression and Random Coefficient
Models, modelling parameter instability
7. Specification Issues: Endogeneity and spatial
econometrics: instrumental variable approaches and
simultaneous equation models
8. Limited Dependent Variable Models: Binary and Ordered
Choice Models
9. Limited Dependent Variable Models with Pooled and Panel
Data
10. Specification Issues in Limited Dependent Variable
Models: Dynamics and Heterogeneity
11. Wrap Up, Q&A
THE SIMPLE LINEAR REGRESSION
MODEL
Regression models help investigating bivariate and multivariate
relationships between variables, where we can hypothesize that 1 variable
depends on another variable or a combination of other variables.
-2 0 2 4 6
x
yi x i i
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
y7
epsilon7
yhat7
y1
y
alpha
yhat1
0 1 2 3 4 5 6 7 8 9 10
0 2 4 6 8 10 x
x
y Fitted values
ŷi ˆ ˆ x i ˆ i ˆ i yi ˆ ˆ x i
Minimize the sum of all squared deviations from the line (squared
residuals)
This is done mathematically by the statistical program at hand
the values of the dependent variable (values on the line) are called
predicted values of the regression (yhat): 4.97,6.03,7.10,8.16,9.22,
10.28,11.34,12.41,13.47,14.53 – these are very close to the „true
values“; the estimated alpha = 4.97 and beta = 1.06
OLS REGRESSION
Ordinary least squares regression: minimizes the squared
residuals
yi x i i i yi x i
ŷi ˆ ˆ x i ˆ i
n n
i i i min
(Y
i 1
Yˆ ) 2
(
ˆ ) 2
i 1
15
Components:
10
Component plus residual
DY: y; at least 1 IV: x
5
Constant or 0
Regression coefficient,
slope: beta
-10
-2 0 2 4 6
Error term, residuals: epsilon x
DERIVATION
Linear relationship between x (explanatory variable) and y (dependent
variable)
yi x i i
n n
(Y Yˆ ) (ˆ )
i 1
i i
2
i 1
i
2
min
n
ˆ yi ˆ x i ˆ y ˆ x
i 1
y
n
i y
ˆ i 1
n
x
i 1
i x
y
n
y xi x
i
Cov x, y
ˆ i 1
X ' X 1X ' y
n
V ar x
xi x
2
i 1
Naturally we still have to verify whether ̂ and ̂
really minimize the sum of squared residuals and
satisfy the second order conditions of the
minimizing problem. Thus we need the second
derivatives of the two functions with respect to
alpha and beta which are given by the so called
Hessian matrix (matrix of second derivatives). (I
spare the mathematical derivation)
(x i x)(yi y)
ˆ yx i 1
n
(x
i 1
i x) 2
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
beta=1.06
alpha
0 1 2 3 4 5 6 7 8 9 10
x
y Fitted values
Properties of the OLS estimator:
Cov i , j E i j 0, i j
If all Gauss-Markov assumptions are met than the OLS estimators alpha
and beta are BLUE – best linear unbiased estimators:
best: variance of the OLS estimator is minimal, smaller than the
variance of any other estimator
linear: if the relationship is not linear – OLS is not applicable.
unbiased: the expected values of the estimated beta and alpha
equal the true values describing the relationship between x and y.
INFERENCE
Is it possible to generalize the regression results for the sample under
observation to the universe of cases (the population)?
Can you draw conclusions for individuals, countries, time-points beyond
those observations in your data-set?
Significance tests are designed to answer exactly these questions.
If a coefficient is significant (p-value<0.10, 0.05, 0.01) then you can
draw conclusions for observations beyond the sample under
observation.
But…
Only in case the samples matches the characteristics of the population
This is normally the case if all (Gauss-Markov) assumptions of OLS
regressions are met by the data under observation.
If this is not the case the standard errors of the coefficients might be
biased and therefore the result of the significance test might be wrong
as well leading to false conclusions.
SIGNIFICANCE TEST: THE T-TEST
2
2
,
N *Var X
The t-test:
T-test for significance: testing the H0 (Null-Hypothesis) that beta
equals zero: H0: beta=0; HA: beta≠0
The test statistic follows a student t distribution under the Null
ˆ r ˆ r
t n 2
SE ˆ SSR
N *Var X
ˆ ˆ
t n 2
SE ˆ SSR
N *Var X
One can reject the null-hypothesis even though it is true (type 1 error,
alpha error):
alpha=Pr[Type I Error]
=Pr[rejecting the H0 | H0 is true]
Or not reject the null-hypothesis even though it is wrong (type 2 error, beta
error)
beta=Pr[Type II Error]
=Pr[accepting the H0 | Ha is true]
CONFIDENCE INTERVALS
Significance tests assume that hypotheses come before
the test: beta≠0. however, the significance test leaves us
with some vacuum since we know that beta is different from
zero but since we have a probabilistic theory we are not
sure what the exact value should be.
i 1
Explained
n
(Estimation) Sum of Squares (SSE):
SSE yˆi y
2
i 1
i 1 i 1
GOODNESS OF FIT
How well does the explanatory variable explain the dependent
variable?
How well does the regression line fit the data?
(Yˆ Yˆ ) 2
SSE SSR
R
2 i 1
n
1
SST SST
(Y Y ) 2
i 1
Properties of R²:
E ˆ 0
If an estimator is unbiased, then its probability distribution has an expected
value equal to the parameter it is supposed to be estimating. Unbiasedness
does not mean that the estimate we get with any particular sample is equal to
the true parameter or even close. Rather the mean of all estimates from
infinitely drawn random samples equals the true parameter.
3.5
3
2.52
Density
1.5 1
.5
0
.5 .6 .7 .8 .9 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 2 2.1
b1
SAMPLING VARIANCE OF AN
ESTIMATOR
Efficiency: is a relative measure between two estimators – measures
the sampling variance of an estimator: V(beta)
Let ̂ and be two unbiased estimator of the true parameter . With
variances V ˆ and V . Then ̂ is called to be relative more efficient
than .
than if is smaller V
V ˆ
The property of relative efficiency only helps us to rank two unbiased
estimators.
.08
.4
.06
.3
Density
Density
.04
.2
.02
.1
0
-20 - 10 0 10 20 -4 -2 0 2 4
b1 b2
TRADE-OFF BETWEEN BIAS AND
EFFICIENCY
With real world data and the related problems we sometimes have only
the choice between a biased but efficient and an unbiased but
inefficient estimator. Then another criterion can be used to choose
between the two estimators, the root mean squared error (RMSE). The
RMSE is a combination of bias and efficiency and gives us a measure
of overall performance of an estimator.
RMSE:
1 K
RMSE
2.5
2
ˆ ˆ true
K k 1
2
MSE E true
2
xtfevd ˆ
1.5
MSE Var Bias , true
Density
2
ˆ ˆ
1
.5
fixed effects
k measures the number
0
-.5 0 .5 1 1.5 2 2.5 3 3.5 4 4.5 5 5.5 6 6.5 7 7.5 of experiments, trials or
coefficient of z3 simulations
ASYMPTOTIC PROPERTIES OF
ESTIMATORS
We can rule out certain silly estimators by studying the asymptotic or large
sample properties of estimators.
We can say something about estimators that are biased and whose
variances are not easily found.
Asymptotic analysis involves approximating the features of the sampling
distribution of an estimator.
Consistency: how far is the estimator likely to be from the true parameter as
we let the sample size increase indefinitely.
If N→∞ the estimated beta equals the true beta:
lim Pr ˆ n 0, p lim ˆ n ,
n
lim E ˆ n
n
yi 1x i1 2 x i2 ... k x ik i
n n n n
x x2 x x1 yi y xi1 x1 xi 2 x2 xi 2 x2 yi y
2
i2 i1
ˆ1 i 1 i 1 i 1 i 1
2
n n
n
x x1 x x2 xi1 x1 xi 2 x2
2 2
i1 i2
i 1 i 1 i 1
ˆ X ' X 1 X ' y
n
SST1 xi1 x1
2
i 1
n
xˆi1 x1
2
SSE
R12 for the regression of xi1 on xi 2 : R12 i 1
n
xi1 x1
SST 2
i 1
SD ˆ1 SE ˆ1
SST1 1 R12
F – TEST: TESTING MULTIPLE
LINEAR RESTRICTIONS
t-test (as significance test) is associated with any OLS coefficient.
We also want to test multiple hypotheses about the underlying parameters
beta_0…beta_k.
F
SSR SSR / q r ur
SSRur / n k 1
F Fq , n k 1
The F-test, tests multiple restriction: e.g. all coefficients jointly equal zero:
H0: beta_0=beta_1=…=beta_k=0
Ha: H0 is not true, thus at least one beta differs from zero
The F-statistic (or F-ratio) is defined as:
The F-statistic is F distributed under the
Null-Hypothesis.
F-test for overall significance of a regression: H0: all coefficients are jointly
zero – in this case we can also compute the F-statistic by using the R² of the
Regression:
R² / k
1 R ² / n k 1
SSR_r: Sum of Squared Residuals of the restricted model (constant
only)
SSR_ur: Sum of Squared Residuals of the unrestricted model (all
regressors)
SSR_r can be never smaller than SSR_ur F is always non-negative
k – number of explanatory variables (regressors), n – number of
observations, q – number of exclusion restrictions (q of the variables
have zero coefficients): q = df_r – df_ur (difference in degrees of
freedom between the restricted and unrestricted models; df_r > df_ur)
The F-test is a one sided test, since the F-statistic is always non-
negative
We reject the Null at a given significance level if F>F_critical for this
significance level.
If H0 is rejected than we say that all explanatory variables are jointly
statistically significant at the chosen significance level.
THUS: The F-test only allows to not reject H0 if all t-tests for all single
variables are insignificant too.
Goodness of Fit in multiple Regressions:
As with simple binary regressions we can define SST, SSE and SSR. And
we can calculate the R² in the same way.
BUT: R² never decreases but tends to increase with the number of
explanatory variables.
THUS, R² is a poor tool for deciding whether one variable of several
variables should be added to a model.
We want to know whether a variable has a nonzero partial effect on y in
the population.
Adjusted R²: takes the number of explanatory variables into account since
the R² increases with the number of regressors:
n 1
R 2
adj 1
nk
1 R 2
k is the number of explanatory variables and n the number of observations
COMPARING COEFFICIENTS
The size of the slope parameters depends on the scaling of the
variables (on which scale a variables is measured), e.g. population in
thousands or in millions etc.
To be able to compare the size effects of different explanatory variables
in a multiple regression we can use standardized coefficients:
ˆ x j ˆ j
bˆ j for j 1,..., k
ˆ y
Standardized coefficients take the standard deviation of the dependent
and explanatory variables into account. So they describe how much y
changes if x changes by one standard deviation instead of one unit. If x
changes by 1 SD – y changes by b_hat SD. This makes the scale of
the regressors irrelevant and we can compare the magnitude of the
effects of different explanatory variables (the variables with the largest
standardized coefficient is most important in explaining changes in the
dependent variable).
SESSION 2
Testing for Violations of Gauss-Markov Assumptions
Solutions to deal with most common violations
Specification Issues:
Non-linear Relationships
Interaction Effects and Dummy Variables
PROBLEMS IN MULTIPLE REGRESSIONS:
1. MULTICOLINEARITY
The higher the correlation the larger the population variance of the
coefficients, the less efficient the estimation and the higher the
probability to get erratic point estimates. Multicolinearity can result in
numerically unstable estimates of the regression coefficients (small
changes in X can result in large changes to the estimated regression
coefficients).
.06 Belgium
Belgium
Japan
Netherlands
Leverage
Japan
Belgium Belgium Switzerland
IrelandSweden
UKFinland Ireland Ireland
.04
Belgium
Ireland Ireland
Ireland
UK
UK Japan
Japan
Belgium Ireland
UKJapan Switzerland
Ireland
Austria Austria
Netherlands
UK
UK Austria
Belgium
UK
Italy Switzerland
Switzerland
Ireland Ireland
UKBelgium
Japan
Belgium
Italy Finland
UKCanadaAustralia
Netherlands
Netherlands
UK Australia
Ireland
Italy
ItalyItaly
Belgium
UK AustraliaFinlandUK Japan Australia Sweden
Belgium
US UK
Austria
Ireland
US Canada
Ireland US Japan Switzerland
Switzerland
Switzerland
France
Italy
UK
Belgium
Italy
Canada
Italy Italy
Italy
Germany
UK Australia
US
UKFinland
Ireland
UK
UK Japan
Italy
GermanyJapan
Canada
Japan Australia
Australia Switzerland
.02
Italy
Japan
CanadaCanada
Germany
US Canada
Germany
Ireland
Netherlands
Ireland Canada
Germany
Denmark Ireland
Netherlands Japan Australia
Denmark Switzerland
Ireland
Ireland
US Netherlands
Germany
Denmark
Netherlands
Germany
Ireland
Belgium
Germany
CanadaItalyUS
Ireland
Belgium
Finland
Netherlands
Italy
Ireland
Finland Ireland
Canada
Australia Switzerland
Switzerland
SwedenSweden
Japan
Germany
Germany
Ireland
Ireland
Italy
Canada
Finland
France
Japan
Italy
US
UK Sweden
Canada
UK Ireland
Netherlands
Finland
Germany
Canada
ItalyNorway
Italy France
Belgium
Italy
Italy Ireland
Ireland Australia
France
Switzerland
Australia Switzerland
Switzerland
France
Japan
US
Austria
Belgium US
Netherlands
Ireland
Canada
Netherlands
Canada
Italy Netherlands
Netherlands
Netherlands
BelgiumBelgium
Australia
UK Denmark AustraliaSwitzerland
Australia Switzerland
Australia Switzerland
Switzerland
Finland
UK
UK
US Ireland
Italy
UK Germany
Canada
Norway
UK
France
US
Norway
Austria
Finland
Austria
Canada
UK
Canada
DenmarkItaly
Canada
Finland
Austria
Canada Netherlands
Netherlands
Norway
Canada
Finland
Austria
Finland
SwedenNetherlands
Germany
Denmark
Germany
Japan
US Belgium
Italy
Italy
NorwayBelgium
Australia
Netherlands Australia
Australia Sweden Sweden
Sweden
Sweden
Denmark
Belgium
Japan
Belgium
US
Netherlands
Belgium
US
US Austria
Finland
Austria
France
Belgium
UK Canada
Finland
Norway Norway
Denmark
Italy
Netherlands
Italy
Norway
Norway
Belgium
Denmark
Netherlands Canada
Italy Finland
Belgium
Sweden
Netherlands
Sweden
Australia
Austria
Netherlands
US
Denmark
France
Austria
Canada France
Austria
Australia
N France
Australia
orway Norway Denmark
Switzerland
FranceNorway
Denmark Denmark Sweden
Japan
Netherlands
Austria
ItalyNorway
Canada Finland
Belgium
Austria Austria
Finland
Sweden
Norway
Netherlands
Austria
Austria
Sweden
USFrance Norway
Sweden
Netherlands
Norway
Finland
Norway
Japan
Sweden
US France
Sweden
Norway
Japan Australia
Norway
Norway
Canada
CanadaSweden
Austria
Germany
Japan Australia
Netherlands
Australia
Australia
Norway Denmark
DenmarkDenmark
Norway
DenmarkDenmark
Denmark
Denmark Denmark
Sweden
Denmark Sweden Sweden
US
Belgium
Austria
Denmark
Austria Austria
Finland
Finland
Japan
UK
US
Finland
US
Finland
Sweden
Sweden
Germany
Norway
Germany
Canada
Norway
US
Denmark
US
Austria
Austria France
Sweden
France
Australia Norway France
Denmark
Netherlands
Norway
France France
France
Germany Denmark Denmark DenmarkSweden Sweden Sweden Sweden
Netherlands
Austria
Germany
Germany
US DenmarkFinland
France
Germany
US
Austria
Canada
Australia
France
Finland Sweden
Finland
Denmark
France
France France
France
Norway
France
France
Australia Germany
Australia
Netherlands
Finland
Finland France
Germany
Australia
France
France Finland
France
Germany Sweden
Germany
Netherlands Norway
0
jacknife, bootstrap:
Are both tests and solutions at the same time: they show whether single
observations have an impact on the results. If so, one can use the
jacknifed and bootstrapped coefficients and standard errors which are
more robust to outliers than normal OLS results.
Jacknife: takes the original dataset, runs the same regression N-1 times,
leaving one observation out at a time.
Example command in STATA: „jacknife _b _se, eclass: reg spend unem growthpc
depratio left cdem trade lowwage fdi skand “
Bootstrapping is a re-sampling technique: for the specified number of
repetitions, the same regression is run for a different sample randomly
drawn from the original dataset.
Example command: „bootstrap _b _se, reps(1000): reg spend unem growthpc
depratio left cdem trade lowwage fdi skand “
3. OMITTED VARIABLE BIAS
The effect of omitted variables that ought to be included:
Suppose the dependent variable y depends on two explanatory
variables:
yi 0 1 xi1 2 xi 2 i
But you are unaware of the importance of x2 and only include x
yi 0 1 xi1 i
Solutions:
Robust Huber-White sandwich estimator (GLS)
White Heteroskedasticity consistent VC estimate: manipulates the
variance-covariance matrix of the error term.
More substantially: include omitted variables
Dummies for groups of individuals or countries that are assumed to behave
more similar than others
Tests for Heteroskedasticity:
a. Breusch-Pagan LM test for known form of Heteroskedasticity:
groupwise 2
n
T 2
si
LM 2
2 i 1 s
1
V i | xi 2
H0:
Ha: V i | xi i2
2
1. Estimate the model under H0ei
2. Compute squared residuals:
3. Use squared residuals as dependent variable of auxiliary
regression: RHS: all regressors, their quadratic forms and
interaction terms
ei2 0 1 xi 2 ... k 1 xik k 1 xi22 k 1 xi 2 xi 3 ... q xik2 i
2
Normal Variance of beta:
2
,
N *Var X
ˆ ˆ 1 ˆ
Robust White VC matrix: V
n
n X ' X X ' DX X ' X
1 1
ˆ ˆ
Dˆ diag ei2
i 1 i 1
1
N
N
X i ' X i X i ' yi
ˆ ˆ 1 ˆ 1
i 1 i 1
X
1
1
Estimated covariance matrix: ' X
Omega matrix with heteroscedastic error structure and
contemporaneously correlated errors, but in principle
FGLS can handle all different correlation structures…:
12 21 31 n1
12 22 32 n 2
13 23 32 n3
2n 3n n2
1n
HETEROSKEDASTICITY IN POOLED DATA
Is a bigger problem than in pure cross-sections since the error term
of one unit i can be correlated with the error term of unit j at the same
point in time and different points in time!
Solutions:
• White Heteroskedasticity consistent VC estimate
• More substantially: include omitted variables
• Parks-Kmenta approach: FGLS
• Beck/ Katz: Panel Corrected Standard Errors
• More theoretically: Dummies for groups of individuals or countries
that are assumed to behave more similar than others
THE PARKS-KMENTA APPROACH
Basically a Feasible Generalized Least Squares Estimator which
takes the two-dimensional nature of the pooled data into account.
B&K suggest to use OLS and correct only for the panel structure in the
standard errors:
where
For panel models with contemporaneously correlated and panel
heteroscedastic errors, Omega is an NT x NT block diagonal
matrix with an N x N matrix of contemporaneous covariances,
Sigma, along the diagonal. To estimate Omega, we need an
estimate of Sigma. Since the OLS estimates are consistent, one
can use the OLS residuals to provide a consistent estimate of
Sigma. Let e_it be the OLS residual for unit i at time t. A typical
element of Sigma is estimated by:
Tests:
Durbin-Watson, Durbin’s m, Breusch-Godfrey test
Regress e on lag(e)
Autocorrelation
The error term in t1 is dependent on the error term in t0: not controlling
for autocorrelation violates on of the basic assumptions of OLS and may
bias the estimation of the beta coefficients
i t i t 1 it
The residual of a regression model picks up the influences of those
variables affecting the DV that have not been included in the regression
equation. Thus, persistence in excluded variables is the most frequent
cause of autocorrelation.
Autocorrelation does make no predictions about a trend, though a trend
in the DV is often a sign for serial correlation.
Positive autocorrelation: rho is positive: it is more likely that a positive
value of the error-term is followed by a one and a negative by a negative
one.
Negative autocorrelation: rho is negative: it is more likely that a positive
value of the error-term is followed by a negative one and vice versa.
DW test for first order AC: T
et et 1
2
d t 2
T
e
t 1
2
t
S et xt xt l et et 1 xt xt' l xt l xt'
* 1 T 2 ' 1 p T
T t 1 T t 1 t l 1
1
l 1
p 1
OR a simpler Test:
yit ( 0 yi ,t 1 it )
1
xit
FIRST DIFFERENCE MODELS
y i t x i t i t
2. An estimate of the correlation in the residuals is then obtained by the
following auxiliary regression:
i t i t 1 it
3. A Cochrane-Orcutt transformation is applied for observations
t=2,…,n
yi t yi t 1 x i t x i t 1 i t
4. And the transformation for t=1 is as follows:
1 2 y1
1 2 x1 1 2 1
Test:
Ramsay RESET F-test gives a first indication for the whole model
In general, we can use acprplot to verify the linearity assumption
against an explanatory variable – though this is just “eye-balling”
Theoretical expectations should guide the inclusion of squared terms.
Augmented component plus residual
-10 -5 0 5 10
.2 .4 0 .6
institutional openness to trade standardized
1 .8
• Different functional forms give parameter estimates that have different substantial
interpretations. The parameters of the linear model have an interpretation as marginal
effects. The elasticities will vary depending on the data. In contrast the parameters of
the log-log model have an interpretation as elasticities. So the log-log model assumes
a constant elasticity over all values of the data set. Therefore the coefficients of a log-
linear model can be interpreted as percentage changes – if the explanatory variable
changes by one percent the dependent variable changes by beta percent.
• The log transformation is only applicable when all the observations in the data set are
positive. This can be guaranteed by using a transformation like log(X+k) where k is a
positive scalar chosen to ensure positive values. However, careful thought has to be
given to the interpretation of the parameter estimates.
• For a given data set there may be no particular reason to assume that one functional
form is better than the other. A model selection approachyi is
to
1 estimate
x i 2 x i2 competing
i
models by OLS and choose the model with the highest R-squared.
include an additional squared term of the IV to test for U-shape and inverse U-
shape relationships. Careful with the interpretation! The size of the two
coefficients (linear and squared) determines whether there is indeed a u-shaped
or inverse u-shaped relationship.
Hausken, Martin, Plümper 2004: Government Spending and Taxation in
Democracies and Autocracies, Constitutional Political Economy 15, 239-
59.
polity_sqr polity govcon
0 0 20.049292
0.25 0.5 18.987946
1 1 18.024796
2.25 1.5 17.159842
4 2 16.393084
6.25 2.5 15.724521
9 3 15.154153
12.25 3.5 14.681982
16 4 14.308005
20.25 4.5 14.032225
25 5 13.85464
30.25 5.5 13.775251
36 6 13.794057
42.25 6.5 13.911059
49 7 14.126257
56.25 7.5 14.43965
64 8 14.851239
72.25 8.5 15.361024
81 9 15.969004
90.25 9.5 16.67518
100 10 17.479551
The „u“ shaped relationship between democracy and government
spending:
21
government consumption in % of GDP
20
19
18
17
16
15
14
13
0 2 4 6 8 10
degree of democracy
2. INTERACTION EFFECTS
Two explanatory variables do not only have a
direct effect on the dependent variable but also a
combined effect
yi 1 x 1i 2 x 2i 3 x1i x 2i i
45
Low cristian democratic
portfolio
High cristian democratic
portfolio
40
35
Low unemployment High unemployment
70
6050
spend
40 30
20
0 5 10 15 20
unem
0 50 100 150
trade
.015
Marginal Effect of unemployment
.01
.5 1
.005
0
0
0 50 100 150
international trade exposure
Thick dashed lines give 95% confidence interval.
Thin dashed line is a kernel density estimate of trade.
3. DUMMY VARIABLES
An explanatory variable that takes on only the
values 0 and 1
Example: DV: spending, IV: whether a country
is a democracy (1) or not (0).
yi D i