F (A) P (X A) : Var (X) 0 If and Only If X Is A Constant Var (X) Var (X+Y) Var (X) + Var (Y) Var (X-Y)
F (A) P (X A) : Var (X) 0 If and Only If X Is A Constant Var (X) Var (X+Y) Var (X) + Var (Y) Var (X-Y)
F ( a )=P ( X a )
Properties:
F(-) = 0
F(+) = 1
c ,
P ( X >c )=1F ( c )
and
b ,
P ( a< X b )=F ( a )F ( b )
Expected Value
For any constant
c ,
E ( c )=c
constant
For any constants
and
b ,
E ( aX +b )=aE ( X )+ b
Variance
2 Var (X ) E [ ( X )2 ]
2=E ( X 22 X+ 2 ) =E ( X 2 ) 2 2 + 2= E ( X 2 ) 2
Standard deviation
sd ( X ) Var ( X )
Z=
E ( Z )=
1
E ( X ) = =0
Var ( Z )=Var
X = 2 Var ( X )= 2 =1
Covariance
XY Cov(X , Y ) E [ ( X X ) ( Y Y ) ]
Cov ( X , Y )=0
Var ( aX +bY )=E [ ( aX+ bY )2 ]E2 ( aX +bY )=E ( a2 X 2 +2 abXY +b2 Y 2 )E ( aX + bY ) E ( aX + bY )=a 2 E ( X 2) + 2
Conditional expectation
E [ c ( X )X ]=c ( X )
E [ a ( X ) Y +b ( X ) X ]=a ( X ) E ( Y | X )+ b ( X )
c (X)
for any functions
a(X)
and
b(X)
E ( Y |X )= E(Y )
Correlation coefficient
Suppose we want to know the relationship between amount of education
and annual earnings in the working population. We could let X denote
education and Y denote earnings and then compute their covariance. But
the answer we get will depend on how we choose to measure education and
earnings. A covariance property implies that the covariance between
education and earnings depends on whether earnings are measured in
dollars or thousands of dollars, or whether education is measured in months
or years. It is pretty clear that how we measure these variables has no
bearing on how strongly they are related. But the covariance between them
does depend on the units of measurement. The fact that the covariance
depends on units of measurement is a deficiency that is overcome by the
correlation coefficient.
1 Corr ( X , Y ) 1 , if
Corr ( X ,Y ) =0
X = Z
i=1
T=
X
n
Unbiasedness
An estimator, W, of a parameter is unbiased if its expected value is
the parameter : E(W) =
Unbiasedness does not mean that we always get the correct value for
the parameter . It means that we get the correct value in
expectation.
n
1
W 1= = Y 1
n i=1
E ( W 1 ) =E
o
1
1
1
Y 1 = E ( Y i )=
( n )=
n i=1
n i=1
n
()
W 2=Y 1
E ( W 2 ) =E ( Y i )=
W 3=( Y 1+Y 2)
Bias:
E ( W 3 ) =E ( Y 1 +Y 2 ) =2 =
W 3=( Y 1+Y 2) /3
Bias:
E ( W 4 )=
E ( Y 1 +Y 2 )
2
1
= =
3
3
3
W 1= =
1
Y
n i=1 1
E ( Y i )=
1
(n 2)=
2
n
n
( )
1
1
Var ( W 1 )=Var Y 1 = 2
n i=1
n i=1
o
W 2=Y 1
Var ( W 2 )=Var ( Y i ) = 2
W 3=( Y 1+Y 2)
Var ( W 3 )=Var ( Y 1 +Y 2 )=2 2
Efficiency
If W1 and W2 are two unbiased estimators of , W1 is efficient
relative to W2 if:
Var(W1) Var(W2) ,for all
One way to compare biased estimators is to compute the mean
squared error (MSE):
MSE(W) E[(W )
The MSE measures how far the estimator is away from , on average
it can be shown that
MSE(W) = Var(W) + Bias(W) 2
For an unbiased estimator
MSE(W) = Var(W)
An estimator W is consistent if both its bias and variance tend to zero as the
sample size increases:
Quantitative data
Measurement bias occurs when the data collected contains errors that are
non-random:
Recall bias: respondents recall some events more vividly than others
e.g., child deaths by guns vs. swimming pools;
Sensitive questions: respondents may not report data accurately
e.g., wages, health conditions;
Faulty
equipment:
equipment
that
exhibits
systematic
measurement errors e.g., a thermometer that is off by 1 degree
Celsius.
y = 0 + 1x + u
The parameters 1 and 0 represent the slope and intercept of the relation
between x and y, respectively.
n
SST ( y i )
i=1
SSE ( i ) 2
i=1
SSR ( i ) = i
i=1
i=1
The total sum of squares corresponds to the sum of the explained sum of
squares and the residual sum of squares: SST = SSE + SSR
Goodness of Fit: the R-squared of a regression corresponds to the fraction of
the sample variation in y that is explained by x
R2 = SSE/SST = 1
SSR/SST
R-squared is a measure of how much of the total variance of the dependent
variable is explained by our model. Regressions with low r-squared can still
be useful. Many models in the social sciences have a low r-squared but are
still informative about the relationship of two variables; a low r-squared just
means that there are other factors affecting our dependent variable.
The main drawback of using simple regression analysis for empirical work is
that it is hard to draw ceteris paribus conclusions about how x affects y.
OLS.1 is often unrealistic (E(u|x) = 0)
Multiple regression analysis is more suitable for ceteris paribus analysis
because it allows to explicitly control for other factors that also affect the
dependent variable.
What happens if we include an irrelevant variable in a multiple regression
analysis?
Expected value does not change -> E(j) = j , E(3) = 0
But variance inflates, which is a bad thing.