Course1 Review
Course1 Review
Christophe Gaillac
Autumn 2024
Motivation
1 Statistical tools
2 Causal inference
3 Additional references
r(X )
X= ∑ sj uj vj′ ,
j =1
Yi = Xi′ β 0 + ε i , ε i ∼ N (0, σ2 ),
with
Xi are 1 × p vectors, Yi and ε i are scalars
The number of covariates p may be larger than the number of
observations n...
Yi = Xi′ β 0 + ε i , ε i ∼ N (0, σ2 ),
with
Xi are 1 × p vectors, Yi and ε i are scalars
The number of covariates p may be larger than the number of
observations n...
1 n
∑
2
βb(λ) ∈ argmin Yi − Xi′ β + λ∥ β∥1 .
β ∈ Rp n i =1
The Lasso minimizes the sum of the empirical average quadratic loss
and a penalty or regularization term λ∥ β∥1 .
1 n
∑
2
βb(λ) ∈ argmin Yi − Xi′ β + λ∥ β∥1 .
β ∈ Rp n i =1
The Lasso minimizes the sum of the empirical average quadratic loss
and a penalty or regularization term λ∥ β∥1 .
4 Aggregate the errors from the previous step and then minimize with
respect to λ:
K 2
1 1
λL = argmin
b
K ∑
n ∑ Yi − Xi′ βbLk (λ) .
λ ≥0 k =1 0 i : Gi =k
E X Y − X ′ β 0 = 0.
xt = Λft + ε t , t = 1, . . . , T ,
where
Λ is a weighting matrix of the size factors p × r “factors loading”;
ft is a vector of factors r × 1;
ε t are idiosyncratic errors.
In matrix form:
X = F Λ′ + E ,
T ×p T ×r r ×p T ×p
xt = Λft + ε t , t = 1, . . . , T ,
where
Λ is a weighting matrix of the size factors p × r “factors loading”;
ft is a vector of factors r × 1;
ε t are idiosyncratic errors.
In matrix form:
X = F Λ′ + E ,
T ×p T ×r r ×p T ×p
xt = Λft + ε t , t = 1, . . . , T ,
where
Λ is a weighting matrix of the size factors p × r “factors loading”;
ft is a vector of factors r × 1;
ε t are idiosyncratic errors.
In matrix form:
X = F Λ′ + E ,
T ×p T ×r r ×p T ×p
One can use Factor models to reduce the dimension of some of the
variables while keeping the meaning of some others using Factor
augmented regression (see Bai (2003)).
Figure: *
I. Statistics and Econometrics Prerequisites 19 / 35
Random trees
Figure: *
I. Statistics and Econometrics Prerequisites 20 / 35
Random trees
Figure: *
I. Statistics and Econometrics Prerequisites 21 / 35
Random trees
Figure: *
I. Statistics and Econometrics Prerequisites 22 / 35
Random forests
We aggregate the trees formed on all possible subsamples of size s
from the training data U1 , . . . , Un :
−1
n
µ̂(x ; U1 , . . . , Un ) =
s ∑ T (x ; Ui1 , . . . , Uis ), (8)
1≤i <...<i ≤n
1 s
1 Statistical tools
2 Causal inference
3 Additional references
∆i = Yi (1) − Yi (0).
1 n Di Yi (1 − Di )Yi
∗
τIPW = ∑ − .
n i =1 p (Xi ) 1 − p (Xi )
n
1 Di (Yi − µ1 (Xi )) (1 − Di )(Yi − µ0 (Xi ))
τbAIPW =
n ∑ µ1 (Xi ) − µ0 (Xi ) + p ( Xi )
−
1 − p (Xi )
.
i =1
√ σ02 (X )
2
d σ (X )
n(τbAIPW − τ0 ) → N 0, Var(τ (X )) + E +E 1 ,
1 − p (X ) p (X )
w 7 → E [S |W = w ]
1 Statistical tools
2 Causal inference
3 Additional references