STAT 135: Linear Regression: Joan Bruna
STAT 135: Linear Regression: Joan Bruna
Joan Bruna
Department of Statistics
UC, Berkeley
May 1, 2015
W ≈ β0 + β1 S .
W ≈ β0 + β1 S .
Is the data consistent with our model? ie, is our model correct?
Is the data consistent with our model? ie, is our model correct?
How can we estimate the parameters of our model? Precision?
F (W , H) ≈
√ 1
F (W , H) ≈ β H , (with β = p ).
9.8/2
P = β0 + β1 Y
P(Y ) = γ0 e γ1 (Y −Y0 ) ,
where γ0 is the population at year Y0 .
P(Y ) = γ0 e γ1 (Y −Y0 ) ,
where γ0 is the population at year Y0 .
How to estimate the growth rate from the data? Simple idea:
transform the data to reveal the linear dependency:
P(Y ) = γ0 e γ1 (Y −Y0 ) ,
where γ0 is the population at year Y0 .
How to estimate the growth rate from the data? Simple idea:
transform the data to reveal the linear dependency:
In these examples,
p−1
X
y ≈ f (x1 , . . . , xp−1 , β0 , . . . , βp−1 ) = β0 + βk xk .
k=1
In these examples,
p−1
X
y ≈ f (x1 , . . . , xp−1 , β0 , . . . , βp−1 ) = β0 + βk xk .
k=1
In these examples,
p−1
X
y ≈ f (x1 , . . . , xp−1 , β0 , . . . , βp−1 ) = β0 + βk xk .
k=1
∂E (β)
= 0 , k = 1...p
∂βk
is Linear.
Then
∂E X
= −2 (yi − β0 − β1 xi ) ,
∂β0
i=1
∂E X
= −2 xi (yi − β0 − β1 xi ) .
∂β1
i=1
2 (
P P P P
xii i y i ) − ( i xi ) ( i xi yi )
βˆ0 = P 2 P 2
,
n i xi − ( i xi )
P P P
n ( i xi yi ) − ( i xi ) ( i yi )
βˆ1 = P 2 P 2
n i xi − ( i xi )
W ≈ 325 + 0.26S .
2 (
P P P P
xii i y i ) − ( i xi ) ( i xi yi )
βˆ0 = P 2 P 2
,
n i xi − ( i xi )
P P P
n ( i xi yi ) − ( i xi ) ( i yi )
βˆ1 = P 2 P 2
n i xi − ( i xi )
W ≈ 325 + 0.26S .
yi = β0 + β1 xi + i , i = 1 . . . n ,
with i ∼ N (0, σ 2 ) iid.
Theorem
Under the previous assumptions, we have
E β̂j = βj , (j = 0, 1)
and
σ2 xi2 nσ 2
P
var βˆ0 = i
, var βˆ1 =
xi2 + ( x i )2 xi2 + ( xi )2
P P P P
n i i n i i
If we denote
1X 1X 1X
sx = (xi −x)2 , sy = (yi −y )2 , sxy = (xi −x)(yi −y )
n n n
i i i
If we denote
1X 1X 1X
sx = (xi −x)2 , sy = (yi −y )2 , sxy = (xi −x)(yi −y )
n n n
i i i
If we denote
1X 1X 1X
sx = (xi −x)2 , sy = (yi −y )2 , sxy = (xi −x)(yi −y )
n n n
i i i
ŷ = βˆ0 + βˆ1 x
ŷ = βˆ0 + βˆ1 x
r r
sy sy
ŷ = y − ρ x +ρ x
sx sx
ŷ = βˆ0 + βˆ1 x
r r
sy sy
ŷ = y − ρ x +ρ x
sx sx
ŷ − y x −x
√ =ρ √ .
sy sx
ŷ = βˆ0 + βˆ1 x
r r
sy sy
ŷ = y − ρ x +ρ x
sx sx
ŷ − y x −x
√ =ρ √ .
sy sx
Let us write
Let us write
Then
ŷ = X β .
1
E (β) = ky − X βk2 .
2
If
E (x) : Rn −→ R ,
the derivative or gradient of E with respect to x is written
∂E
∇x E ∈ Rn , with (∇x E )i = .
∂xi
Example: Say
1 1X 2
E (x) = kxk2 = xi .
2 2
i
If
E (x) : Rn −→ R ,
the derivative or gradient of E with respect to x is written
∂E
∇x E ∈ Rn , with (∇x E )i = .
∂xi
Example: Say
1 1X 2
E (x) = kxk2 = xi .
2 2
i
Then
∇x E = x .
E = F (y ) , y = G (x) ,
with
E ∈ R , y ∈ Rm , x ∈ Rn .
E = F (y ) , y = G (x) ,
with
E ∈ R , y ∈ Rm , x ∈ Rn .
Question: How to compute the gradient of E with respect to x?
E = F (y ) , y = G (x) ,
with
E ∈ R , y ∈ Rm , x ∈ Rn .
Question: How to compute the gradient of E with respect to x?
A: Chain Rule is defined as usual:
T
∂G
∇x E = (y ) ∇y F .
∂x
∇z E = z and (∂β z) = −X .
∇z E = z and (∂β z) = −X .
Therefore
∇x E = −X T z = −X T (y − X β) .
It results that
∇x E = 0 ⇔ X T X βb = X T y ⇔ βb = (X T X )−1 X T y .
It results that
∇x E = 0 ⇔ X T X βb = X T y ⇔ βb = (X T X )−1 X T y .
It results that
∇x E = 0 ⇔ X T X βb = X T y ⇔ βb = (X T X )−1 X T y .
X X
xi2 − xi
1
(X T X )−1
i i
= P 2 X .
n i xi − ( i xi )2 −
P
xi n
i
Lemma
Let Y be a random vector with mean µ and covariance Σ. Then, if
Z = b + AY is a linear transformation, we have
E (Z ) = b + Aµ , and ΣZ ,Z = AΣAT .
Thus
Z = b + AY ,
and
E (Z ) = b + AE (Y ) = b + A(µ + E ()) = b + Aµ .
Thus
Z = b + AY ,
and
E (Z ) = b + AE (Y ) = b + A(µ + E ()) = b + Aµ .
Also,
ΣZ = σ 2 AAT .
Lemma
Under the same assumptions as the previous theorem, let
X = Y T AY ∈ R. Then
E (X ) = Tr (AΣ) + µT Aµ .
s = σ 2 Tr (A) = σ 2 (n − 1) .
Lemma
Let X ∈ Rn be random vector with covariance Σ. If Y = AX ∈ Rp
and Z = BX ∈ Rm are linear transformations of X , where A and B
are constant matrices, then
ΣY ,Z = AΣB T .
Y = Xβ + ,
with
X ∈ Rn×p known and fixed,
β ∈ Rp unknown parameters,
∈ Rn with E () = 0 and Σ = σ 2 1.
Y = Xβ + ,
with
X ∈ Rn×p known and fixed,
β ∈ Rp unknown parameters,
∈ Rn with E () = 0 and Σ = σ 2 1.
β̂ = X † y = (X T X )−1 X T y .
Theorem
If E () = 0, then
E β̂ = β .
Theorem
If E () = 0, then
E β̂ = β .
Theorem
If E () = 0 and Σ = σ 2 1, then
Σβ̂ = σ 2 (X T X )−1 .
Recall that
X X
xi2 − xi
1
(X T X )−1
= P 2 X i i .
P 2
n i xi − ( i xi ) − xi n
i
Recall that
X X
xi2 − xi
1
(X T X )−1
= P 2 X i i .
P 2
n i xi − ( i xi ) − xi n
i
It results that
var βˆ0 cov βˆ0 , βˆ1
Σβ̂ =
cov βˆ0 , βˆ1 var βˆ1
2P 2
−σ 2 i xi
P
1 σ P i xi
= P 2
xi )2 −σ 2 i xi σ2n
P
n x −(
i i i
Theorem
Under the iid Gaussian statistical model Y ∼ N (βX , σ 2 Id),
β̂ = X † y
ê = Y − Ŷ = Y − X β̂
ê = Y − Ŷ = Y − X β̂
Theorem
Under the assumption that errors are uncorrelated with constant
variance σ 2 ,
kêk2
s2 =
n−p
is an unbiased estimator of σ 2 : E s 2 = σ 2 .
Y − Ŷ = Y − X β̂ = (Id − X (X T X )−1 X T )Y .
Y − Ŷ = Y − X β̂ = (Id − X (X T X )−1 X T )Y .
Y − Ŷ = Y − X β̂ = (Id − X (X T X )−1 X T )Y .
It results that
X
kêk2 = (Yi − Ŷi )2 = kY − Ŷ k2 = Y T PX ⊥ Y ,
i
It results that
X
kêk2 = (Yi − Ŷi )2 = kY − Ŷ k2 = Y T PX ⊥ Y , so
i
E kêk2 = E (Y )T PX ⊥ E (Y ) + σ 2 tr (PX ⊥ ) .
It results that
X
kêk2 = (Yi − Ŷi )2 = kY − Ŷ k2 = Y T PX ⊥ Y , so
i
E kêk2 = E (Y )T PX ⊥ E (Y ) + σ 2 tr (PX ⊥ ) .
We have
PX ⊥ E (Y ) = PX ⊥ X β = 0 .
It results that
X
kêk2 = (Yi − Ŷi )2 = kY − Ŷ k2 = Y T PX ⊥ Y , so
i
E kêk2 = E (Y )T PX ⊥ E (Y ) + σ 2 tr (PX ⊥ ) .
We have
PX ⊥ E (Y ) = PX ⊥ X β = 0 .
tr (PX ⊥ ) = tr (Id) − tr (X (X T X )−1 X T ) =
n − tr ((X T X )−1 X T X ) = n − p .
It results that
X
kêk2 = (Yi − Ŷi )2 = kY − Ŷ k2 = Y T PX ⊥ Y , so
i
E kêk2 = E (Y )T PX ⊥ E (Y ) + σ 2 tr (PX ⊥ ) .
We have
PX ⊥ E (Y ) = PX ⊥ X β = 0 .
tr (PX ⊥ ) = tr (Id) − tr (X (X T X )−1 X T ) =
n − tr ((X T X )−1 X T X ) = n − p .
So E kêk2 = σ 2 (n − p).
Σê,ê = PX ⊥ ΣY ,Y PXT⊥ = σ 2 PX ⊥ .
Σê,ê = PX ⊥ ΣY ,Y PXT⊥ = σ 2 PX ⊥ .
Y − Ŷi Yi − Ŷi
pi = p .
s PX ⊥ (i, i) s 1 − PX (i, i)
Ŵ = 325.6 + 0.26S .
β̂ = X † Y ∼ N (β, σ 2 (X T X )−1 ) .
β̂ = X † Y ∼ N (β, σ 2 (X T X )−1 ) .
β̂k − βk
∀k , p ∼ tn−p .
s Ck,k
β̂k − βk
∀k , p ∼ tn−p .
s Ck,k
β̂k − βk
∀k , p ∼ tn−p .
s Ck,k
β̂k − βk
∀k , p ∼ tn−p .
s Ck,k
Σβ̂ = σ 2 (X T X )−1 .
Σβ̂ = σ 2 (X T X )−1 .
Σβ̂ = σ 2 (X T X )−1 .
Σβ̂ = σ 2 (X T X )−1 .
√
Ŷ = β0 + β1 X , withβ0 ∈ (−0.03 ± 0.13) , β1 ∈ (0.459 ± 0.018) .
√
Ŷ = β0 + β1 X , withβ0 ∈ (−0.03 ± 0.13) , β1 ∈ (0.459 ± 0.018) .
√
Ŷ = β0 + β1 X , withβ0 ∈ (−0.03 ± 0.13) , β1 ∈ (0.459 ± 0.018) .
Outliers are extreme values that greatly influence the rest of the
fitted model parameters.
Outliers are extreme values that greatly influence the rest of the
fitted model parameters. The quadratic loss is very sensitive to
such extreme values:
Outliers in X (Monaco). How would β̂ change if we removed
an “extreme” observation (x ∗ , y ∗ ) ?
n
X n
X
(X T X )k,l = xi,k xi,l = xk∗ xl∗ + xi,k xi,l ,
i=1 xi 6 x∗
=
X n
X
(X T Y )k = xi,k yi = xk∗ y ∗ + xi,k yi .
i xi 6=x ∗
Outliers are extreme values that greatly influence the rest of the
fitted model parameters. The quadratic loss is very sensitive to
such extreme values:
Outliers in X (Monaco). How would β̂ change if we removed
an “extreme” observation (x ∗ , y ∗ ) ?
n
X n
X
(X T X )k,l = xi,k xi,l = xk∗ xl∗ + xi,k xi,l ,
i=1 xi 6 x∗
=
X n
X
(X T Y )k = xi,k yi = xk∗ y ∗ + xi,k yi .
i xi 6=x ∗
Model misfits. Suppose that one observation (eg Iceland) does not
follow the specified model. Q: How much is it going to degrade
the overall fit?
Model misfits. Suppose that one observation (eg Iceland) does not
follow the specified model. Q: How much is it going to degrade
the overall fit?
Suppose y∗ = x∗ β + + γ with |γ| ||.
Model misfits. Suppose that one observation (eg Iceland) does not
follow the specified model. Q: How much is it going to degrade
the overall fit?
Suppose y∗ = x∗ β + + γ with |γ| ||. Then
β̂ = X † Y = X † Ỹ + γX∗†
Model misfits. Suppose that one observation (eg Iceland) does not
follow the specified model. Q: How much is it going to degrade
the overall fit?
Suppose y∗ = x∗ β + + γ with |γ| ||. Then
β̂ = X † Y = X † Ỹ + γX∗†
Σβ = σ 2 (X T X )−1 .
Σβ = σ 2 (X T X )−1 .
Ŷ = X β̂ ,
Yˆ∗ = x∗ β̂ .
Its variance is
var Yˆ∗ = x∗ Σβ̂ x∗T = σ 2 x∗ (X T X )−1 x∗T ,
n
σ2 X
1X
var Ŷ (xi ) = xi (X T X )−1 xiT
n n
i i=1
σ2 σ2
= Tr (X (X T X )−1 X T ) = Tr (Idp×p )
n n
p
= σ2 .
n
n
σ2 X
1X
var Ŷ (xi ) = xi (X T X )−1 xiT
n n
i i=1
σ2 σ2
= Tr (X (X T X )−1 X T ) = Tr (Idp×p )
n n
p
= σ2 .
n
Y = Xβ + = θ + .
Y = Xβ + = θ + .
Y = Xβ + = θ + .
Source: Worldbank.org
Joan Bruna STAT 135: Linear Regression
Example
β̂ = (X T X )−1 Y , and
\
GDP(t + 1) = β̂1 CO2 (t) + βˆ2 GDP(t) + · · · + β̂7 Exports(t + 1) .
Also
[ 2
kGDP − GDPk
s2 = .
50 − 7
According to our model, this year’s CO2 emissions for France will
be
5.57 ± 0.78 ( metric tons per capita) .
Joan Bruna STAT 135: Linear Regression
Model Selection
Suppose a model
Y = X β + , β ∈ Rp , X ∈ R1×p ,
with E () = 0, Σ = σ 2 Id, and observations (xi , yi ) , i = 1 . . . n.
Suppose a model
Y = X β + , β ∈ Rp , X ∈ R1×p ,
with E () = 0, Σ = σ 2 Id, and observations (xi , yi ) , i = 1 . . . n.
Suppose a model
Y = X β + , β ∈ Rp , X ∈ R1×p ,
with E () = 0, Σ = σ 2 Id, and observations (xi , yi ) , i = 1 . . . n.
Suppose a model
Y = X β + , β ∈ Rp , X ∈ R1×p ,
with E () = 0, Σ = σ 2 Id, and observations (xi , yi ) , i = 1 . . . n.
Xn
2 2
E R̂tr (S) = E (ŶS (xi ) − Yi ) = E kŶS − Y k .
i=1
It turns out that the training error is a biased estimator of the test
error:
Theorem
E R̂tr (S) = R(S) − 2Tr (ΣŶ ,Y ) .
It turns out that the training error is a biased estimator of the test
error:
Theorem
E R̂tr (S) = R(S) − 2Tr (ΣŶ ,Y ) .
Remarks:
The data is being used twice: to fit the model and then to
estimate the risk.
It turns out that the training error is a biased estimator of the test
error:
Theorem
E R̂tr (S) = R(S) − 2Tr (ΣŶ ,Y ) .
Remarks:
The data is being used twice: to fit the model and then to
estimate the risk.
The Cross-covariance between Ŷ and Y increases as the
model becomes more complex.
It turns out that the training error is a biased estimator of the test
error:
Theorem
E R̂tr (S) = R(S) − 2Tr (ΣŶ ,Y ) .
Remarks:
The data is being used twice: to fit the model and then to
estimate the risk.
The Cross-covariance between Ŷ and Y increases as the
model becomes more complex.
How to estimate the risk more reliably, ie how to choose the
best model size?
1 X
R̂ = R̂k .
K
k≤K
et
The function f (t) = is the logistic function:
1 + et
n T
X xi e β xi
∇`(β) = xi yi −
i=1
1 + e β T xi
n
X
= xi (yi − p(xi , β)) .
i=1
n T
X xi e β xi
∇`(β) = xi yi −
i=1
1 + e β T xi
n
X
= xi (yi − p(xi , β)) .
i=1
n T
X xi e β xi
∇`(β) = xi yi −
i=1
1 + e β T xi
n
X
= xi (yi − p(xi , β)) .
i=1
n T
X xi e β xi
∇`(β) = xi yi −
i=1
1 + e β T xi
n
X
= xi (yi − p(xi , β)) .
i=1
f (tn )
tn+1 = tn − .
f 0 (tn )
f (tn )
tn+1 = tn − .
f 0 (tn )
we have
∇`(β) = X T (y − p) ,
and
∂ 2 `(β)
= −X T WX ,
∂β∂β T
with W a diagonal matrix Wi,i = p(xi , β)(1 − p(xi , β).
β n+1 = β n + (X T WX )−1 X T (y − p)
= (X T WX )−1 X T W (X β n + W −1 (y − p))
= (X T WX )−1 X T Wz ,
with z = X β n + W −1 (y − p).
β n+1 = β n + (X T WX )−1 X T (y − p)
= (X T WX )−1 X T W (X β n + W −1 (y − p))
= (X T WX )−1 X T Wz ,
with z = X β n + W −1 (y − p).
β̂ → β (n → ∞).
β̂ → β (n → ∞).
N (β, (X T WX )−1 ) .
220
o o o o o oo o o oo
oo ooo o
o oo o o
o ooo o
o ooo oooo o
o o
o
o ooo
ooo o ooo o o o o oo o ooo o
o o oo oo oo ooo
ooo oo ooo ooo
ooooo oooooo oo o o ooooooo ooooooo oooo o
oooo o o o oooooooo oooooo o
oooooooo
oo o ooo
oo
ooooooooooo
ooooooo o oo o oooo oooooooooo o oo
ooo oo oooo o oooo
oooooo
oo
ooo
160
o oooooo oo oooo oooo ooooo o o o oo
oo oooo ooooo oo ooooo o o ooooooo oooooo
sbp o
o o
ooo o o o ooo o ooo o ooooooo
o o oo ooooooo oo ooooooo o o
oooooooo oo o oooooo o o o
ooo
oooooo oo o o oo oo
oo
o oooo
o
oo
oo
ooooo
ooo ooooo ooo o oooooooooo o
ooooo ooooooooo
o ooo
ooo oooo
oooooo
oooo oooo
ooooo o
o o
oo
oo ooo
o
oo ooo
ooooo
oo oo
oo
o
oo ooooooo
oooooooooo ooooooooo
oo oooo o oo
ooooooo ooooo oo o o
o oo o o oooooooooo ooooooooooo
oo o
ooooo
oo ooooooo
o
oooo ooo
ooooo ooo
ooo oo ooo
ooooo oo
o
oo
o o o
oooo
o o
o o ooooooo
oo o oo o
oooooooo oo
o
oo
o
o
oo o
o o
oo ooooo ooo
o o ooo o
ooooo o o
o oo
oo
o o
o ooo
o o o oo o
o o oo oo
ooooo
o o
ooooooo
oooo ooooooo
o oooo
oooo oo
ooo ooo
o o o
o o
oo ooo
o
o o o o o o o
o
o o oo
o o
ooo
oo ooooo
o o ooooo
oooooo
oooooo
ooo oooo
ooo
oooooo
ooooo ooooooo
oo ooo oo
oo ooo
o ooooo o oooo oooooooooooooo ooo ooo o ooo oo oooooo
oooo ooo o o
oooooooo ooooo ooooo
ooooo
oooo ooooooo ooo o
oo ooo
o ooo oooo ooo oooo
o
o
ooo o oooooo o oo oo oo o o
oo o
ooo
o o o
o o oo
oo
o
o
oo
oo
ooo
o o
ooooooo oo ooo o o
o oo
o o
ooo o
ooo ooooo
oo oooooooooo ooo
oo oo
ooooo
o
oo oo
oo ooo
oooo
o
oo o oooooo ooo o oooo
ooo o o o
oooo oooo o oo
o o o oooooo o oooo
oo o
o ooooo o
ooooooo o o o o o oooo o oo
oooooooooooooooooo o o o oo o o
ooooo o
o
ooooooo oooo ooooo o o oo o
ooooo o
oooooo
o o o o
oooo o o oo o o oo o o
o o o oo o oooo o o o o oo
100
oooo oo ooo
o o o ooo oooo o o
o
30
o o o o o
o o o o o
o o o
o o o o
20 oo oo o o o
o ooo o ooo o ooooo o ooo o o o
o ooo
o
o oo oo o tobacco ooooo o oo
o ooo oooo o
o o o o o o oo
o oooo o ooooooo oooooo oooo ooo o oooo
o oo ooo oo ooooo ooo
oo oooooo
oooooooo
oo o
oooo ooo oo o
o
o ooooooooooooo o o
ooooooooooo o oooo
o oo oo o o o ooo oo o
10
ooooooo o o o o oo oooo
o ooooooo ooooooo o oo oo o o o
o o o o o o o o o
o o o
ooooo ooooo ooooo
o oooo oooooooo oo ooooo oo ooooooooo
oo oo ooo
o
o o
o
o
o ooo
oo o o
oooo oooo oooo
o oooo
o
ooooo
oooooo oooooo o o
o ooo ooo
o
ooo o
o oo
ooooo oo o
oooooooo
ooo oooo
oo oo o o
oooooo
oooo
o
ooooooo o
ooo ooo
o o
oo oo
oo
oo
o oooooo ooooooo
ooooo ooooooooo o o o o o oo
o oo
o oo
oo o ooooo
oo
oooooo oooo
ooo
oo ooooo
o o ooo
oooooooo
oo oo
oo oo oo ooo o
oo oooo oo ooo ooooooo
o oo oo oooooooo oooooo
o oo oo o o
o oooo
o o oo ooo
oo
oooo
o
oooooo oooo
o
o o o oooo
oooo ooooo
ooo oo
oo
oo
o
oo o
oooo ooo
oo oo o o
oo o ooo
oo o oooooooooooo oooooo
oo
ooo o oo oooooooooooooooo o
o o o
ooo o
ooooo
ooo
o o
oooo
ooo
oooo ooo ooo o o
ooooooooooooo o o oooooo o
ooooooooo o
o
ooo oooo
ooo o o
oooooo o o
ooo
o o
oooooooo oo o ooo o
oooo
oo o o
ooooooo
oo
oo
ooooooo
oo ooooo
oo
ooo oo
ooo oo
ooo
o oo ooo
oooo ooooooooooooooo ooo oo o oo
ooo
oooo
oooo oo
oo
o ooo ooooo
oo ooooooooooooooooooo oo
o oo ooo oooooooo ooooooo o o
ooooooo
oo oo
ooo oooo o oo
ooo
oo o o
oo ooooo ooooooo ooooooo ooo oo oooooooooooo ooo
oooooooooooooooooooo oooo
0
o
oo o o o oo ooo o o o
o o o o o o
10 14
o o o o o o
ooo o o oooo
o o oooo o oo oo oo o o ooo o o
o oo o oo oooo oo oo o ooo oooo o o
o o oo
ooo oo
o ooooooo ooooo
o ooo o oooo oooo oooooo o
o
oooooooo o
o
o oo o
ooooo ooo o o o o o o oo ooo
ooo
o ooo
ldl ooo ooo o o ooo ooooooooo o o oo oo o oo ooooooooooooooooooooooo
oooo o oo
oooooo o oo o
ooooooooooo oo o oo oooo
oooooooooo ooo o o
ooo o o oo oooooooooo
o ooo oo
o
oo
oooo o o o ooooooooo o o ooooooo o
ooooooooooo ooo o oooooooo ooooooo
oooo ooo oo
ooooooo
o
oo
o oooo ooo ooooo o oo
oooooo
oooo o ooo
o oooo ooooo
oooo o ooooooooooo oooooooo oo oooooo oo
oooooooooooo ooooooo
6
o o o o o ooooooooo oo oo o
o o o o
o o oo oooo o oo
o o o oo oooooo ooooooooooo o o oo oo
ooo o
oo
oo o
ooo
o
ooo
ooooooo o o o o ooo
ooooooooooooo
ooooooooo
oooo oo o o o o oo
o ooo
oooooooooooo
oooo oooo oo o o o o o o
o o ooo
oo o o
ooo
o o
ooooo oo o
oooo
ooo o o
o ooo oo oo o
oooo ooo
o oo oo
oo
ooo ooo
ooooo
ooooo oo o o oooo ooooooooooo ooooooooooo
o ooo
oooooooooo o
oo o o ooo oo
o oo
oo oo
ooo
ooooo oooo
o oo
o
oo
oooo oo
oo
o ooo oo ooo o ooooo
oo ooo
oo oooooooooo
oo oooooooo
oo o o o o
oo
o o
o o
oo
o oooo
o oooooooo
oo
oo
oooo
oo ooo oo
oo
ooooooooo
oo oo ooo o
oo o
oo
oooo
oooo oooooooo
oooooooo ooo ooo
ooooooooo
oooo oooooo
o oo oooo
oooooooooo ooooooo
o oooo ooooooooo ooooo
oooooo oooo
oo oooo ooo ooooooo ooo o oooooooooo
ooooooooooo
o oo
ooo
ooo
o o
o o
ooo oooo
oooo ooo
oooooo
o oo oo ooooooo oooo oo o o o o o
ooo o ooooooo oo o o o oo oo
ooooooooooo
oo oooooo oooooooooo ooo ooo
o o
o oooooooo ooo oo
2
ooooooo oo o o o oo o o
o
oooooo
ooo
oooo oo
ooooo
oooo ooooooo
ooooo oooooooooooooooooooo ooo
o oo
ooo
ooo
o oooooo
oo
oo oooooo
oooo oo
ooooooooooooo o o ooooo
ooo oo
oo ooooooo
oooo
o oo
oooo
oooooooooooo oooo
ooo
oo
o o oo oooo
oo
oo
oo ooooo
oo
oooo
oo oooooo
oo
ooo
ooo
oooo oooo
ooooo ooooooooo ooooo
o
oo oooooooo oooooooooooooooooo ooooo oo oo oooo
oooooooooooo ooooo
ooooo oooooo
oo
ooooooooooo
oo oooooooooooooo
oo ooo
oo oooo
oo
0.8
famhist
0.4
0.0
oooooooooooo
oooo
oo
o oo
oooo ooo
oo
o oooooo
o
oo ooo
oo
o oooooooo oooooooooooooooooooo oooo ooo
oo ooo
oo
oo ooooooooo ooo ooooooooooooooooo
oo oo o oo
oooooo
ooo ooooo
ooo ooooooooo
ooo
oo
ooo
o oooooooooo o o o
oooooo ooo
o
oooooo
ooooooo
oo
oo oo ooooo
ooooo
ooo oo
oo ooooo
ooo oooooo
oooooooo o oooo
oo
oo
ooo
oo
oo
o
oooooooo
o
oo
oo
o ooo ooo
ooo oooooooooo
oooo ooooooooooo
ooooo oooo
ooooooooooooooooooooooo
ooo ooo
oooooooo oooooo oo
oooooo
o ooo
o o o o o o oo oo o
45
o
o o oo oo oo o oo
o o oo oo ooo ooo o oo o oo oo o
oooooo o o
oo oooooo oooo oooo oooooooo o oo
35
o o o o oo o oo ooooo o oo o o o o o o
ooo o ooooooooooooo o
oo ooo
o oo o oooooooo o ooo ooo
oooooooooooooo ooooo
o o
oo
oo
o ooo oo oooooo oo ooooo ooooooooooooooo o o oooo
oooo oo
oo
oooo
oo o ooooooo ooo ooooooo o o oo ooo o oo oooo ooo
oo o o o o oo
ooooooo o o
oo o o obesity o o o o
oo ooooo o ooooooo oo
oooo o ooooooo oo oo oo ooooooooo
ooooo oooo
ooooo o oooo o oooo ooo
ooo oo
o ooooooo oo
ooo ooo ooo oooooooo oooooooooooo o o
o oo oo
ooo o oo
o
oo ooo
oooooo oooooo
ooooooooooooooo
o o oo
oo o o
o
oo
ooo o
oo
oooooooooo
o
oooooooo
o
oo
oooo o o oooooooooooo oooooooo
o oooooo oooooooooo o
ooooo
ooooo
o oo oo oooo
oo ooooo oo o
oooo ooo o oo o o o
oo ooo ooooooo ooooo o ooo
oo o
oo oooooo
ooooooo o o o oooooo o
ooo o
ooo o
ooooo oooo ooooooooooooooooo
o oo oooooo o oo oo
oo
o oooo
ooooo
ooooo ooooo
o ooo
oo ooo
25
oo o oo o ooo o o oooooo o oo
oo ooooo oooo ooooooo ooo ooooooo
ooo ooooo
ooo
oooo o
o oooo
o
o oo
ooooo oooo
oo oooooooooooo
o ooooooo o o o o
oo o
ooooo ooo
oooooo
o oo oo
o
oo oooo o ooo oooo
oo
o
o
oo
o
ooo oo
oo
oooo ooo oo
o oo
ooooooooo oo o
oo o
oo oo oo
o oooo o ooo oooooooo
oooooooooo o o o oooo o
oo
oooooo
oooooooo
ooooooooo
oo
oooo
o
ooooooooo
oooooooo
ooooooo ooo oo ooo oooo ooo
o ooooooooooo o o ooooo oooo oooo
o
oooooo
ooooooooo
ooo oooooooo o
oooo ooooo o oo
o oo
o oo
ooooo ooo
oooo
ooo ooooo ooo oo oo ooooo o
oooooo
ooooooooooo oooo oo o oooooooooo
oooooooo o o o
o o o
ooo
o oooo ooooo oooo
oo oo
15
o o o o o o
ooo o oo o oo o oo o o o
o o o oo o o
oo o
100
o o oo o o o ooo o o o
oo ooooo o o o o o o ooo ooo oo oooo oo o ooo oo o oo o
ooooo ooooo oo oo oo oooooo o ooooooooo oo o o
o
o ooo oooooo
o o
o alcohol oo ooo oooooo ooooo
oo oo oooo o ooo o o oo
oooooooo oo o oooooooooooooooo o o
ooo
o o
oo o oooo oooo ooo oo o o
oooooo oo oo oooooooooooo
50
ooo o o o oo
o ooo oooo o o o o o o o o o oo o o o o o o oooo o oooo oo oooo ooo o oooooooo o ooo ooo o o oo ooooooo oooo oo oooo oooo oooooo o o
o o o o o o o o o o o o o o o
o o oooooo o o o o o o o o o o o oo oo
60
o oo o oooooo oooo ooooooooo o o
oooooooo
oo
oo
oooooo
ooooo
oooooo oooo
oooo oo oo
oooo o ooo o oooo ooooooooo oo ooooooooo o oo oo o oo oo
oo oo o
ooooo oo
ooo ooo ooo o oo oooo o
oooooo oooooooooo oo
o ooooo oo oo
oo oooo ooo
oooo ooo
ooo o
ooooo
ooo
oooooooo
oo
oo oo oo o
ooooooooo oo oo
oo o o o o o oo o o o o
oo o
o o o o o o o o
o o ooo oooo o
o oo ooooooooooooo ooooo oo
ooooooooooo o oo
o o o
o oo o o ooooo ooooo oo oooo ooooo o o ooo ooooooo
oooooooooo
oo
oooo
o
ooooooooooooo ooo o o o o
o o oooo
o ooo o oo
ooooo ooo
o
oooooooooooooo o o
ooo ooo
ooo o ooooooo ooo oo
ooo ooooooooo oo oooooooo ooo oo o
oooooo
o oooooo
ooooooooo
ooooo o
ooo o oo ooooo oo
oooo oooooooooooo o ooooo
ooooo oooooo o
ooo
ooo oo
ooo
o oo
o
o oo o oo ooo
ooooooooooooooo o
o o o oo oooo oo
ooooooo ooo
oo ooo o ooo
ooo
oo oooo oooo oo o o oooo oooo oo oo oo o oooo
oo oooooo o
oooo o oo o
oo ooooo o o o
o oo ooo
oo o
ooooooo ooo
oo oo o
40
ooooo oo ooooooooooo ooooooo oo
o oo o ooo oooooo o ooo ooooooo oo
o ooo ooo
ooooo o
o o oo
o oo ooo
o oo
oo o oo
ooooooooo oooooooo oo oo o oo oo
oooo ooooo o oo age
ooo oo oooo
o
oooooooooo oooo ooo o o o
oo o o o ooooooooooo
ooooo
ooooooo oo o
oooooooo oooooo oo oooo
o o
oooooooooo oooo
o
oooo
o oo ooo oo oo o ooo ooooo o ooooo o oooo
ooooooooo
o o o oooo ooo
oo
ooooooo oo o
ooo ooo
ooo o o o
ooooooooo o o o o o
o
o o oo
ooo o o
oooo o o oooo ooooooo o oo oo o oo ooooo oooooooo oooooo o
oo ooooo oooooooooo oo o
ooo o o o oooo o oo oo o ooooooooo o o oo o
20
oooo ooo
o oooo oo oo
o ooooo ooooooooooooo ooo oo oo o o o o o oooo
ooooo o
oooo ooooooooo ooooo o
o oo ooooooooooooo o
o ooo o
oo ooooo
ooooo
oo ooooooo oo o oo oo
100 160 220 2 6 10 14 15 25 35 45 20 40 60
FIGURE 4.12. A scatterplot matrix of the South African heart disease data.
Each plot shows a pair ofJoan
risk Bruna
factors, and STAT
the cases and
135: controls
Linear are color coded
Regression
Example: South African Heart disease
β̂ std(
c β̂)
(intercept) -4.13 0.964
sbp 0.006 0.006
tobacco 0.080 0.026
ldl 0.185 0.057
famhist 0.939 0.225
obesity -0.035 0.029
alcohol 0.001 0.004
age 0.043 0.010
(results from Hastie&Tibshirani)
βˆk
.
c βˆk )
std(
βˆk
.
c βˆk )
std(
βˆk − βk
∼ N (0, 1) .
c βˆk )
std(
βˆk
.
c βˆk )
std(
βˆk − βk
∼ N (0, 1) .
c βˆk )
std(
The Wald Test tests the null hypothesis βk = 0. Reject the null
hypothesis if
|βˆk |
Zk = ≥ z(1 − α/2) .
c βˆk )
std(
β̂ std(
c β̂) Z -score
(intercept) -4.13 0.964 -4.28
sbp 0.006 0.006 1.023
tobacco 0.080 0.026 3.034
ldl 0.185 0.057 3.22
famhist 0.939 0.225 4.178
obesity -0.035 0.029 -1.18
alcohol 0.001 0.004 0.136
age 0.043 0.010 4.184
(results from Hastie&Tibshirani)
β̂ std(
c β̂) Z -score
(intercept) -4.13 0.964 -4.28
sbp 0.006 0.006 1.023
tobacco 0.080 0.026 3.034
ldl 0.185 0.057 3.22
famhist 0.939 0.225 4.178
obesity -0.035 0.029 -1.18
alcohol 0.001 0.004 0.136
age 0.043 0.010 4.184
(results from Hastie&Tibshirani)