PCA Machine Learning
PCA Machine Learning
Cedric Flamant
1
Outline
1. Introduction:
a. Why Dimensionality Reduction?
b. Linear Algebra (Recap).
c. Statistics (Recap).
To find a more meaningful basis to express our data filtering the noise
and revealing the hidden structure.
6
Symmetric matrices
Consider a design (or data) matrix consists of n observations and p
predictors:
Similar for
such that:
(orthogonal)
(normalized)
➢ Hence, they form an orthonormal basis.
Spectrum:
Orthogonal Matrix:
Eigen-decomposition:
Similar for
Same eigenvalues.
Transformed eigenvectors:
13
Centered Model Matrix
16
PCA
The data with the greatest variance lie on the first axis
(first principal component) and so on.
PCA finds:
Hence, PCA indicates that there may be fewer variables that are essentially
responsible for the variability of the response.
positive
negative
vs
When the data lie on a nonlinear manifold in the predictor space, then
linear methods are likely to be ineffective.
CS109A, PROTOPAPAS, RADER, 31
TANNER
Kernel PCA for Nonlinear Dimensionality Reduction
Thank you