Principal Component Analysis
Principal Component Analysis
Principal Component Analysis
max
C
( )
= max
w: w =1
w
T
Cw
= w
*T
Cw
*
Orthogonal decomposition
Total variance is conserved
Maximizing parallel variance =
Minimizing perpendicular variance
Rubber band computer
Aw= qyx
Aw= q y y
( )
x x
( )
Correlation/covariance rule
presynaptic activity x
postsynaptic activity y
Hebbian
Stochastic gradient ascent
Assume data has zero mean
Linear perceptron
y = w
T
x
Aw= qxx
T
w
E =
1
2
w
T
Cw
C = xx
T
Aw = q
cE
cw
Preventing divergence
Bound constraints
Ojas rule
Aw= q yx y
2
w
( )
Converges to principal component
Normalized to unit vector
Multiple principal components
Project out principal component
Find principal component of remaining
variance
clustering vs. PCA
Hebb: output x input
binary output
first-order statistics
linear output
second-order statistics
Data vectors
x
a
means ath data vector
ath column of the matrix X.
X
ia
means matrix element X
ia
ith component of x
a
x is a generic data vector
x
i
means ith component of x
Correlation matrix
Generalization of second moment
x
i
x
j
=
1
m
X
ia
a=1
m
X
ja
xx
T
=
1
m
XX
T