Spectral Clustering 2
Spectral Clustering 2
matrix
matrix
Spectral Clustering
Empirically very successful
Authors disagree:
Which eigenvectors to use
How to derive clusters from these
eigenvectors
How the choice of value affects the
results?
What would be the optimal choice for ?
Example 2 (not so simple)
Example 2 cont-d I
Example 2 cont-d II
Example 2 cont-d III
Example 2 cont-d IV
Spectral Clustering
Algorithm Ng, Jordan, and
Weiss
Motivation
Given a set of points
S s1 ,..., sn R l
Let X
(d) =X(1:d, :) be top d rows of X: the d
principal eigenvectors scaled by the square
root of the corresponding eigenvalue
A =X TX
d (d) (d) is the best rank-d approximation
to A with respect to Frobenius norm (||A||
F
2
= a ij )
2
The Polarization Theorem
II
Build Y(d) by normalizing the columns of X(d)
to unit length
Let ij be the angle btw xi,xj – columns of
X(d)
Claim
As A is projected to successively lower
ranks A(N-1), A(N-2), … , A(d), … , A(2), A(1), the
sum of squared angle-cosines (cos ij)2 is
strictly increasing
Brand-Huang algorithm
Basic strategy: two alternating
projections:
Projection to low-rank
Projection to the set of zero-
diagonal doubly stochastic matrices
(all rows and columns sum to unity)
stochastic matrix has all rows and
columns sum to unity
Brand-Huang algorithm II
While {number of EV=1}<2 do
APA(d)PA(d) …
Projection is done by suppressing the negative
eigenvalues and unity eigenvalue.