Lecture 15
Lecture 15
Lecture 15
Lecture 15
Recap: Clustering 2
Hierarchical clustering
Evaluation
Term vs. document space clustering
Multi-lingual docs
Feature selection
Labeling
Linear Algebra
Background
(right) eigenvector
eigenvalue
Matrix-vector multiplication
3 0 0
S 0 2 0
0 0 0
1
v1 0
0
0
v2 1
0
0
v3 0
1
Sx S (2v1 4v2 6v 3 )
Sx 2 Sv1 4 Sv2 6 Sv 3 21v1 42 v2 6 3v 3
w n , wT Sw 0, then if Sv v 0
Example
2 1
S
1 2
Let
Then
Real, symmetric.
2
S I
1
1
1
1
2
( 2 ) 1 0.
Eigen/diagonal Decomposition
Let
be a square matrix with mlinearly
independent eigenvectors (a non-defective
Uniqu
matrix)
e for
Theorem: Exists an eigen decomposition
diagonal
Diagonal elements of
are eigenvalues of
distinc
t
eigenvalues
... vn
...
2 1
S
; 1 1, 2 3.
1 2
1 and 1
The eigenvectors
1
1
Inverting, we haveU
1 / 2 1 / 2
1
/
2
1
/
2
1 1
formU
1
1
Recall
UU1 =1.
1 1 1 0 1 / 2 1 / 2
Then, S=U U1 =
1 1 0 3 1 / 2 1 / 2
Example continued
Lets divide U (and multiply U1) by 2
1 / 2 1 / 2 1 0 1 / 2
Then, S=
1 / 2 1 / 2 0 3 1 / 2
Q
1/ 2
1/ 2
(Q-1= QT )
If
is a symmetric matrix:
where Q is orthogonal:
Q-1= QT
(everything is real)
Exercise
0 1
1 0
0 1
1 0
1 2
2 3
2 2
2 4
Time out!
A UV
mm mn
V is nn
diag 1... r
Singular values.
SVD example
Let
1 1
A 0 1
1 0
0
1/ 2
1/ 2
2/ 6
1/ 6
1/ 6
1/ 3 1 0
1/ 2
1/ 3 0
3
1/ 2
1 / 3 0 0
1/ 2
1/ 2
Low-rank Approximation
Ak
min
A X
X :rank ( X ) k
Frobenius norm
Low-rank Approximation
Ak i 1 i ui viT
k
Approximation error
min
X :rank ( X ) k
A X
A Ak
k 1
Latent Semantic
Analysis via SVD
What it is
But why?
Various extensions
Document clustering
Relevance feedback (modifying query vector)
Geometric foundation
ring
jupiter
meaning 1
planet
...
space
voyager
saturn
...
meaning 2
contribution to similarity, if
used in 1st meaning, but not
if in 2nd
car
company
dodge
ford
Goals
qk q U k
T
1
k
Empirical evidence
Empirical evidence
Precision
250
0.367
300
0.371
346
0.374
Failure modes
Negated phrases
Boolean queries
Block 2
m
terms
0s
Block k
= non-zero entries.
0s
Block 2
m
terms
0s
Block k
Block 2
m
terms
0s
Block k
= non-zero entries.
Block 1
Block 2
Block k
Simplistic picture
Topic 1
Topic 2
Topic 3
For text, the terms are features and the docs are
objects.
Could be opinions and users more in 276B.
Can work with low-rank approximation.
If entries are missing (e.g., users opinions), can
recover if dimensionality is low.
Resources
http://www.cs.utk.edu/~berry/lsi++/
http://lsi.argreenhouse.com/lsi/LSIpapers.html
Dumais (1993) LSI meets TREC: A status report.
Dumais (1994) Latent Semantic Indexing (LSI) and
TREC-2.
Dumais (1995) Using LSI for information filtering:
TREC-3 experiments.
M. Berry, S. Dumais and G. O'Brien. Using linear
algebra for intelligent information retrieval. SIAM
Review, 37(4):573--595, 1995.