Eigenvectors 2
Eigenvectors 2
9- Dimensionality
Reduction
Computer and Information Science
University of Massachusetts Dartmouth
Attribute Dimensions and Orders
• Dimensions
• 1D: scalar
• 2D: two-dimensional vector
• 3D: three-dimensional
vector
• >3D: multi-dimensional
vector
• Orders 1-st order 2-nd order
• scalars
• vectors
• matrix
• tensors (high-order) vector matrix
Bivariate Data Representations
• Parametric Model
• Use a parameterized family of probability distributions to describe the
nature of a set of data (Moghaddam & Pentland, 1997).
• The data distribution is empirically assumed or estimated.
• Learning is conducted by measuring a set of fixed parameters, such as
mean and variance.
• Effective for the large sample but degrade for complicated data
distribution.
• Nonparametric Model
• Distribution free.
• Learning is conducted by measuring the pair-wise data relationship in
both global and local manners.
• Effective and robust due to the reliance on fewer assumptions and
parameters.
• Work for cases with small-sample, high-dimensionality, and complicated
data distribution.
Linear Models
• Two representative models
• Principal Component Analysis (PCA)
• Linear Discriminant Analysis (LDA)
• PCA is trying to captures the “principal” variations in the
data
• It is computed by finding the Eigenvectors of the
covariance matrix of the data
• Geometrically, PCA finds the largest variations
directions of the underlying data
• LDA on the other hand considers the label information
which maximizes the distance between classes, and
minimizes the distance within a class
Courtesy of Prof. Zhu Li, Hong Kong Polytechnic University.
Principal Component Analysis
• Two views: Assuming zero
• Variance mean, line is
represented
• Reconstruction as , where
w is the basis,
s.t. =1.
• Maximize the data
variance in the lower-
dimensional space
• Objective:
• Find a subspace spanned by to
minimize the reconstruction error
• We add an additional constraint to
make this problem trackable:
Face reconstruction using different
number of eigenvectors
Principal Component Analysis
Eigenvector Eigenfaced
Courtesy of Prof. Zhu Li, Hong Kong Polytechnic University.
Linear Discriminant Analysis
• Instead of PCA, it finds the discriminant subspace by including class label info in
subspace modeling (Supervised learning).
– Compute within class scatter
– Compute between class scatter
– Maximize between scatters and minimize within scatters
•
LDA
𝑥𝑗
𝜇𝑖 𝑢𝑗
Within-Class Scatter Matrix
• For all samples from Class , we add them
together
Within-class scatter
matrix of class-i
• If we have more than one classes (most likely!),
the within-class scatter matrix is:
Between-Class Scatter Matrix
• Only Two Classes
𝑇
( ( 𝜇1 − 𝜇2 ) ( 𝜇1 − 𝜇 2) ) 𝑤=𝑤 𝑆𝑏 𝑤
2
( 𝑤 ( 𝜇1 −𝜇 2) ) =𝑤
𝑇 𝑇 𝑇
: Between-class
• More than Two Classes scatter matrix
Learning Objective
• To achieve two goals, we will
PCA LDA
Courtesy of Prof. Zhu Li, Hong Kong Polytechnic University.
PCA vs. LDA
• PCA performs
worse under
this condition
• LDA (FLD-Fisher
Linear Discriminant)
provides
better low
dimensional
representation.
http://en.wikipedia.org/wiki/Manifold
Manifold Learning
Swiss Roll
Dimensionality
Reduction
http://www.cs.toronto.edu/~roweis/lle/
LEA for Pose Manifold
Yun Fu, et. al. “Locally Adaptive Subspace and Similarity Metric Learning for Visual Clustering and Retrieval”, CVIU, Vol. 110, No. 3, pp: 390-402, 2008.
LEA for Expression Manifold
Yun Fu, et. al. “Locally Adaptive Subspace and Similarity Metric Learning for Visual Clustering and Retrieval”, CVIU, Vol. 110, No. 3, pp: 390-402, 2008.