MLSP-6 Dimensionality Reduction
MLSP-6 Dimensionality Reduction
Processing (MLSP)
Dimensionality Reduction
Course Instructor
Prof. Jyotsna Singh
curse of dimensionality
• In practice, the curse of dimensionality means that, for a
given sample size, there is a maximum number of features
above which the performance of our classifier will degrade
rather than improve
In most cases, the additional information that is lost by
discarding some features is compensated by a more accurate
mapping in the lower dimensional space
• How do we beat the curse of dimensionality?
• By incorporating prior knowledge
• By providing increasing smoothness of the
target function
• By reducing the dimensionality
Dimensionality Reduction
• In pattern recognition, Dimension Reduction is
defined as-
• It is a process of converting a data set having
vast dimensions into a data set with lesser
dimensions.
• It ensures that the converted data set conveys
similar information concisely.
Dimensionality Reduction
Dimensionality Reduction
Signal representation versus classification
Linear Discriminant Analysis
• LDA is a dimensionality reduction technique used
as a pre-processing step for Pattern recognition or
Machine Learning Application.
• LDA is similar to PCA, But LDA in addition finds the
axis that maximises the separation between
multiple classes.
• It projects the N dimensional feature space onto a
smaller subspace M (M<=N), While maintaining
the class discriminatory Information.
Linear Discriminant Analysis
Linear Discriminant Analysis, two-classes
Fisher linear discriminant
Linear Discriminant Analysis, two-classes
Linear Discriminant Analysis, two-classes
Fisher LDA for C classes
Fisher LDA for C classes
Fisher LDA for C classes
PCA Algorithm
• Step-01:
• Step-05:
• Calculate the eigen values and eigen vectors of
the covariance matrix.
• λ is an eigen value for a matrix M if it is a
solution of the characteristic equation |M – λI| =
0.
• So, we have-
• From here,
• (2.92 – λ)(5.67 – λ) – (3.67 x 3.67) = 0
• 16.56 – 2.92λ – 5.67λ + λ2 – 13.47 = 0
• λ2 – 8.59λ + 3.09 = 0
• Solving this quadratic equation, we get λ = 8.22, 0.38
• Thus, two eigen values are λ1 = 8.22 and λ2 = 0.38.
•
• Clearly, the second eigen value is very small compared to
the first eigen value.
• So, the second eigen vector can be left out.
•
• Eigen vector corresponding to the greatest eigen value is
the principal component for the given data set.
• So. we find the eigen vector corresponding to eigen value
λ1.
• We use the following equation to find the eigen
vector-
• MX = λX
• where-
• M = Covariance Matrix
• X = Eigen vector
• λ = Eigen value
•
• Substituting the values in the above equation, we
get-
• Solving these, we get-
• 2.92X1 + 3.67X2 = 8.22X1
• 3.67X1 + 5.67X2 = 8.22X2
• On simplification, we get-
• 5.3X1 = 3.67X2 ………(1)
• 3.67X1 = 2.55X2 ………(2)