Exp9 MLAI2
Exp9 MLAI2
09
Aim:- To apply dimensionality reduction techniques (e.g., PCA , LDA or ICA) on a dataset
and observe the performance of machine learning models.
Theory:-
Dimensionality Reduction:
Dimensionality reduction is the process of reducing the number of input variables or features in a
dataset. This is essential in machine learning to remove redundant data, minimize overfitting, and
improve model performance.
Common methods include PCA, LDA, and ICA.
1. Principal Component Analysis (PCA):
o PCA is an unsupervised learning algorithm that reduces the dimensionality of the
data by finding the directions (principal components) that maximize the variance in
the dataset.
o It transforms the data into a new coordinate system where the greatest variance by
any projection of the data comes to lie on the first principal component, the second
greatest variance on the second component, and so on.
o Applications: Data visualization, noise reduction, feature extraction.
2. Linear Discriminant Analysis (LDA):
o LDA is a supervised learning algorithm that maximizes the distance between
different class labels and minimizes the variance within the same class.
o While PCA focuses on finding directions of maximum variance in the entire
dataset, LDA focuses on maximizing the separability between known categories or
labels.
o Applications: Classification, reducing dimensions in labeled data.
3. Independent Component Analysis (ICA):
o ICA is used to transform data into statistically independent components, which is
helpful in scenarios where the underlying signals are non-Gaussian.
o Unlike PCA, which maximizes variance, ICA looks for independent sources of
variation and is often used for blind source separation.
o Applications: Blind source separation (e.g., separating mixed sound signals),
feature extraction.
Code: