Love Report 1
Love Report 1
(2023-24)
SEMINAR REPORT
On
MACINE LEARNING
(PRINCIPAL COMPONENT ANALYSIS)
Submitted
By
NAME: LAHU FULSING AJMERA
ROLL NO: CSEC333
TABLE CONTENT
3
1 Introduction 4
4.1 Advantages 8
4.2 Limitations 8
5 Conclusion 9
6 References 9
ABSTRACT
4
1. INTRODUCTION
selected to form the new feature space. These components capture the most
significant variance in the data.
Consider a dataset with two features: height and weight. These features may be
correlated, meaning that an increase in height could be associated with an
increase in weight. PCA will identify a new axis (the first principal component)
that captures the greatest variance, often representing the linear combination of
height and weight. The second principal component will capture the remaining
variance, orthogonal to the first component.
This transformation allows us to reduce the data from two dimensions (height
and weight) to one dimension, while preserving as much of the original variance
as possible.
4.1 Advantages
Reduces Dimensionality: PCA reduces the number of features, making it easier
and faster to process large datasets.
Improves Model Performance: By removing irrelevant or redundant features,
PCA can enhance the performance of machine learning models, especially in
terms of generalization.
Noise Reduction: PCA can filter out noise from the data, resulting in more
robust models.
Uncorrelated Features: After PCA, the transformed features (principal
components) are uncorrelated, which simplifies subsequent modelling
techniques.
4.2 Limitations
Linear Assumption: PCA assumes that the relationships between features are
linear, which may not be the case in all datasets. Non-linear dimensionality
reduction techniques (e.g., t-SNE, autoencoders) may be more appropriate in
such cases.
Interpretability: While PCA reduces the dimensionality of the data, the new
features (principal components) may not have an intuitive interpretation,
making it harder to understand what the components represent.
Loss of Information: Although PCA preserves the most important information,
some data variation is inevitably lost, particularly when reducing to a very low
number of dimensions.
5. CONCLUSION
10
Machine learning has transformed the way we approach data analysis, enabling
more accurate predictions, better decision-making, and automated systems.
Principal Component Analysis (PCA) plays a vital role in machine learning,
particularly in tasks involving high-dimensional data. By reducing the number
of features, PCA helps mitigate the curse of dimensionality, enhances model
performance, and facilitates data visualization. While PCA is an effective tool
for dimensionality reduction, it is important to consider its assumptions and
limitations, and to choose the appropriate technique based on the nature of the
data and the problem at hand.
As machine learning continues to evolve, techniques like PCA will remain
essential for dealing with complex, high-dimensional datasets, ensuring that
machine learning models are both efficient and accurate.
6.REFERENCES