Feature Extraction: - Saheni Patra
Feature Extraction: - Saheni Patra
Feature Extraction: - Saheni Patra
-SAHENI PATRA
Feature vector
Important
terminologies
Applications of PCA
Standardization
Singular Value
Decomposition
WHAT IS FEATURE
EXTRACTION?
Feature extraction means are just select a particular features. For training purpose,we have to extract the features and send to those
features to particular machine as a input.
Feature extraction involves transforming high dimensional data into space of fewer dimensional. High dimensional data space
means all the the data sets of the featured dataset and lower dimensional dataset means we have to select only the relevant features.
There is one bike image and feature extraction algorithm means the different parts & different subsets which is called feature extraction.
TYPES OF METHOD
Here, we take some very complex dataset with lots of variables. Then we run it through the
PCA and we reduce the variables.
dimensionality
• Here, for our ease we can consider that we get two principle components PC1 & PC2.
• Comparing both the principl components,vee find that the data points are sufficiently spaced in the PC1.
• Whereas, in PC2 they are less spaced which makes the observation and further calculations much difficult. Therefore, we
accept the PC1 and not PC2 as the data point are more spaced.
IMPORTANT TERMINOLOGIES
HOW PCA WORKS?
The steps to perform PCA are following:
1. Standardize the data.
2. Compute the covariance matrix of the features from the dataset.
3. Perform eigendecompositon on the covariance matrix.
4. Order the eigenvectors in decreasing order based on the magnitude of their corresponding eigenvalues.
STANDARDIZATION
In this step, we will standardize our dataset. Such as in a particular column, the features with high variance are more
important compared to the features with lower variance.
If the importance of features is independent of the variance of the feature, then we will divide each data
item in a column with the standard deviation of the column. Here we will name the matrix as Z.
The process Involves the removal of mean from the variable values and scaling the data with respect to the
standard deviation.
COVARIANCE MATRIX COMPUTATION
Covariance matrix is used to express the correlation between any two or more attributes in a maultidimensional dataset.
• Positive covariance indicate that the value of one variable is directly proportional to other variable.
• Negetive covariance indicate that the value of one variable is inversely proportional to other variable.
Here we can see the covariance table for more than two attributes in a multidimensional dataset.
EIGEN VALUES & EIGEN VECTORS
Eigen values and eigen vectors are the mathematical values that are extracted from the covariance table. Now we
need to calculate the eigenvalues and eigenvectors for the resultant covariance matrix Z. Eigenvectors or the
covariance matrix are the directions of the axes with high information. And the coefficients of these eigenvectors
are defined as the eigenvalues.
Here we decide whether we must keep or disregard the less significant principal components that we have
generated in the above steps.
APPLICATION OF PCA IN MACHINE
LEARNING
• PCA is used to visualize multidimensional image.
• It is used to reduce the number of dimensions in healthcare data.
• PCA can help resize an image.
• PCA helps to find patterns in the high-dimensional datasets.
FACTOR ANALYSIS
Just like PCA, Factor Analysis is also a model that allows reducing information in a
larger number of variables into a smaller number of variables. In Factor Analysis we
call those “latent variables”.
Factor Analysis is based on a model called the common factor model. It starts from
the principle that there a certain number of factors in a data set, and that each of the
measured variables captures a part of one or more of those factors.
SINGULAR VALUE DECOMPOSITION
1. https://www.geeksforgeeks.org/ml-principle-component/analysispca/amp/
2. https://www.simplilearn.com/tutorials/machine-learning-tutorial/principal-c
omponent-analysis
3. https://www.machinelearningplus.com/machine-learning/principal-compone
nts-analysis-pca-better-explained/
4. https://slideplayer.com/slide/5256003/
THANK
YOU