Machine Unit4
Machine Unit4
org/feature-selection-techniques-in-machine-learning/
feature extraction
Process:
Process:
3. Kernel PCA
Kernel Principal Component Analysis (Kernel PCA) is an extension of
Principal Component Analysis (PCA) that enables the technique to handle
non-linear data. While traditional PCA identifies the principal components
(the directions of maximum variance) in the original feature space, Kernel
PCA first maps the data into a higher-dimensional space using a non-linear
function and then performs PCA in this new space. This allows Kernel PCA
to capture the complex structures and relationships within the data that
linear PCA might miss.
Key Concepts
Process:
Process:
Process
Process:
7. Factor analysis
. Factor Analysis (FA) is a statistical method used for identifying the underlying
relationships between observed variables. It is commonly used for dimensionality
reduction and fea+ture extraction in machine learning. FA assumes that the
observed variables are influenced by a smaller number of unobserved variables
called factors. These factors can be thought of as the underlying structure that
explains the patterns of correlations within the observed data.
Purpose:
Key Concepts:
https://www.geeksforgeeks.org/ml-content-based-recommender-system/
https://www.geeksforgeeks.org/collaborative-filtering-ml/
In machine learning, particularly in Support Vector Machines (SVMs), a hyperplane is a crucial
concept used for classification tasks. To understand it better, let's break it down:
A hyperplane is a subspace whose dimension is one less than that of its ambient space. In simpler
terms:
- In a 2-dimensional space (like a piece of paper), a hyperplane is a line.
- in higher dimensions, it's a bit harder to visualize, but the idea extends
similarly: in an n-dimensional space, a hyperplane is an (𝑛−1)-dimensional flat
affine subspace.
Support Vector Machines are supervised learning models used for classification and regression
analysis. In the context of binary classification, the goal of an SVM is to find the hyperplane that best
separates the data into two classes.
1. **Maximal Margin**:
- SVM aims to find the hyperplane that maximizes the margin between the two classes. The margin
is defined as the distance between the hyperplane and the nearest points from either class, which
are called support vectors.
- By maximizing this margin, the SVM ensures that the classifier is not only accurate but also robust
to new data points, reducing the risk of overfitting.
2. **Support Vectors**:
- Support vectors are the data points that lie closest to the hyperplane and are most difficult to
classify. These points are crucial because they determine the position and orientation of the
hyperplane.
- The optimal hyperplane is the one that is equidistant from the support vectors of each class.
1. Mathematical Formulation:
𝑤⋅𝑥+𝑏=0w⋅x+b=0
Here, 𝑤w is the weight vector (normal to the hyperplane), 𝑥x is the
input feature vector, and 𝑏b is the bias term.
For classification, the decision rule is:
Class=sign(𝑤⋅𝑥+𝑏)Class=sign(w⋅x+b)
This means that a point is classified based on which side of the
hyperplane it falls on.
- For classification, the decision rule is:
In a 2-dimensional space (2D), imagine plotting the data points on a graph. If you have two classes,
say Class A and Class B, the SVM will find a line (hyperplane) that best separates these two classes.
Ideally, this line will be placed in such a way that the distance to the nearest points (support vectors)
from both classes is maximized.
For more complex datasets where a linear hyperplane cannot separate the classes effectively, SVMs
use a technique called the **kernel trick**. Kernels transform the data into a higher-dimensional
space where a hyperplane can then separate the classes. Common kernels include polynomial
kernels, radial basis function (RBF) kernels, and others.
### Summary
In summary, the hyperplane in SVMs is a critical concept that defines the decision boundary
between classes in a classification problem. The primary goal is to find the hyperplane that
maximizes the margin, ensuring a robust and accurate classification model. This is achieved by
focusing on the support vectors and potentially using kernel methods for non-linear separation.
Convergence means two random variables are present and difference between their probability is
less means two variables value are match with each other then we called as convergence.
Logistic regression:
https://www.geeksforgeeks.org/understanding-logistic-regression/
clustering:::
https://www.javatpoint.com/clustering-in-machine-learning
Agglomerative