SVM Fully Translated Fixed
SVM Fully Translated Fixed
DEFINITION
SVM (Support Vector Machine) is a supervised learning algorithm that can be used for classification
or regression problems.
This algorithm helps find an optimal hyperplane to linearly separate data into two different classes.
SVM is a suitable method for classification problems with large feature spaces where the objects to
be classified are
In a classification problem, the data will be divided into two classes: a positive label class and a
1. Positive Class:
Data points in this class are those identified by the SVM algorithm as belonging to a specific group
Example: In an email classification problem, the positive class could represent spam emails.
2. Negative Class:
Data points in this class are the opposite of the positive class, meaning they are in the opposing
group.
Example: In the email classification problem, the negative class could represent non-spam (regular)
emails.
PRINCIPLE OF SVM
(image illustrating how Support Vector Machines (SVM) work in a two-class classification problem,
Data groups:
- Yellow triangles: Represent data belonging to one class (possibly the negative class).
- Blue diamonds: Represent data belonging to the other class (possibly the positive class).
Hyperplanes:
- Positive Hyperplane: This is the line parallel to the optimal hyperplane and closest to the points of
It limits the distance from the positive points to the optimal hyperplane.
- Negative Hyperplane: This is the line parallel to the optimal hyperplane and closest to the points of
It limits the distance from the negative points to the optimal hyperplane.
Optimal Hyperplane:
This is the central line that divides the data into two groups such that the distance between this
data points of both groups is maximized. The goal of the SVM algorithm is to find the hyperplane
Maximum Margin:
Maximum Margin is the distance between the positive and negative hyperplanes. The larger this
Support Vectors:
Support Vectors are the data points closest to the optimal hyperplane, and these points determine
hyperplane. In the image, the points on the positive and negative hyperplanes are called Support
KERNEL FUNCTION
In cases where data cannot be separated by a hyperplane, kernel functions are used to transform
Advantages of SVM:
Efficient in high-dimensional spaces: Particularly useful for problems with large numbers of
Memory-efficient: Only stores important points (support vectors) necessary for training and
classification.
Flexible with Kernels: SVM can handle both linear and nonlinear problems through the use of
different kernels.
Disadvantages of SVM:
Does not provide probability: SVM only classifies data without providing a probability score.
CONCLUSION
SVM is a powerful classification method, particularly useful in image processing, text classification,
thanks to its flexible use of kernel functions. However, it has limitations when dealing with very