Support Vector Machines (II) : CMSC 422
Support Vector Machines (II) : CMSC 422
Machines (II)
CMSC 422
MARINE CARPUAT
[email protected]
REVIEW
The Maximum Margin Principle
• Find the hyperplane with maximum
separation margin on the training data
Support Vector Machine (SVM)
Characterizing the margin
Let’s assume the entire training data is correctly classified
by (w,b) that achieve the maximum margin
Solving the SVM Optimization Problem
(assuming linearly separable data)
Solving the SVM Optimization Problem
(assuming linearly separable data)
Note
• Given 𝛼 the solution for w, b has the same form as in the
separable case
• 𝛼 is again sparse, nonzero 𝛼𝑛 ’s correspond to support vectors
Support Vectors
in the Non-Separable Case
We now have 3 types of support vectors!
(1)
(2)
(3)
Notes on training
• Solving the quadratic problem is O(N^3)
– Can be prohibitive for large datasets
Note
• Kernelized SVM needs the
support vectors at test time!
• While unkernelized SVM can
just store w
Example: decision boundary of an
SVM with an RBF Kernel
What you should know
• What are Support Vector Machines
• How to train SVMs
– Which optimization problem we need to solve
• Geometric interpretation
- What are support vectors and what is their
relationship with parameters w,b?
• How do SVM relate to the general formulation of
linear classifiers
• Why/how can SVMs be kernelized