13.1 Support Vector Machine
13.1 Support Vector Machine
Age
2
Classification of customers
Income
Age
3
Classification of customers
Income
Age
4
Classification of customers
Income
Age
5
Classification of customers
Income
Age
6
Support vector machine
• Support vector machine (SVM) is a method for the classification of both linear and
nonlinear data.
• It uses a nonlinear mapping to transform the original training data into a higher
dimension. Within this new dimension, it searches for the linear optimal separating
hyperplane: a “decision boundary” separating the tuples of one class from another.
Non-linear SVM
• Datasets that are linearly separable work out great:
0 x
0 x
0 x
Sec. 15.2.3
Feature spaces
General idea: the original feature space can always be mapped to
some higher-dimensional feature space where the training set is
separable:
Φ: x → φ(x)
Sec. 15.2.3
Kernels
• Why use kernels?
• Make non-separable problem separable.
• Map data into better representational space
• Common kernels
Multiclass classification
One-versus-all/one-versus-rest
• Given m classes, we train m binary classifiers, one for each class.
• Classifier j is trained using data of class j as the positive class, and the
remaining data as the negative class. It learns to return a positive
value for class j and a negative value for the rest.
All-versus-all/one-versus-one
𝒎𝒎(𝒎𝒎−𝟏𝟏)
• Given m classes, we construct binary classifiers
𝟐𝟐
combination nCr, where n=m classes and r=2.
• A classifier is trained using data of the two classes.