Machine Learning Introduction
Machine Learning Introduction
---------------------------------------------------------------------------
Human or Machine?
The Chinese Room Argument
Information Processing?
Topics in AI
What we already achieved in AI?
• Board games – Chess, Checkers, Go, etc
• Solving Puzzles – Sudoku, etc
• Route finding in a map
• Image/speech enhancement
– Creating high resolution images, noise
suppression, …
•
Machine Learning?
• ML is a subset of AI.
• We focus on mathematical/algorithmic
aspects of learning which can be programmed
over a machine.
• We look into various learning paradigms.
• Prerequisites of the course are
– Probability theory
– Linear algebra
– Calculus
Evaluation/Examinations
DOG
This picture as it is
may not be in the Child has done more than
training set just remembering
What is learning (pattern recognition)?
• Child has learnt what is it that is common
among dogs … and, what is it that is common
among cats… also, what are the distinguishing
features/attributes.
• Child has learnt the pattern (regularity) behind
all dogs and the pattern behind all cats.
• Child then recognized a test image as having a
particular pattern that is unique to dogs.
Basic concepts
Object
x1 Feature vector X
- A vector of observations (measurements).
x
2 X
X - X is a point in feature space .
xd
Task
- To design a classifer (decision rule) f : Y
which decides about the class label based on X.
An example
is a set of persons
Object
Feature vector
h - A vector of observations (height, weight).
Person w
Task
- To design a classifer (decision rule) f : Y
- given height and weight of a person, classify him/her.
Feature extraction
Task: to extract features which are good for classification.
Good features: • Objects from the same class have similar feature values.
• Objects from different classes have different values.
Normal
persons
weight(w) Overweight
persons
height(h)
f(x1,x2) = w1x1+w2x2+b = 0
Class 1
x1
Perceptron
• Perceptron is the name given to the linear
classifier.
• If there exists a Perceptron that correctly
classifies all training examples, then we say
that the training set is linearly separable.
• In 1960s Rosenblatt gave an algorithm for
Perceptron learning for linearly separable
data.
Perceptron
• For linearly separable data many classifiers are
possible.
All being doing equally good
on training set, which one is
Class 2
good on the unseen test set?
Class 1
Maximizing the Margin SVM
Var1 IDEA : Select the
separating
hyperplane that
maximizes the
margin!
Margin
Width
Margin
Width
Var2
Artificial Neural Networks
Generative Models
• Bayes
– Naïve Bayes
• Graphical models
– Belief networks
Remember…
• No classifier is inherently better than any
other: you need to make assumptions to
generalize
Slide
Slide
credit:
credit:
D. D.
Hoiem
Hoiem
Generalization
Then we find (learn) a set of basis faces which best represent the differences
between them
That is, apply PCA and choose top Eigen vectors (Eigen faces)
We can then store each face as a set of weights for those basis faces
Eigenfaces: the idea
• Think of a face as being a weighted combination of some “component” or
“basis” faces
…
Eigenfaces: representing faces
• These basis faces can be differently weighted to represent any face
-8029 -1183 2900 -2088 1751 -4336 1445 -669 4238 -4221 6193 10549
SOME CHALLENGES …
• King – man + woman = Queen
• Face – emotion + surprise = Surprised face
• Interpretable models.
THANK YOU