Perceptron PDF
Perceptron PDF
1. Feed-Forward
Network
2. Recurrent
Network
Network Structure
Two main network structures
1. Feed-Forward
Network
2. Recurrent
Network
Learning Paradigms
Supervised Learning:
Given training data consisting of pairs of
inputs/outputs, find a function which correctly
matches them
Unsupervised Learning:
Given a data set, the network finds patterns and
categorizes into data groups.
Reinforcement Learning:
No data given. Agent interacts with the
environment calculating cost of actions.
Simple Perceptron
The perceptron is a single layer feed-forward
neural network.
Simple Perceptron
Simplest output function
= 0.2
0
w= 1
0.5
0 = w 0w 1 x1w 2 x 2
= 0x 10.5x 2
x 2 = 2x1
Learning Example
= 0.2
0
w= 1
0.5
x1 = 1, x2 = 1
wTx > 0
Correct classification,
no action
Learning Example
= 0.2
0
w= 1
0.5
x1 = 2, x2 = -2
w 0 = w 00.21
w 1 = w 10.22
w 2 = w 20.22
Learning Example
= 0.2
0.2
w = 0.6
0.9
x1 = 2, x2 = -2
w 0 = w 00.21
w 1 = w 10.22
w 2 = w 20.22
Learning Example
= 0.2
0.2
w = 0.6
0.9
x1 = -1, x2 = -1.5
wTx < 0
Correct classification,
no action
Learning Example
= 0.2
0.2
w = 0.6
0.9
x1 = -2, x2 = -1
wTx < 0
Correct classification,
no action
Learning Example
= 0.2
0.2
w = 0.6
0.9
x1 = -2, x2 = 1
w 0 = w 00.21
w 1 = w 10.22
w 2 = w 20.21
Learning Example
= 0.2
0
w = 0.2
1.1
x1 = -2, x2 = 1
w 0 = w 00.21
w 1 = w 10.22
w 2 = w 20.21
Learning Example
= 0.2
0
w = 0.2
1.1
x1 = 1.5, x2 = -0.5
w 0 = w 00.21
w 1 = w 10.21.5
w 2 = w 20.20.5
Learning Example
= 0.2
0.2
w = 0.5
1
x1 = 1.5, x2 = -0.5
w 0 = w 00.21
w 1 = w 10.21.5
w 2 = w 20.20.5
Perceptron Convergence Theorem
The theorem states that for any data set which is
linearly separable, the perceptron learning rule is
guaranteed to find a solution in a finite number of
iterations.
=x k ...x 1w 0
.
Perceptron Convergence Theorem
As we assume linear separability, a solution w*
where wx(k) > 0, x(1)...x(k) T1. Multiply both
sides by the solution w* to get
w w k 1 = w x 1...w x k
j=1
k 2
= max x j
Perceptron Convergence Theorem
But now we have a conflict between the
equations, for sufficiently large values of k
2 2
2 2 k
w k 1 k w k 1 2
w
So, we can state that k cannot be larger than
some value kmax for which the two equations are
both satisfied.
2 2 2
k max w
k max = 2
k max = 2
w
Perceptron Convergence Theorem
Thus it is proved that for k = 1, k, w(0) = 0,
given that a solution vector w* exists, the
perceptron learning rule will terminate after at
most kmax iterations.
The End