MLP and Backpropagation
MLP and Backpropagation
x1
xn
xn
1st question:
what do the extra layers gain you? Start with looking at
what a single layer can’t do
3
Perceptron Learning Theorem
• Recap: A perceptron (threshold unit) can
learn anything that it can represent (i.e.
anything separable with a hyperplane)
4
The Exclusive OR problem
A Perceptron cannot represent Exclusive OR
since it is not linearly separable.
5
6
Minsky & Papert (1969) offered solution to XOR problem by
combining perceptron unit responses using a second layer of
Units. Piecewise linear classification using an MLP with
threshold (perceptron) units
+1
+1
7
Three-layer networks
x1
x2
Input
Output
xn
Hidden layers
8
Properties of architecture
• No connections within a layer
• No direct connections between input and output layers
• Fully connected between layers
• Often more than 3 layers
• Number of output units need not equal number of input units
• Number of hidden units per layer can be more or less than
input or output units
m
y
f
(
i
w
x
i
j
b
)
j
i
j
1