Issues in Backpropagation Algorithm
Issues in Backpropagation Algorithm
Algorithm
Common heuristic to attempt to
alleviate the problem of local
minima include:
Representation Power of
FeedFarward Networks
Hypothesis Space Search and
Inductive Bias
Hidden Layer Representations
• One intriguing property of Backpropagation ability to discover useful
intermediate representation at the hidden unit layers inside the
network.
Generalization, Overfitting, and
Stopping Criterion
Advance Topics In ANN
Alternative Error Minimization
Procedure
c) Recurrent network
Dynamically Modify Network
Structure
Example: Face
Recognition
●
From Machine Learning by Tom M. Mitchell
● Input: 30 by 32 pictures of people with the
following properties:
Wearing eyeglasses or not
Facial expression: happy, sad, angry,
neutral
Direction in which they are looking: left, right, up,
● straight ahead
Output: Determine which category it fits into
for
one of these properties (we will talk about
direction)
Input
Encoding
●
Each pixel is an input
30*32 = 960 inputs
●
The value of the pixel (0 – 255) is linearly mapped
onto the range of reals between 0 and 1
Output
Encoding
●
Could use a single output node with the
classifications assigned to 4 values (e.g. 0.2, 0.4,
0.6, and 0.8)
●
Instead, use 4 output nodes (one for each value)
1-of-N output encoding
Provides more degrees of freedom to the network
●
Use values of 0.1 and 0.9 instead of 0 and 1
The sigmoid function can never reach 0 or 1!
● Example: (0.9, 0.1, 0.1, 0.1) = left, (0.1, 0.9, 0.1,
0.1) = right, etc.
Network structure
Inputs
3 Hidden Units
3 Hidden Units
x1
x2
. 3 Hidden Units
Outputs
.
.
x960
Other Parameters
•●training rate:η = 0.3 momentum: α = 0.3
•●Used full gradient descent (as opposed to stochastic)
•●Weights in the output units were initialized to
•small random variables, but input weights were initialized to 0