Chapter 3
Chapter 3
SUPERVISED LEARNING
NETWORK
DEFINITION OF SUPERVISED LEARNING
NETWORKS
Training and test data sets
x1 w1
w0
w2
x2 o
. n
wi xi
.
. wn i=0
n
xn 1 if w xi >0
{ -1 otherwise
f(xi)= i=0
i
PERCEPTRON LEARNING
wi = wi + wi
wi = (t - o) xi
where
t = c(x) is the target value,
o is the perceptron output,
Is a small constant (e.g., 0.1) called learning rate.
If the output is correct (t = o) the weights wi are not
changed
Output Values
Output Layer
Adjustable
Weights
Input Layer
Input Signals (External Stimuli)
LAYERS IN NEURAL NETWORK
The input layer:
• Introduces input values into the network.
• No activation function or other processing.
Adaline network uses Delta Learning Rule. This rule is also called
as Widrow Learning Rule or Least Mean Square Rule. The delta
rule for adjusting the weights is given as (i = 1 to n):
USING ADALINE NETWORKS
Initialize
Initialize • Assign random weights to all links
Training
• Feed-in known inputs in random
sequence
• Simulate the network
Training • Compute error between the input and
the output (Error Function)
• Adjust weights (Learning Function)
• Repeat until total error < ε
Thinking Thinking
• Simulate the network
• Network will respond to any input
• Does not guarantee a correct solution
even for trained inputs
MADALINE NETWORK
Inputs
Hiddens
I0
Outputs
h0
I1 o0
h1
I2 o1
h2 Outputs
I3 Hiddens
Inputs
MULTILAYER FEEDFORWARD NETWORK:
ACTIVATION AND TRAINING
For feed forward networks:
• A continuous function can be
• differentiated allowing
• gradient-descent.
• Back propagation is an example of a gradient-descent
technique.
• Uses sigmoid (binary or bipolar) activation function.
In multilayer networks, the activation function is
usually more complex than just a threshold function,
like 1/[1+exp(-x)] or even 2/[1+exp(-x)] – 1 to allow for
inhibition, etc.
GRADIENT DESCENT
Gradient-Descent(training_examples, )
Image processing.
Signature verification.
Bioinformatics.
Perceptron,
Adaline,
Madaline,
Backpropagation Network,
Radial Basis Function Network.