Supervised Learning Network
Supervised Learning Network
NETWORK
DEFINITION OF SUPERVISED LEARNING
NETWORKS
Training and test data sets
x1 w1
w0
w2
x2 o
. n
wi xi
.
. wn i=0
n
xn 1 if w xi >0
{ -1 otherwise
f(xi)= i=0
i
PERCEPTRON LEARNING
wi = wi + wi
wi = (t - o) xi
where
t = c(x) is the target value,
o is the perceptron output,
Is a small constant (e.g., 0.1) called learning rate.
If the output is correct (t = o) the weights wi are not
changed
For
each If y!=t
s:t
Output Values
Output Layer
Adjustable
Weights
Input Layer
Input Signals (External Stimuli)
LAYERS IN NEURAL NETWORK
The input layer:
• Introduces input values into the network.
• No activation function or other processing.
Y
Initialize weights and
bias and α
If Ei=Es
Calculate error
Ei=Σ(t-yin)2
For
each wi (new) wi (old ) (t yin ) xi
s:t b(new) b(old ) (t yin )
Adaline network uses Delta Learning Rule. This rule is also called
as Widrow Learning Rule or Least Mean Square Rule. The delta
rule for adjusting the weights is given as (i = 1 to n):
USING ADALINE NETWORKS
Initialize
Initialize • Assign random weights to all links
Training
• Feed-in known inputs in random
sequence
Training • Simulate the network
• Compute error between the input and
the output (Error Function)
• Adjust weights (Learning Function)
• Repeat until total error < ε
Thinking
Thinking
• Simulate the network
• Network will respond to any input
• Does not guarantee a correct solution
even for trained inputs
MADALINE NETWORK
Inputs
Hiddens
I0
Outputs
h0
I1 o0
h1
I2 o1
h2 Outputs
I3 Hiddens
Inputs
MULTILAYER FEEDFORWARD NETWORK:
ACTIVATION AND TRAINING
For feed forward networks:
• A continuous function can be
• differentiated allowing
• gradient-descent.
• Back propagation is an example of a gradient-descent
technique.
• Uses sigmoid (binary or bipolar) activation function.
In multilayer networks, the activation function is
usually more complex than just a threshold function,
like 1/[1+exp(-x)] or even 2/[1+exp(-x)] – 1 to allow for
inhibition, etc.
Image processing.
Signature verification.
Bioinformatics.