Unsupervised Learning Networks
Unsupervised Learning Networks
UNSUPERVISED LEARNING
NETWORKS
April 2007 1
UNSUPERVISED LEARNING
No help from the outside.
Learning by doing.
April 2007 2
FEW UNSUPERVISED
LEARNING NETWORKS
There exists several networks under this category, such as
Max Net,
Mexican Hat,
Kohonen Self-organizing Feature Maps,
Learning Vector Quantization,
Counterpropagation Networks,
Hamming Network,
Adaptive Resonance Theory.
April 2007 3
COMPETITIVE LEARNING
Output units compete, so that eventually only
one neuron (the one with the most input) is active
in response to each output pattern.
April 2007 4
MAX NET
Max Net is a fixed weight competitive net.
Max Net serves as a subnet for picking the node whose
input is larger. All the nodes present in this subnet are fully
interconnected and there exist symmetrical weights in all
these weighted interconnections.
The weights between the neurons are inhibitory and
fixed.
The architecture of this net is as shown below:
April 2007 5
April 2007 6
MEXICAN HAT NETWORK
Kohonen developed the Mexican hat network which is a more
generalized contrast enhancement network compared to the
earlier Max Net.
April 2007 7
MEXICAN HAT FUNCTION OF
LATERAL CONNECTION
April 2007 8
MEXICAN HAT NETWORK
The lateral connections are used to create a
competition between neurons. The neuron with the
largest activation level among all neurons in the
output layer becomes the winner. This neuron is the
only neuron that produces an output signal. The
activity of all other neurons is suppressed in the
competition.
April 2007 11
• Let x, y be two bipolar vectors
• Use x.y = a-d,
a-no. Of components in which
vectors agree and d- no. Of
components in which vectors
disagree.
• n-total no. Of components
• Now Hamming distance=a-d
• n = a+d or d= n-a
April 2007 12
• Since x.y = a-d
• x.y = a-(n-a)
= 2a-n
2a = x.y+n
a =1/2(x.y) + ½(n)
• So weights can be set to one-half the
exemplar vector
• Bias can be set to n/2 initially
• Calculating the unit with largest net
input, net can locate the unit closest
to
April exemplar vector
2007
ARCHITECTURE OF HAMMING NET
April 2007 14
The Hamming network consists of two layers.
• The first layer computes the difference
between the total number of
components and Hamming distance
between the input vector x and the
stored pattern of vectors in the
feedforward path.
• Layered structure.
April 2007 17
SELF-ORGANIZING FEATURE MAP
April 2007 18
SELF-ORGANIZING NETWORKS
Discover significant patterns or features in
the input data.
April 2007 19
KOHONEN SELF-ORGANIZING
FEATURE MAP (KSOFM)
The Kohonen model provides a topological mapping.
April 2007 21
COMPETITION OF KSOFM
Each neuron in an SOM is
assigned a weight vector with the
same dimensionality N as the input
space.
April 2007 22
CO-OPERATION OF KSOFM
The activation of the winning neuron is spread to
neurons in its immediate neighborhood.
This allows topologically close neurons to become
sensitive to similar patterns.
April 2007 25
KSOFM ALGORITHM
April 2007 26
KSOFM ALGORITHM
April 2007 27
EXAMPLE OF KSOFM
April 2007 28
EXAMPLE OF KSOFM
Find the winning neuron using the Euclidean distance:
April 2007 29
EXAMPLE OF KSOFM
The updated weight vector W3 at iteration (p+1) is
determined as:
April 2007 30
LEARNING VECTOR QUANTIZATION
(LVQ)
This is a supervised version of vector quantization.
Classes are predefined and we have a set of labeled
data.
The goal is to determine a set of prototypes that best
represent each class.
April 2007 31
BASIC SCHEME OF LVQ
Step 1: Initialize prototype vectors for different classes.
Step 2: Present a single input.
Step 3: Identify the closest prototype, i.e., the so-called
winner.
Step 4: Move the winner
- closer toward the data (same class),
- away from the data (different class).
VARIANTS OF LVQ
LVQ 1
LVQ 2
LVQ 2.1
LVQ 3
April 2007 32
COUNTERPROPAGATION NETWORK
Another variant of the BPN is the
counterpropagation network (CPN).
April 2007 37
BASIC STRUCTURE OF FULL CPN
April 2007 38
39
• X* and y* are approximations to x and y
set of vectors.
• Network is designed to approximate a
continuous function.
• Vectors x and y flow in a counterflow
manner to yield x* and y*.
• No. of nodes in hidden layer of CPN is
greater than that of BPN for a given
accuracy level.
• CPN has a greater speed of learning.
40
• Uses hybrid learning unsupervised
learning(instar) and supervised
learning(outstar).
• Weight updation rule on winning
cluster units is (kohonen learning)
vi j (new) vij (old ) [ xi vij (old )], i 1...n
wkJ (new) wk J (old ) [ yk wkJ (old )], k 1....m
• Weights between winning unit J and
output units are adjusted
as(Grossberg)
u Jk (new) u Jk (old ) a[ yk u Jk (old )], k 1....m
t Ji (new) t Ji (old ) b[ xi t Ji (old )], i 1...n 41
FIRST PHASE OF CPN
April 2007 42
SECOND PHASE OF CPN
April 2007 43
CPN LEARNING PROCESS
April 2007 44
CPN LEARNING PROCESS
April 2007 45
CPN LEARNING PROCESS
April 2007 46
CPN LEARNING PROCESS
April 2007 47
COUNTERPROPAGATION NETWORK
After the first phase of the training, each hidden-layer
neuron is associated with a subset of input vectors.
After the training has finished, the network maps the training
vectors onto output vectors that are close to the desired
ones. The more hidden units, the better the mapping.
April 2007 49
Training of CPN
April 2007 50
April 2007 51
Forward Only Counterpropagation
Network
April 2007 52
Forward Only
Counterpropagation Network
• Simplified version of Full CPN
• Approximation of th function y=f(x)
but not x=f(y) can be done
• Only x vectors are used to form the
clusters on kohonen units in first
phase of training.
• First weights between input and
cluster layer are trained.
April 2007 53
• This is a specific network with known
target.
• Winning cluster unit sends its signal
to output layer.
• At the output layer difference
between wJk and target yk is
calculated.
• Based on this weights between
cluster and output layer are updated.
April 2007 54
Weight updation between input units
to cluster units
April 2007 55
ADAPTIVE RESONANCE THEORY
(ART) NETWORK
Adaptive Resonance Theory (ART) is a family of
algorithms for unsupervised learning developed by
Carpenter and Grossberg.
April 2007 56
ARCHITECTURES OF ART NETWORK
April 2007 57
ART1 Network
• It was developed to solve the problem
of instability of feed forward nets
• For each pattern presented to the
network, appropriate cluster unit is
chosen.
• Weights to cluster unit are adjusted to
let the cluster unit learn the pattern.
• This network controls degree of
similarity of patterns placed on same
cluster unit.
April 2007 58
ART1 Network
• These networks possess stability –
plasticity properties.
• I/p pattern should not be presented
on same cluster unit when the
pattern is presented each time.
• Based on this stability is defined as
that wherein a pattern is not
presented to previous cluster units.
It is achieved by reducing learning
rates.
April 2007 59
ART1 Network
• The ability of the pattern to respond
to a new pattern equally at any stage
of learning is called as plasticity.
• Stability –plasticity can be resolved
bya network that includes bottom-
up(i/p-o/p) competitive learning
combined with top-down (o/p-i/p)
learning.
April 2007 60
ART1 Network
• Instability of instar-outstar n/w is
resolved by reducing lerning rate
gradually to zero by freezing learned
categories.(react to new data)
63
FUNDAMENTAL ALGORITHM OF ART
NETWORK
64
Supplemental unit
65
Training algorithm
66
67
68
69
70
71
April 2007 72
ART2 NETWORK
April 2007 73
BASIC ARCHITECTURE OF ART2
April 2007 74
ART2 ALGORITHM
April 2007 75
ART 2 Algorithm
April 2007 76
April 2007 77
April 2007 78
ART2 Algorithm
April 2007 79
SUMMARY
This chapter discussed on the various unsupervised
learning networks like:
Max Net
Mexican Hat
Kohonen Self-organizing Feature Maps
Learning Vector Quantization
Counterpropagation Networks
Hamming Network
Adaptive Resonance Theory
April 2007 80