NN&DL Unit-II Unsupervise Learning Networks
NN&DL Unit-II Unsupervise Learning Networks
MAXNET
Maxnet network was developed by Lippmann in 1987.
The Maxner serves as a sub net for picking the node whose input is larger.
All the nodes present in this subnet are fully interconnected and there exist symmetrical weights in
all these weighted interconnections.
Architecture of Maxnet
The architecture of Maxnet is a fixed symmetrical weights are present over the weighted
interconnections. The weights between the neurons are inhibitory and fixed.
The Maxnet with this structure can be used as a subnet to select a particular node whose net input is
the largest.
Testing algorithm
Step – 0: Initial weights and initial activations are set. The weight is set as [0 < ε < 1/m], where "m"
is the total number of nodes. Let
Step – 3: Save the activations obtained for use in the next iteration. For j = 1 to m,
Step – 4: Finally, test the stopping condition for convergence of the network. The following is the
stopping condition: If more than one node has a nonzero activation, continue; else stop.
Hamming Network
The Hamming network is a two-layer feedforward neural network for classification of binary bipolar
n-tuple input vectors using minimum Hamming distance denoted as D H(Lippmann, 1987).
The first layer is the input layer for the n-tuple input vectors.
The second layer (also called the memory layer) stores p memory patterns.
A p-class Hamming network has p output neurons in this layer. The strongest response of a neuron is
indicative of the minimum Hamming distance between the stored pattern and the input vector.
Hamming Distance
Hamming distance of two vectors, x and y of dimension n
x.y = a - d
where: a is number of bits in agreement in x & y (No. of Similar bits in x & y), and d is number of
bits different in x and y (No. of Dissimilar bits in x & y).
The value "a - d" is the Hamming distance existing between two vectors. Since, the total number of
components is n, we have,
n=a+d
i.e., d = n - a
On simplification, we get
Mr. Mohammed Afzal, Asst. Professor in AIML
Mob: +91-8179700193, Email: [email protected]
From the above equation, it is clearly understood that the weights can be set to one-half the exemplar
vector and bias can be set initially to n/2
Initialize the bias for storing the "m" exemplar vectors. For j = 1 to m,
Step – 4: Maxnet is found to iterate for finding the exemplar that best matches the input
patterns.
Training Algorithm
Step – 0: Initialize the weights with Random values and the learning rate
Step – 1: Perform Steps 2-8 when stopping condition is false.
Step – 2: Perform Steps 3-5 for each input vector x.
Step – 3: Compute the square of the Euclidean distance, i.e., for each j = i to m,
Training Algorithm
Step – 0: Initialize the reference vectors. This can be done using the following steps.
From the given set of training vectors, take the first "m" (number of clusters) training
vectors and use them as weight vectors, the remaining vectors can be used for training.
o Assign the initial weights and classifications randomly.
o K-means clustering method.
o Set initial learning rate α
Step – 1: Perform Steps 2-6 if the stopping condition is false.
If Euclidean distance method is used, find the cluster unit Z j whose squared distance
from input vectors is the smallest
If there occurs a tie in case of selection of winner unit, the unit with the smallest
index is the winner. Take the winner unit index as J.
Step – 5: Update the weights over the calculated winner unit Zj.
Step – 13: Update the weights from unit Zj to the output layers.
If Euclidean distance method is used, find the cluster unit Z j whose squared distance
from input patterns is the smallest
If there exists a tie in the selection of winner unit, the unit with the smallest index is
chosen as the winner.
Step – 5: Perform weight update for unit Zj. For i= 1 to n,
Step – 13: Update the weights from unit Zj to the output layers.
Operating Principal
The main operation of ART classification can be divided into the following phases −
Recognition phase − The input vector is compared with the classification presented at every node in
the output layer. The output of the neuron becomes “1” if it best matches with the classification applied,
otherwise it becomes “0”.
Comparison phase − In this phase, a comparison of the input vector to the comparison layer vector is
done. The condition for reset is that the degree of similarity would be less than vigilance parameter.
Search phase − In this phase, the network will search for reset as well as the match done in the above
phases. Hence, if there would be no reset and the match is quite good, then the classification is over.
Otherwise, the process would be repeated and the other stored pattern must be sent to find the correct
match.
Basic of Adaptive Resonance Theory (ART) Architecture
The adaptive resonant theory is a type of neural network that is self-organizing and competitive. It can be
of both types, the unsupervised ones (ART1, ART2, ART3, etc) or the supervised ones (ARTMAP).
Generally, the supervised algorithms are named with the suffix “MAP”. But the basic ART model is
unsupervised in nature and consists of:
o Cluster Unit (F2 layer) − This is a competitive layer. The unit having the largest net input is
selected to learn the input pattern. The activation of all other cluster units are set to 0.
o Reset Mechanism − The work of this mechanism is based upon the similarity between the top-down
weight and the input vector. Now, if the degree of this similarity is less than the vigilance parameter,
then the cluster is not allowed to learn the pattern and a rest would happen.
The F1 layer accepts the inputs and performs some processing and transfers it to the F2 layer that best
matches with the classification factor. There exist two sets of weighted interconnections for controlling
the degree of similarity between the units in the F1 and the F2 layer. The F2 layer is a competitive layer.
The cluster unit with the large net input becomes the candidate to learn the input pattern first and the rest
F2 units are ignored. The reset unit makes the decision whether or not the cluster unit is allowed to learn
the input pattern depending on how similar its top-down weight vector is to the input vector and to the
decision. This is called the vigilance test. Thus, we can say that the vigilance parameter helps to
incorporate new memories or new information. Higher vigilance produces more detailed memories, lower
vigilance produces more general memories.
Supplement Unit − Actually the issue with Reset mechanism is that the layer F2 must have to be inhibited
under certain conditions and must also be available when some learning happens. That is why two
supplemental units namely, G1 and G2 is added along with reset unit, R. They are called gain control units.
These units receive and send signals to the other units present in the network. ‘+’ indicates an excitatory
signal, while ‘−’ indicates an inhibitory signal.
Step – 2: Continue step 3-9, when the stopping condition is not true.
Step – 3: Continue step 4-6 for every training input.
Step – 4: Set activations of all F1 a and F1 units as follows
Step – 10: Now, after calculating the norm of vector x and vector s, we need to check the reset
condition as follows –
Step – 12: The stopping condition for algorithm must be checked and it may be as follows –