SC Unit-2
SC Unit-2
UNIT –II
ARTIFICIAL NEURAL NETWORKS
1 Analyze the Back propagation of Neural Network with neat diagram. [L4][CO2] [12M]
Back propagation Network
The back propagation learning algorithm is one of the most important
developments in neural networks (Bryson and Ho, 1969; Werbos, 1974; Lecun,
1985; Parker, 1985;umelhart, 1986). This learning algorithm is applied to multilayer
feed-forward networks consisting of processing elements with continuous differentiable
activation functions. The networks associated with back-propagation learning algorithm
are also called back- propagation networks. (BPNs).
For a given set of training input-output pair, this algorithm provides a procedure for
changing the weights in a BPN to classify the given input patterns correctly. The
basic concept for this weight update algorithm is simply the gradient descent
method. This is a methods were error is propagated back to the hidden unit. Back
propagation network is a training algorithm.
Step 0: Initialize weights and learning rate (take some small random values).
Step 1: Perform Steps 2-9 when stopping condition is false.
Step 2: Perform Steps 3-8 for each training pair.
Step 3: Each input unit receives input signal xi and sends it to the hidden unit
(i = l to n).
Step 4: Each hidden unit zj (j = 1 to p) sums its weighted input signals to
calculate net input:
= 0 + ∑
=1
= ()
and send the output signal from the hidden unit to the input of output
layer units.
Step 5: For each output unit (k = 1 to m), calculate the net input:
= 0 + ∑
=1
= ()
= ( − )′()
The term gets multiplied with the derivative of () to calculate the
error term:
= ′( )
Step 8: Each output unit (yk, k = 1 to m) updates the bias and weights:
Step 9: Check for the stopping condition. The stopping condition may be
certain number of epochs reached or when the actual output equals the
target output.
2 Discuss Self –Organizing Map algorithm and its features . [L2][CO2] [12M]
Kohonen Self-Organizing Feature Maps
This topology has 24 nodes in the distance-2 grid, 16 nodes in the distance-1
grid, and 8 nodes in the distance-0 grid, which means the difference between
each rectangular grid is 8 nodes. The winning unit is indicated by #.
Course Code: 19CS0544 R19
This topology has 18 nodes in the distance-2 grid, 12 nodes in the distance-1
grid, and 6 nodes in the distance-0 grid, which means the difference between
each rectangular grid is 6 nodes. The winning unit is indicated by #.
Architecture
Step 1 − Initialize the weights, the learning rate α and the neighborhood
topological scheme.
Step 2 − Continue step 3-9, when the stopping condition is not true.
Architecture
Following figure shows the architecture of LVQ which is quite similar to the
architecture of KSOM. As we can see, there are “n” number of input units
and “m” number of output units. The layers are fully interconnected with
having weights on them.
Course Code: 19CS0544 R19
Variants
Three other variants namely LVQ2, LVQ2.1 and LVQ3 have been developed
by Kohonen. Complexity in all these three variants, due to the concept that the
winner as well as the runner-up unit will learn, is more than in LVQ.
LVQ2
As discussed, the concept of other variants of LVQ above, the condition of
LVQ2 is formed by window. This window will be based on the following
parameters −
Course Code: 19CS0544 R19
Hamming Network
In most of the neural networks using unsupervised learning, it is essential to
compute the distance and perform comparisons. This kind of network is
Hamming network, where for every given input vectors, it would be clustered
Course Code: 19CS0544 R19
into different groups. Following are some important features of Hamming
Networks −
Max Net
This is also a fixed weight network, which serves as a subnet for selecting the
node having the highest input. All the nodes are fully interconnected and there
exists symmetrical weights in all these weighted interconnections.
Architecture
It uses the mechanism which is an iterative process and each node receives
inhibitory inputs from all other nodes through connections. The single node
whose value is maximum would be active or winner and the activations of all
other nodes would be inactive. Max Net uses identity activation function with
Architecture
Following are some important points to keep in mind about discrete Hopfield
network −
This model consists of neurons with one inverting and one non-
inverting output.
The output of each neuron should be the input of other neurons but not
the input of self.
Weight/connection strength is represented by wij.
Connections can be excitatory as well as inhibitory. It would be
excitatory, if the output of the neuron is same as the input, otherwise
inhibitory.
Weights should be symmetrical, i.e. wij = wji
Training Algorithm
Testing Algorithm
Step 1 − Initialize the weights, which are obtained from training algorithm by
using Hebbian principle.
Step 2 − Perform steps 3-9, if the activations of the network is not consolidated.
Step 3 − For each input vector X, perform steps 4-8.
Step 4 − Make initial activation of the network equal to the external input
vector X as follows −
Course Code: 19CS0544 R19
The architecture of BAM network consists of two layers of neurons which are
connected by directed weighted pare interconnections. The network dynamics
involve two layers of interaction. The BAM network iterates by sending the
signals back and forth between the two layers until all the neurons reach
equilibrium. The weights associated with the network are bidirectional. Thus,
BAM can respond to the inputs in either layer.
Figure shows a BAM network consisting of n units in X layer and m units in Y
layer. The layers can be connected in both directions(bidirectional) with the
result the weight matrix sent from the X layer to the Y layer is W and the
weight matrix for signals sent from the Y layer to the X layer is W T. Thus, the
Weight matrix is calculated in both directions.
Determination of Weights
Let the input vectors be denoted by s(p) and target vectors by t(p). p = 1, ... , P.
Then the weight matrix to store a set of input and target vectors, where
s(p) = (s1(p), .. , si(p), ... , sn(p))
t(p) = (t1(p), .. , tj(p), ... , tm(p))
Course Code: 19CS0544 R19
Each cell must have storage capability and logical circuits for matching its
content with external argument.
7 a Generalize the Adaptive Resonance Theory Neural Network [L6][CO2] [8M]
Architecture of ART1
Input unit (F1 layer) − It further has the following two portions −
F1(a) layer (Input portion) − In ART1, there would be no
processing in this portion rather than having the input vectors only. It
is connected to F1(b) layer (interface portion).
F1(b) layer (Interface portion) − This portion combines the signal
from the input portion with that of F2 layer. F1(b) layer is connected
to F2 layer through bottom up weights bij and F2 layer is connected to
F1(b) layer through top down weights tji.
Cluster Unit (F2 layer) − This is a competitive layer. The unit having
the largest net input is selected to learn the input pattern. The activation of
all other cluster unit are set to 0.
Reset Mechanism − The work of this mechanism is based upon the
similarity between the top-down weight and the input vector. Now, if the
degree of this similarity is less than the vigilance parameter, then the
cluster is not allowed to learn the pattern and a rest would happen.
Supplement Unit − Actually the issue with Reset mechanism is that the
layer F2 must have to be inhibited under certain conditions and must also be
available when some learning happens. That is why two supplemental units
namely, G1 and G2 is added along with reset unit, R. They are called gain
control units. These units receive and send signals to the other units present in
the network. ‘+’ indicates an excitatory signal, while ‘−’ indicates an inhibitory
signal.
Course Code: 19CS0544 R19
Target recognition:
Fuzzy ARTMAP neural network can be used for automatic classification of targets
depend on their radar range profiles. Tests on synthetic data show the fuzzy ARTMAP
can result in substantial savings in memory requirements when related to k nearest
neighbor(kNN) classifiers. The utilization of multi wave length profiles mainly
improves the performance of both kinds of classifiers.
Medical diagnosis:
Medical databases present huge numbers of challenges found in general information
management settings where speed, use, efficiency, and accuracy are the prime concerns.
A direct objective of improved computer-assisted medicine is to help to deliver
intensive care in situations that may be less than ideal. Working with these issues has
stimulated several ART architecture developments, including ARTMAP-IC.
Signature verification:
Course Code: 19CS0544 R19
Automatic signature verification is a well known and active area of research with
various applications such as bank check confirmation, ATM access, etc. the training of
the network is finished using ART1 that uses global features as input vector and the
verification and recognition phase uses a two-step process. In the initial step, the input
vector is coordinated with the stored reference vector, which was used as a training set,
and in the second step, cluster formation takes place.
Mobile control robot: Nowadays, we perceive a wide range of robotic devices. It is
still a field of research in their program part, called artificial intelligence. The human
brain is an interesting subject as a model for such an intelligent system. Inspired by the
structure of the human brain, an artificial neural emerges. Similar to the brain, the
artificial neural network contains numerous simple computational units, neurons that
are interconnected mutually to allow the transfer of the signal from the neurons to
neurons. Artificial neural networks are used to solve different issues with good
outcomes compared to other decision algorithms.
8 Illustrate the Support Vector Machine with neat diagram. [L3][CO2] [12M]
SVM:
The goal of the SVM algorithm is to create the best line or decision boundary
that can segregate n-dimensional space into classes so that we can easily put
the new data point in the correct category in the future. This best decision
boundary is called a hyperplane
Types of SVM
o Linear SVM: Linear SVM is used for linearly separable data, which
means if a dataset can be classified into two classes by using a single
straight line, then such data is termed as linearly separable data, and
classifier is used called as Linear SVM classifier.
o Non-linear SVM: Non-Linear SVM is used for non-linearly separated
Course Code: 19CS0544 R19
data, which means if a dataset cannot be classified by using a straight
line, then such data is termed as non-linear data and classifier used is
called as Non-linear SVM classifier.
Applications of SVM
Face detection
image classification
Medical Diagnosis
text categorization, etc.
Support Vectors:
The data points or vectors that are the closest to the hyperplane and which
affect the position of the hyperplane are termed as Support Vector. Since these
vectors support the hyperplane, hence called a Support vector.
Linear SVM:
Non-Linear SVM:
By adding the third dimension, the sample space will become as below image:
Course Code: 19CS0544 R19
This rule, one of the oldest and simplest, was introduced by Donald Hebb in his
book The Organization of Behavior in 1949. It is a kind of feed-forward,
unsupervised learning.
Basic Concept − This rule is based on a proposal given by Hebb, who wrote −
From the above postulate, we can conclude that the connections between two
neurons might be strengthened if the neurons fire at the same time and might
weaken if they fire at different times.
Now the output ‘y’ can be calculated, as explained earlier on the basis of the net
input, and activation function being applied over that net input can be expressed
as follows –
Describe the structure of back propagation neural network and derive [L2][CO2] [12M]
the learning rule for the back propagation algorithm.
10
It is same as question 1 answer