0% found this document useful (0 votes)
4 views81 pages

Lecture 2

The document provides an overview of Artificial Neural Networks (ANNs), including their history, functionality, and comparison with biological brains. It discusses key concepts such as learning paradigms, types of neural networks, and the structure of artificial neurons. The document emphasizes the significance of ANNs in replicating brain functions and solving complex problems through parallel processing.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views81 pages

Lecture 2

The document provides an overview of Artificial Neural Networks (ANNs), including their history, functionality, and comparison with biological brains. It discusses key concepts such as learning paradigms, types of neural networks, and the structure of artificial neurons. The document emphasizes the significance of ANNs in replicating brain functions and solving complex problems through parallel processing.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 81

Vivekanand Education Society’s

Institute of Technology,
Chembur, Mumbai

ARTIFICIAL NEURAL NETWORK AND FUZZY LOGIC


FA C U LT Y N A M E : A S H W I N I S AWA N T
A S S I S TA N T P R O F E S S O R
D E PA RT M E N T O F E L E C T R O N I C S & T E L E C O M M U N I C AT I O N E N G I N E E R I N G
Artificial Neural Networks

Ashwini Sawant
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 2
Agenda
 History of Artificial Neural Networks
 What is an Artificial Neural Networks?
 How it works?
 Learning
 Learning paradigms
 Supervised learning
 Unsupervised learning
 Reinforcement learning
 Applications areas
 Advantages and Disadvantages
History of the Artificial Neural Networks
 History of the ANNs stems from the 1940s, the decade of the first electronic
computer.
 However, the first important step took place in 1957 when Rosenblatt introduced
the first concrete neural model, the perceptron. Rosenblatt also took part in
constructing the first successful neurocomputer, the Mark I Perceptron. After this,
the development of ANNs has proceeded as described in Figure.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 4


History of the Artificial Neural Networks
 Rosenblatt's original perceptron model contained only one layer. From this, a
multi-layered model was derived in 1960. At first, the use of the multi-layer
perceptron (MLP) was complicated by the lack of a appropriate learning
algorithm.
 In 1974, Werbos came to introduce a so-called backpropagation algorithm for the
three-layered perceptron network.
History of the Artificial Neural Networks
 In 1986, The application area of the MLP networks remained rather limited until the
breakthrough when a general back propagation algorithm for a multi-layered
perceptron was introduced by Rummelhart and Mclelland.
 in 1982, Hopfield brought out his idea of a neural network. Unlike the neurons in
MLP, the Hopfield network consists of only one layer whose neurons are fully
connected with each other.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6


History of the Artificial Neural Networks
 Since then, new versions of the Hopfield network have been developed. The
Boltzmann machine has been influenced by both the Hopfield network and
the MLP.
History of the Artificial Neural Networks
 in 1988, Radial Basis Function (RBF) networks were first introduced by
Broomhead & Lowe. Although the basic idea of RBF was developed 30
years ago under the name method of potential function, the work by
Broomhead & Lowe opened a new frontier in the neural network
community.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 8


History of the Artificial Neural Networks
 in 1982, A totally unique kind of network model is the Self-Organizing Map
(SOM) introduced by Kohonen. SOM is a certain kind of topological map
which organizes itself based on the input patterns that it is trained with. The
SOM originated from the LVQ (Learning Vector Quantization) network the
underlying idea of which was also Kohonen's in 1972.
History of Artificial Neural Networks
Since then, research on artificial neural networks has
remained active, leading to many new network types, as
well as hybrid algorithms and hardware for neural
information processing.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 10


Evolution of neural networks
Year Neural network Designer Description
1943 McCulloch and Pitts McCulloch and Pitts Arrangement of
neuron neurons is combination
of logic gate. Unique
feature is thresh hold
1949 Hebb network Hebb If two neurons are
active, then their
connection strengths
should be increased.
1958,1959,1962,1988,1 Perceptron Frank Rosenblatt, Block, Weights are adjusted to
960 Adaline Minsky and Papert reduce the difference
Widrow and Hoff between the net input
to the output unit and
the desired output

11
Contd…
1972 Kohonen self- Kohonen Inputs are clustered
organizing feature to obtain a fired
map output neuron.
1982, Hopfield network John Hopfield and Based on fixed
1984, Tank weights.
1985, Can act as associative
1986, memory nets
1987
1986 Back propagation Rumelhart, Hinton i) Multilayered
network and Williams ii) Error propagated
backward from
output to the
hidden units

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 1212


Contd..
1988 Counter Grossberg Similar to kohonen
propagation network
network
1987-1990 Adaptive resonance Carpenter and Designed for binary
Theory(ART) Grossberg and analog inputs.
1988 Radial basis function Broomhead and Resemble back
network Lowe propagation
network , but
activation function
used is Gaussian
function
1988 Neo cognitron Fukushima For character
recognition.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 1313
Artificial Neural Network
 An artificial neural network consists of a pool of simple
processing units which communicate by sending signals to
each other over a large number of weighted connections.
 A Neural Network is a machine that is designed to model
the way in which the brain performs a particular task or
function of interest; the network is usually implemented
by using electronic components or is simulated in software
on a digital computer.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 14


Artificial Neural Network
 A set of major aspects of a parallel distributed model include:
▪ a set of processing units (cells).
▪ a state of activation for every unit, which equivalent to the output of the unit.
▪ connections between the units. Generally each connection is defined by a weight.
▪ a propagation rule, which determines the effective input of a unit from its external inputs.
▪ an activation function, which determines the new level of activation based on the
effective input and the current activation.
▪ an external input for each unit.
▪ a method for information gathering (the learning rule).
▪ an environment within which the system must operate, providing input signals and _ if
necessary _ error signals.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 15


Computers vs. Neural Networks
“Standard” Computers Neural Networks

 one CPU highly parallel processing

fast processing units slow processing units

reliable units unreliable units

static infrastructure dynamic infrastructure

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 16


Brain Vs computer
Term Brain Computer
Speed Execution time is few milliseconds Execution time is few nano seconds

Processing Perform massive parallel operations Perform several parallel operations


simultaneously simultaneously. It is faster the
biological neuron
Size and complexity Number of Neuron is 1011 and It depends on the chosen application
number of interconnections is 1015. and network designer.
So complexity of brain is higher than
computer

Storage capacity i) Information is stored in i) Stored in continuous memory


interconnections or in synapse location.
strength. ii) Overloading may destroy older
ii) New information is stored without locations.
destroying old one. iii) Can be easily retrieved
iii) Sometimes fails to recollect
information

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 1717


Contd…
Tolerance i) Fault tolerant i) No fault tolerance
ii) Store and retrieve ii) Information corrupted
information even if the network
interconnections fails connections
iii) Accept redundancies disconnected.
iii) No redundancies
Control mechanism Depends on active CPU
chemicals and neuron Control mechanism is very
connections are strong or simple
weak

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 1818


Why Artificial Neural Networks?
There are two basic reasons why we are interested in building
artificial neural networks (ANNs):

• Technical viewpoint: Some problems such as


character recognition or the prediction of future
states of a system require massively parallel and
adaptive processing.

• Biological viewpoint: ANNs can be used to


replicate and simulate components of the human
(or animal) brain, thereby giving us insight into
natural information processing.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 19
Artificial Neural Networks
• The “building blocks” of neural networks are the neurons.
• In technical systems, we also refer to them as units or nodes.
• Basically, each neuron
 receives input from many other neurons.
 changes its internal state (activation) based on the current input.
 sends one output signal to many other neurons, possibly including
its input neurons (recurrent network).

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 20


Artificial Neural Networks
• Information is transmitted as a series of electric impulses, so-
called spikes.

• The frequency and phase of these spikes encodes the


information.

• In biological systems, one neuron can be connected to as many


as 10,000 other neurons.

• Usually, a neuron receives its information from other neurons


in a confined area, its so-called receptive field.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 21


How do ANNs work?
 An neural network (ANN) is either a hardware
artificial
implementation or a computer program which strives to simulate the
information processing capabilities of its biological exemplar. ANNs are
typically composed of a great number of interconnected artificial neurons.
The artificial neurons are simplified models of their biological
counterparts.
 ANN is a technique for solving problems by constructing software that
works like our brains.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 22


How do our brains work?
▪ The Brain is A massively parallel information processing
system.
▪ Our brains are a huge network of processing elements. A
typical brain contains a network of 10 billion neurons.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 23


How do our brains work?
▪ A processing element

Dendrites: Input
Cell body: Processor
Synaptic: Link
Axon: Output
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 24
How do our brains work?
▪ A processing element

A neuron is connected to other neurons through about 10,000


synapses

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 25


How do our brains work?
▪ A processing element

A neuron receives input from other neurons. Inputs are combined.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 26


How do our brains work?
▪ A processing element

Once input exceeds a critical level, the neuron discharges a spike ‐


an electrical pulse that travels from the body, down the axon, to
the next neuron(s)
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 27
How do our brains work?
▪ A processing element

The axon endings almost touch the dendrites or cell body of the
next neuron.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 28


How do our brains work?
▪ A processing element

Transmission of an electrical signal from one neuron to the next is


effected by neurotransmitters.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 29


How do our brains work?
▪ A processing element

Neurotransmitters are chemicals which are released from the first neuron
and which bind to the
Second.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 30
How do our brains work?
▪ A processing element

This link is called a synapse. The strength of the signal that


reaches the next neuron depends on factors such as the amount of
neurotransmitter available.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 31
How do ANNs work?

An artificial neuron is an imitation of a human neuron


20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 32
How do ANNs work?
• Now, let us have a look at the model of an artificial neuron.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 33


How do ANNs work?
............
Input xm x2 x1

Processing ∑
∑= X1+X2 + ….+Xm =y

Output y

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 34


How do ANNs work?
Not all inputs are equal
............
xm x2 x1
Input
wm ..... w2 w1
weights
Processing ∑ ∑= X1w1+X2w2 + ….+Xmwm
=y

Output y

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 35


How do ANNs work?
The signal is not passed down to the
next neuron verbatim
............
xm x2 x1
Input
wm ..... w2 w1
weights
Processing ∑
Transfer Function
f(vk)
(Activation Function)

Output y
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 36
The output is a function of the input, that is affected by the weights,
and the transfer functions

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 37


Artificial Neural Networks
 An ANN can:
1. compute any computable function, by the appropriate
selection of the network topology and weights values.
2. learn from experience!
▪ Specifically, by trial‐and‐error

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 38


Learning by trial‐and‐error
Continuous process of:
➢Trial:
Processing an input to produce an output (In terms of ANN: Compute
the output function of a given input)
➢Evaluate:
Evaluating this output by comparing the actual output with
the expected output.
➢Adjust:
Adjust the weights.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 39


Neural Networks
-N E U R A L NE T W O R K WA S I N S P I R E D B Y T H E D E S I G N A N D F U N C T I O N I N G O F
HUMAN BRAIN AND COMPONENTS.
-D E F I N I T I O N :
-I N F O R M AT I O N P R O C E S S I N G M O D E L T H AT I S I N S P I R E D B Y T H E WAY
B I O L O G I C A L N E RV O U S S Y S T E M ( I . E ) T H E B R A I N , P R O C E S S I N F O R M AT I O N .
-A N N I S C OM P O S E D O F L A R G E N U M B E R O F H I G H LY I N T E R C O NN E C T E D
P R O C E S S I N G E L E M E N T S ( N E U R O N S ) W O RK I N G I N U N I S O N TO S O LV E
PROBLEMS.
-I T I S C ON F I G U R E D F O R S P E C I A L A P P L I C AT I O N S U C H A S PAT T E R N
R E C O G N I T I O N A N D D ATA C L A S S I F I C AT I O N T H R O U G H A L E A R N I N G
PROCESS.
-8 5 - 9 0 % A C C U R AT E .

40
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 40
Advantages of Neural Networks
• A Neural Network can be an “expert” in analyzing the category of information given to it.
• Answers “ what-if” questions
• Adaptive learning
– Ability to learn how to do tasks based on the data given for training or initial experience.
• Self organization
– Creates its own organization or representation of information it receives during learning time.
• Real time operation
– Computations can be carried out in parallel.
• Fault tolerance via redundant information coding
– Partial destruction of neural network cause degradation of performance.
– In some cases, it can be retained even after major network damage.
• In future, it can also used to give spoken words as instructions for machine.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 4141


This figure shows the multi
disciplinary point of view of Neural
Networks

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 4242


Application Scope of Neural
Networks
1. Animal behavior 12. Air traffic control
2. Appraisal and valuation of property, etc., 13. Fraud detection
3. Betting on horse races, stock markets 14. Hand writing and typewriting
4. Criminal sentencing 15. Lake water levels
5. Complex physical and chemical process 16. Machinery controls
6. Data mining, cleaning and validation 17. Medical diagnosis
7. Direct mail advertisers 18. Music composition
8. Echo patterns 19. Photos and finger prints
9. Economic modeling 20. Recipes and chemical formulation
10. Employee hiring 21. Traffic flows
11. Expert consultants 22. Weather prediction

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 43


Fuzzy Logic
Lofti Zadeh, Professor at University of California.
An organized method for dealing with imprecise data
Fuzzy logic includes 0 and 1 as extreme cases of truth (or "the state of matters" or "fact") but
also includes the various states of truth in between so that, for example, the result of a
comparison between two things could be not "tall" or "short" but ".38 of tallness.“
Allows partial membership
Implemented in small, embedded micro controllers to large , networked, multichannel PC or
work station.
Can be implemented in hardware, software or in both.
It mimics how a person would make decisions.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 4444


Genetic algorithm
How genes of parents combine to form those of their children.
Create an initial population of individuals representing possible solutions to
solve a problem
Individual characters determine whether they are less or more fit to the
population
The more fit members will take high probability.
It is very effective in finding optimal or near optimal solutions.
Generate and test strategy.
Differ from normal optimization and search procedures in:
◦ Work with coding of the parameter set
◦ Work with multiple points
◦ Search via sampling( a blind search)
◦ Search using stochastic opeartors
In business, scientific and engineering circles, etc.,

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 4545


Hybrid System
Three types:-
Neuro Fuzzy hybrid system
◦ Combination of fuzzy set theory and neural networks
◦ Fuzzy system deal with explicit knowledge that can be explained and understood
◦ Neural network deal with implicit knowledge acquired by learning
◦ Advantages are:
◦ Handle any kind of information
◦ Manage imprecise, partial, vague or imperfect information
◦ Resolve conflicts by collaboration and aggregation.
◦ Self learning, self organizing and self tuning capability
◦ No need for prior knowledge of relationship of data
◦ Mimic human decision making system
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 4646
Contd..
Neuron genetic hybrid system
◦ Topology optimization
◦ Select a topology for ANN , common one is back propagation
◦ Genetic algorithm training
◦ Learning of ANN is formulated ad weight optimization problem, usually mean
squared error as a fitness measure
◦ Control parameter optimization
◦ Learning rate, momentum rate, tolerance level. Etc., are optimized using GA.
Fuzzy genetic hybrid system
◦ Creating the classification rules for a fuzzy system where objects are classified
by linguistic terms.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 4747
Soft computing
Two major problem solving techniques are:
◦ Hard computing
◦ Deals with precise model where
accurate solutions are achieved.
◦ Soft computing
◦ deals with approximate model to give
solution for complex problems
◦ Prof. Lotfi Zadeh introduced it.
◦ Ultimate goal-emulate the human mind
◦ It is a combination of GA, Neural
Network and FL.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 4848


Artificial Neural Network : An
Introduction
 Resembles the characteristic of biological neural network.
 Nodes – interconnected processing elements (units or neurons)
 Neuron is connected to other by a connection link.
 Each connection link is associated with weight which has information about
the input signal.
 ANN processing elements are called as neurons or artificial neurons , since
they have the capability to model networks of original neurons as found in
brain.
 Internal state of neuron is called activation or activity level of neuron, which
is the function of the inputs the neurons receives.
 Neuron can send only one signal at a time.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 4949
Basic Operation of a Neural Net
X1 and X2 – input neurons.
Y- output neuron
Weighted interconnection links- W1
and W2.
Net input calculation is :

Output is :

Output= function

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5050


Contd…
The function to be applied over the net input is called activation function.
Weight involved in ANN is equal to the slope of linear straight line (y=mx).

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5151


Biological Neural Network
Has three main parts
◦ Soma or cell body-where cell nucleus is located
◦ Dendrites-where the nerve is connected to the cell body
◦ Axon-which carries the impulses of the neuron
Electric impulse is passed between synapse and dendrites.
Synapse- Axon split into strands and strands terminates into small bulb like organs called
as synapse.
It is a chemical process which results in increase /decrease in the electric potential inside
the body of the receiving cell.
If the electric potential reaches a thresh hold value, receiving cell fires & pulse / action
potential of fixed strength and duration is send through the axon to synaptic junction of the
cell.
After that, cell has to wait for a period called refractory period.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5252
Contd..
In this model net input is calculated by

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5353


Terminology Relation Between Biological And
Artificial Neuron
Biological Neuron Artificial Neuron
Cell Neuron
Dendrites Weights or
interconnections
Soma Net input
Axon Output

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5454


Characteristics of ANN:
Neurally implemented mathematical model
Large number of processing elements called neurons exists here.
Interconnections with weighted linkage hold informative knowledge.
Input signals arrive at processing elements through connections and
connecting weights.
Processing elements can learn, recall and generalize from the given data.
Computational power is determined by the collective behavior of neurons.
◦ ANN is a connection models, parallel distributed processing models, self-
organizing systems, neuro-computing systems and neuro morphic system.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5555
Basic models of ann
Models are based on three entities
◦ The model’s synaptic interconnections.
◦ The training or learning rules adopted for updating and adjusting the connection weights.
◦ Their activation functions
The arrangement of neurons to form layers and the connection pattern formed within and
between layers is called the network architecture.
Five types:
◦ Single layer feed forward network
◦ Multilayer feed-forward network
◦ Single node with its own feedback
◦ Single-layer recurrent network
◦ Multilayer recurrent network

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5656


Single layer Feed- Forward Network
Layer is formed by taking processing
elements and combining it with other
processing elements.
Input and output are linked with each
other
Inputs are connected to the
processing nodes with various
weights, resulting in series of outputs
one per node.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5757


Multilayer feed-forward network
Formed by the interconnection of several
layers.
Input layer receives input and buffers input
signal.
Output layer generated output.
Layer between input and output is called
hidden layer.
Hidden layer is internal to the network.
Zero to several hidden layers in a network.
More the hidden layer, more is the complexity
of network, but efficient output is produced.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5858
Feed back network
If no neuron in the output layer is an input to a node in the
same layer / proceeding layer – feed forward network.
If outputs are directed back as input to the processing
elements in the same layer/proceeding layer –feedback
network.
If the output are directed back to the input of the same layer
then it is lateral feedback.
Recurrent networks are networks with feedback networks
with closed loop.
Fig 2.8 (A) –simple recurrent neural network having a
single neuron with feedback to itself.
Fig 2.9 – single layer network with feedback from output
can be directed to processing element itself or to other
processing element/both.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 5959
Maxnet –competitive interconnections having fixed
weights.
On-center-off-surround/lateral inhibiton structure –
each processing neuron receives two different classes of
inputs- “excitatory” input from nearby processing
elements & “ inhibitory” elements from more distantly
located precessing elements. This type of
interconnection is shown below

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6060


Multilayer Recurrent Network
Processing element output can be directed
back to the nodes in the preceding layer,
forming a multilayer recurrent network.
Processing element output can be directed
to processing element itself or to other
processing element in the same layer.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6161


learning
Two broad kinds of learning in ANNs is :
i) parameter learning – updates connecting weights in a neural net.
ii) Structure learning – focus on change in the network.
Apart from these, learning in ANN is classified into three categories as
◦ i) supervised learning
◦ ii) unsupervised learning
◦ iii) reinforcement learning

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6262


Supervised learning
Learning with the help of a teacher.
Example : learning process of a small child.
 Child doesn’t know read/write.
 Their each & every action is supervised by a teacher
In ANN, each input vector requires a corresponding target vector, which
represents the desired output.
The input vector along with target vector is called training pair.
The input vector results in output vector.
The actual output vector is compared with desired output vector.
If there is a difference means an error signal is generated by the
network.
It is used for adjustment of weights until actual output matches desired
output.
20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6363
Unsupervised learning
Learning is performed without the help of a teacher.
Example: tadpole – learn to swim by itself.
In ANN, during training process, network receives input
patterns and organize it to form clusters.
From the Fig. it is observed that no feedback is applied
from environment to inform what output should be or
whether they are correct.
The network itself discover patterns, regularities, features/
categories from the input data and relations for the input
data over the output.
Exact clusters are formed by discovering similarities &
dissimilarities so called as self – organizing.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 64


64
Reinforcement learning
Similar to supervised learning.
Learning based on critic information is called
reinforcement learning & the feedback sent is
called reinforcement signal.
The network receives some feedback from the
environment.
Feedback is only evaluative.
The external reinforcement signals are processed
in the critic signal generator, and the obtained
critic signals are sent to the ANN for adjustment
of weights properly to get critic feedback in
future.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6565


Activation functions
To make work more efficient and for exact output, some force or activation is given.
Like that, activation function is applied over the net input to calculate the output of an ANN.
Information processing of processing element has two major parts: input and output.
An integration function (f) is associated with input of processing element.
Several activation functions are there.
1. Identity function:
 it is a linear function which is defined as
f(x) =x for all x
 The output is same as the input.
2. Binary step function
 it is defined as

where θ represents thresh hold value.


It is used in single layer nets to convert the net input to an output that is bianary. ( 0 or 1)

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6666


Contd..
3. Bipolar step function:
• It is defined as

• where θ represents threshold value.


• used in single layer nets to convert the net input to an output that is bipolar (+1 or -1).
4. Sigmoid function
used in Back propagation nets.
Two types:
a) binary sigmoid function
-logistic sigmoid function or unipolar sigmoid function.
-it is defined as
where λ – steepness parameter.
-The derivative of this function is
f’(x) = λ f(x)[1-f(x)]. The range of sigmoid function is 0 to 1.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6767


Contd..
b) Bipolar sigmoid function

where λ- steepness parameter and the sigmoid range is between -1 and +1.
- The derivative of this function can be
[1+f(x)][1-f(x)]
- It is closely related to hyberbolic tangent function, which is written as

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6868


Contd..
The derivative of the hyberbolic tangent function is
h’(x)= [1+h(x))][1-h(x)]
5. Ramp function

The graphical representation of all these function is given in the upcoming Figure

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 6969


20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 7070
Important terminologies
Weight
 The weight contain information about the input signal.
 It is used by the net to solve the problem.
 It is represented in terms of matrix & called as connection matrix.
 If weight matrix W contains all the elements of an ANN, then the set of all W matrices will determine the
set of all possible information processing configuration.
 The ANN can be realized by finding an appropriate matrix W.
 Weight encode long-term memory (LTM) and the activation states of network encode short-term memory
(STM) in a neural network.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 71


71
Contd..
Bias
◦ Bias has an impact in calculating net input.
◦ Bias is included by adding x0 to the input vector x.
◦ The net output is calculated by

◦ The bias is of two types


◦ Positive bias
◦ Increase the net input
◦ Negative bias
◦ Decrease the net input

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 7272


Contd..
Threshold
◦ It is a set value based upon which the final output is calculated.
◦ Calculated net input and threshold is compared to get the network output.
◦ The activation function of threshold is defined as

◦ where θ is the fixed threshold value

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 7373


Contd..
Learning rate
 Denoted by α.
 Control the amount of weight adjustment at each step of training.
 The learning rate range from 0 to 1.
 Determine the rate of learning at each step
Momentum Factor
 Convergence is made faster if a momentum factor is added to the weight updation process.
 Done in back propagation network.
Vigilance parameter
 Denoted by ρ.
 Used in Adaptive Resonance Theory (ART) network.
 Used to control the degree of similarity.
 Ranges from 0.7 to 1 to perform useful work in controlling the number of clusters.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 7474


Mcculloch-pitts neuron
Discovered in 1943.
Usually called as M-P neuron.
M-P neurons are connected by directed weighted paths.
Activation of M-P neurons is binary (i.e) at any time step the neuron may fire or may not fire.
Weights associated with communication links may be excitatory(wgts are
positive)/inhibitory(wgts are negative).
Threshold plays major role here. There is a fixed threshold for each neuron and if the net input to
the neuron is greater than the threshold then the neuron fires.
They are widely used in logic functions.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 7575


Contd…
A simple M-P neuron is shown in the
figure.
It is excitatory with weight (w>0) /
inhibitory with weight –p (p<0).
In the Fig., inputs from x1 to xn
possess excitatory weighted
connection and Xn+1 to xn+m has
inhibitory weighted interconnections.
Since the firing of neuron is based on
threshold, activation function is
defined as

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 76


76
Contd…
For inhibition to be absolute, the threshold with the activation function should satisfy the
following condition:
θ >nw –p

Output will fire if it receives “k” or more excitatory inputs but no inhibitory inputs where
kw≥θ>(k-1) w
- The M-P neuron has no particular training algorithm.
- An analysis is performed to determine the weights and the threshold.
- It is used as a building block where any function or phenomenon is modeled based on a
logic function.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 7777


Linear separability
It is a concept wherein the separation of the input space into regions is based on whether the
network response is positive or negative.
A decision line is drawn to separate positive or negative response.
The decision line is also called as decision-making line or decision-support line or linear-
separable line.
The net input calculation to the output unit is given as

The region which is called as decision boundary is determined by the relation

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 7878


Contd..
Consider a network having positive
response in the first quadrant and
negative response in all other
quadrants with either binary or bipolar
data.
Decision line is drawn separating two
regions as shown in Fig.
Using bipolar data representation,
missing data can be distinguished
from mistaken data. Hence bipolar
data is better than binary data.
Missing values are represented by 0
and mistakes by reversing the input
values from +1 to -1 or vice versa.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 7979


Hebb network
Donald Hebb stated in 1949 that “ In brain, the learning is performed by the change in the synaptic gap”.
When an axon of cell A is near enough to excite cell B, and repeatedly or permanently takes place in firing it, some
growth process or metabolic change takes place in one or both the cells such than A’s efficiency, as one of the cells
firing B, is increased.
According to Hebb rule, the weight vector is found to increase proportionately to the product of the input and the
learning signal.
In Hebb learning, two interconnected neurons are ‘on’ simultaneously.
The weight update in Hebb rule is given by
 Wi(new) = wi (old)+ xi y.

It is suited more for bipolar data.


If binary data is used, the weight updation formula cannot distinguish two conditions namely
 A training pair in which an input unit is “on” and the target value is “off”
 A training pair in which both the input unit and the target value is “off”.

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 8080


Flowchart of training
algorithm
Steps:
◦ 0: First initialize the weights.
◦ 1: Steps 2-4 have to be performed for each input training vector and target output pair, s:t
◦ 2: Input activations are set. The activation function for input layer is identity function.
◦ Xi =Si for i=1 to n
◦ 3: Output activations are set.
◦ 4: Weight adjustment and bias adjustments are performed.
◦ Wi(new) = wi(old)+xiy
◦ b(new)=b(old)+y
◦ In step 4, the weight updation formula can be written in vector form as
◦ w(new)=w(old)+y.
◦ Change in weight is expressed as
◦ Δw=xy
Hence,
w(new)=w(old)+Δw
Hebb rule is used for pattern association, pattern categorization, pattern classification
and over a range of other areas

20/02/2022 ASHWINI SAWANT, ASSISTANT PROFESSOR, VESIT 8181

You might also like