IT 701 Soft Computing Unit I - 1722317885
IT 701 Soft Computing Unit I - 1722317885
Comparison between Biological Neuron and Artificial Neuron (Brain vs. Computer)
It is called supervised learning because the process of algorithm learning from the training
dataset can be thought of as a teacher supervising the learning process. We know the correct
answers; the algorithm iteratively makes predictions on the training data and is corrected by
the teacher. Learning stops when the algorithm achieves an acceptable level of performance.
During training, the input vector is presented to the network, which results in an output vector.
This output vector is the actual output vector. Then the actual output vector is compared with
the desired (target) output vector. If there exists a difference between the two output vectors
then an error signal is generated by the network. This error is used for adjustment of weights
until the actual output matches the desired (target) output.
In this type of learning, a supervisor or teacher is required for error minimization. Hence, the
network trained by this method is said to be using supervised training methodology. In
supervised learning it is assumed that the correct “target” output values are known for each
input pattern.
Unsupervised Learning
Unsupervised learning is where you only have input data (X) and no corresponding output
variables.
The goal for unsupervised learning is to model the underlying structure or distribution in the
data in order to learn more about the data.
In ANNs following unsupervised learning, the input vectors of similar type are grouped without
the use of training data to specify how a member of each group looks or to which group a
number belongs. In the training process, the network receives the input patterns and organizes
these patterns to form clusters. When a new input pattern is applied, the neural network gives
an output response indicating the class to which the input pattern belongs. If for an input, a
pattern class cannot be found then a new class is generated.
These are called unsupervised learning because unlike supervised learning above there is no
correct answer and there is no teacher. Algorithms are left to their own devises to discover and
present the interesting structure in the data.
From the working of unsupervised learning it is clear that there is no feedback from the
environment to inform what the outputs should be or whether the outputs are correct. In this
case the network must itself discover patterns, regularities, features or categories from the
input data and relations for the input data over the output.
Linear separability is a powerful technique which is used to learn complicated concepts that are
considerably more complicated than just hyperplane separation.
Let f be the XOR function which expects two binary inputs and generates a binary output .Let us
try to represent the XOR function by means of an SLP with two input neurons i 1, i2 and one
output neuron:
Figure 1.8: Single layer perceptron
Here we use the weighted sum as propagation function, a binary activation function with the
threshold value and the identity as output function. Depending on i 1 and i2 the output becomes
the value 1 if the following holds:
----- (1)
------- (2)
With a constant threshold value, the right part of in equation 2 is a straight line through a
coordinate system defined by the possible outputs oi1 und oi2 of the input neurons i1 and i2.
Figure 1.9: Linear separation of n=2 inputs of the input neuron i1 and i2 by 1-D line, A and B
show the corner belonging to sets of XOR function that are to be separated
For a positive wi2,Ω the output neuron fires for input combinations lying above the generated
straight line. For a negative wi2,Ω it would fire for all input combinations lying below the straight
line. Note that only the four corners of the unit square are possible inputs because the XOR
function only knows binary inputs.
In order to solve the XOR problem, we have to turn and move the straight line so that input set
A = {(0, 0), (1, 1)} is separated from input set B = {(0, 1), (1, 0)} this is, obviously, impossible.
Generally, the input parameters of n man input neurons can be represented in an n
dimensional cube which is separated by an SLP through an (n−1)-dimensional hyper plane. Only
sets that can be separated by such a hyper plane, i.e. which are linearly separable, can be
classified by an SLP.
Figure 1.10: Linear separation of n=3 inputs of the input neuron i1, i2 and i3 by 2-D plane
Unfortunately, it seems that the percentage of the linearly separable problems rapidly
decreases with increasing, which limits the functionality of the SLP. Additionally, tests for linear
separability are difficult. Thus, for more difficult tasks with more inputs we need something
more powerful than SLP. The XOR problem itself is one of these tasks, since a perceptron that is
supposed to represent the XOR function already needs a hidden layer.
Activation Function
Let us assume a person is performing some task. To make the task more efficient and to obtain
exact output, some force or activation may be given. This activation helps in achieving the exact
output. In the similar way, the activation function is applied over the net input to calculate the
output of ANN.
It’s just a thing (node) that you add to the output end of any neural network. It is also known as
Transfer Function. It can also be attached in between two Neural Networks.
The information processing of a processing element can be viewed as consisting of two major
parts: input and output. An integration function (f) is associated with the input of a processing
element. This function serves to combine activation, information or evidence from an external
source or other processing elements into a net input to the processing element.
Models of ANN
Feed Forward Network
A feed forward neural network is an artificial neural network wherein connections between the
units do not form a cycle. As such, it is different from recurrent neural networks.
The feed forward neural network was the first and simplest type of artificial neural network
devised. In this network, the information moves in only one direction, forward, from the input
nodes, through the hidden nodes (if any) and to the output nodes.
Learning Rule
Basically, learning means to do and adapt the change in itself as and when there is a change in
environment. ANN is a complex system or more precisely we can say that it is a complex
adaptive system, which can change its internal structure based on the information passing
through it.
Learning rule or Learning process is a method or a mathematical logic. It improves the Artificial
Neural Network’s performance and applies this rule over the network. Thus learning rules
updates the weights and bias levels of a network when a network simulates in a specific data
environment.
Applying learning rule is an iterative process. It helps a neural network to learn from the existing
conditions and improve its performance.
Basic Concept − this rule is based on a proposal given by Hebb, who wrote −
“When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes
part in firing it, some growth process or metabolic change takes place in one or both cells such
that A’s efficiency, as one of the cells firing B, is increased.”
From the above postulate, we can conclude that the connections between two neurons might
be strengthened if the neurons fire at the same time and might weaken if they fire at different
times.
Δwji(t)=αxi(t).yj(t)
Here, Δwji(t)= Increment by which the weight of connection increases at time step t
α = The positive and constant learning rate
xi(t)= The input value from pre-synaptic neuron at time step t
yi(t)= The output of pre-synaptic neuron at same time step t
Basic Concept − The base of this rule is gradient-descent approach, which continues forever.
Delta rule updates the synaptic weights so as to minimize the net input to the output unit and
the target value.