Module 1 Ann
Module 1 Ann
Content.
• Introduction to ANN
• Architecture of ANN
• Types of perceptron
I. Introduction to ANN
The term "Artificial neural network" refers to a biologically inspired
sub-field of artificial intelligence modeled after the brain. An Artificial
neural network is usually a computational network based on biological
neural networks that construct the structure of the human brain. Similar
to a human brain has neurons interconnected to each other, artificial
neural networks also have neurons that are linked to each other in
various layers of the networks. These neurons are known as nodes.
Artificial neural network tutorial covers all the aspects related to the
artificial neural network. In this tutorial, we will discuss ANNs,
Adaptive resonance theory, Kohonen self-organizing map, Building
blocks, unsupervised learning, Genetic algorithm, etc.
The typical Artificial Neural Network looks something like the given
figure.
Dendrites from Biological Neural Network represent inputs in Artificial
Neural Networks, cell nucleus represents Nodes, synapse represents
Weights, and Axon represents Output.
Dendrites Inputs
Synapse Weights
Axon Output
There are around 1000 billion neurons in the human brain. Each neuron
has an association point somewhere in the range of 1,000 and 100,000.
In the human brain, data is stored in such a manner as to be distributed,
and we can extract more than one piece of this data when necessary from
our memory parallelly. We can say that the human brain is made up of
incredibly amazing parallel processors.
Input Layer:
As the name suggests, it accepts inputs in several different formats
provided by the programmer.
Hidden Layer:
Output Layer:
The artificial neural network takes input and computes the weighted sum
of the inputs and includes a bias. This computation is represented in the
form of a transfer function.
After ANN training, the information may produce output even with
inadequate data. The loss of performance here relies upon the
significance of missing data.
Hardware dependence:
ANNs can work with numerical data. Problems must be converted into
numerical values before being introduced to ANN. The presentation
mechanism to be resolved here will directly impact the performance of
the network. It relies on the user's abilities.
III. Perceptron
Types of Perceptron:
The Perceptron algorithm learns the weights for the input signals in
order to draw a linear decision boundary.
Step 1: Multiply all input values with corresponding weight values and
then add to calculate the weighted sum. The following is the
mathematical expression of it:
Y=f(∑wi*xi + b)
Backward Stage: In the backward stage, weight and bias values are
modified per the model’s requirement. The backstage removed the error
between the actual output and demands originating backward on the
output layer. A multilayer perceptron model has a greater processing
power and can process linear and non-linear patterns. Further, it also
implements logic gates such as AND, OR, XOR, XNOR, and NOR.
Advantages:
Helps us obtain the same accuracy ratio with big and small data.
Disadvantages:
3. Initially, weights are multiplied with input features, and then the
decision is made whether the neuron is fired or not.
6. If the added sum of all input values is more than the threshold
value, it must have an output signal; otherwise, no output will be
shown.
Limitation of Perceptron Model
IV Activation Function
Definition
Activation Function
o Linear Function
Uses: The output layer is the only location where the activation
function's function is applied.
o Sigmoid Function
o Tanh Function
Uses: - Since its values typically range from -1 to 1, the mean again for
hidden layer of a neural network will be 0 or very near to it. This helps
to centre the data by getting the mean close to 0. This greatly facilitates
learning for the following layer.
Equation:
Currently, the ReLU is the activation function that is employed the most
globally. Since practically all convolutional neural networks and deep
learning systems employ it.
However, the problem is that all negative values instantly become zero,
which reduces the model's capacity to effectively fit or learn from the
data. This means that any negative input to a ReLU activation function
immediately becomes zero in the graph, which has an impact on the final
graph by improperly mapping the negative values.
o Softmax Function
Although it is a subclass of the sigmoid function, the softmax function
comes in handy when dealing with multiclass classification issues.
A very logical choice for the output layer is the sigmoid function if your
input is for binary classification. If our output involves multiple classes,
Softmax can be quite helpful in predicting the odds for each class.
1. Non Linearity
2) Continuously differentiable
3) Range
When the range of the activation function is finite, gradient-based
training methods tend to be more stable, because pattern presentations
significantly affect only limited weights.
4) Monotonic
When activation functions have this property, the neural network will
learn efficiently when its weights are initialized with small random
values.
When the activation function does not approximate identity near the
origin, special care must be used when initializing the weights.
The single layer perceptron is a very simple algorithm that can only learn
linear decision boundaries. This means that it is not well suited for more
complex problems where the data is not linearly separable. Additionally,
the single layer perceptron is a batch learning algorithm, which means
that it cannot learn from new data points incrementally. The single layer
perceptron is susceptible to local minima, meaning that it may not find
the global optimum solution to a problem.
The single layer perceptron is a neural network with only one hidden
layer. It is a simple model that is used to learn how to classify data. The
single layer perceptron is limited to linearly separable data, which means
that it can only learn to classify data that can be separated by a line. This
limitation means that the single layer perceptron cannot learn to classify
data that is not linearly separable.
2. Limited Memory Capacity and Range of Capabilities
Other Limitations
Fourth, single layer perceptron is also sensitive to noise. This means that
if there is any noise in the training data, it will adversely affect the
performance of the model.
This network has three main layers that combine to form a complete
Artificial Neural Network. These layers are as follows:
Input Layer
Hidden Layer
It is the heart of all Artificial neural networks. This layer comprises all
computations of the neural network. The edges of the hidden layer have
weights multiplied by the node values. This layer uses the activation
function.
Output Layer
This layer gives the estimated output of the Neural Network. The
number of nodes in the output layer depends on the type of problem. For
a single targeted variable, use one node. N classification problem, ANN
uses N nodes in the output layer.
Each input node passes the vector input value to the hidden layer.
In the hidden layer, each edge has some weight multiplied by the
input variable. All the production values from the hidden nodes
are summed together. To generate the output