Neural NetworksChapter2Sup
Neural NetworksChapter2Sup
Introduction
Artificial Neural Networks (ANNs) are computational models designed to mimic
the functioning of the human brain. They are widely used in pattern
recognition, data classification, and decision-making tasks. This chapter
introduces the fundamental concepts, evolution, models, learning methods,
activation functions, and essential terminologies related to ANNs.
o
4. Single-layer recurrent network
o In this model, neurons are connected in a way that allows
feedback loops within the layer. This means the output of some
neurons can influence their own inputs or those of other neurons
in the same layer.
o
2.5 Learning Methods in Neural Networks
2.5.1 Supervised Learning:
• The network learns from labeled training data.
• Example: Image classification (labels such as "cat" or "dog").
• Uses techniques like Backpropagation and Gradient Descent.
•
2.6 Activation Functions and Their Equations
Activation functions determine how the weighted sum of inputs is
transformed into an output signal. Here are the commonly used activation
functions:
• Equation:
• Characteristics:
o Used in linear regression and some output layers.
o No non-linearity, making it unsuitable for complex problems.
• Equation:
•
• Characteristics:
o Used in simple binary classification tasks.
o where thetha represents the threshold value. This function is
most widely used in single-layer nets to convert the net input to
an output that is a binary (1 or 0).
• Equation:
• where thetha represents the threshold value. This function is also used
in single-layer nets to convert the net input to an output that is
bipolar(+ 1 or -1).
•
• Characteristics:
o Used in threshold-based decision-making tasks.
o Lacks smooth transitions, making it unsuitable for deep learning.
•
•
• Characteristics:
• If the nerwork uses a binary data, it is better to convert it to bipolar
form and use the bipolar sigmoidal activation function or hyperbolic
tangent function.
o Introduces non-linearity.
o Used in probabilistic models and logistic regression.
o Suffers from the vanishing gradient problem, making deep
networks difficult to train.
• Equation:
• Characteristics:
o Efficient for deep learning due to its simple derivative.
o Helps mitigate the vanishing gradient problem.
o Can suffer from the dying ReLU problem, where neurons become
inactive for negative inputs.
Overview
• Introduced in 1943 by Warren McCulloch and Walter Pitts.
• First mathematical model of a neuron, based on binary threshold logic.
• It is a simple feedforward neural network that performs logic operations.
• Used to model basic decision-making processes.
Key Characteristics
• Binary Inputs and Outputs: Only 0 or 1.
• Threshold-Based Decision: Output is 1 if the weighted sum is above the
threshold.
• Performs Logical Operations: AND, OR, NOT operations can be
implemented.
• Limitations: Cannot solve non-linearly separable problems like XOR.
2.9 Hebbian Neural Model
Overview
• Introduced in 1949 by Donald Hebb.
• Based on the principle "Neurons that fire together, wire together."
• It strengthens connections between neurons that activate
simultaneously.
Key Characteristics
• Hebb's Rule: If two neurons activate together, their connection
strength increases.
• Self-organizing learning: No external supervision is needed.
• Used in associative learning: Helps in memory storage and pattern
recognition.
• Limitations: No weight decay, leading to unbounded growth of weights.
2.11 Summary
• ANNs are inspired by biological neural networks.
• They consist of layers of neurons with weighted connections.
• Learning occurs through weight adjustments.
• Activation functions define neuron output behavior.
• Different models and learning techniques are used for various
applications.