0% found this document useful (0 votes)
12 views4 pages

DL Mod 1 Final

The document discusses the structure and functioning of neural networks, including single-layer and multilayer feedforward networks, as well as recurrent networks. It explains key components such as synapses, activation functions, and the importance of bias, along with the advantages of neural networks like adaptive learning and fault tolerance. Additionally, it addresses the XOR problem, demonstrating how a multilayer perceptron can solve it through non-linear transformations.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
12 views4 pages

DL Mod 1 Final

The document discusses the structure and functioning of neural networks, including single-layer and multilayer feedforward networks, as well as recurrent networks. It explains key components such as synapses, activation functions, and the importance of bias, along with the advantages of neural networks like adaptive learning and fault tolerance. Additionally, it addresses the XOR problem, demonstrating how a multilayer perceptron can solve it through non-linear transformations.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

Models of a NeuronA neuron is a fundamental information-processing unit Network Architectures 1.

A single-layer feedforward network is a


used in neural networks. It operates based on several components that simple neural network where the neurons are arranged in layers. In this basic
collectively process signals and produce an output. The block diagram of a form, the network has:Input layer: A set of source nodes that send input
typical neuron model signals.Output layer: A set of neurons that process the inputs and produce
the output.In this structure, the input layer projects directly onto the output
layer, and there are no connections or feedback from the output layer to the
input layer. This setup is called a "feedforward" network because the
information flows only in one direction—from the input to the
output.Single-layer refers to the output layer of computation nodes
(neurons). The input layer does not perform any computation, so it isn't
counted as a "layer" in the sense of processing 2. multilayer feedforward
network, there are one or more hidden layers in addition to the input and
output layers. The hidden layers are intermediate layers that process
information between the input and output layers. These neurons are called
hidden neurons or hidden units, and they are called "hidden" because their
outputs are not directly visible to the outside of the network.Key
1. Synapses (Connecting Links)Each synapse connects an input signal Characteristics:The input signals are passed to the neurons in the first hidden
xjx_j to the neuron and is associated with a synaptic layer.The output of each hidden layer is then used as input to the next layer
weight wkjw_{kj}.The synaptic weight wkjw_{kj} (whether another hidden layer or the output layer).The output of the last
defines the strength of the connection between input layer (the output layer) is the network’s final result.Each layer is typically
xjx_j and the neuron. It can be fully connected to the next, meaning that every neuron in one layer is
positive or negative. 2. Adder (Linear Combiner)The connected to every neuron in the subsequent layer.(iii) Recurrent
input signals are summed, each weighted by its Networksdiffers from a feedforward network in that it includes feedback
corresponding synaptic weight. This summation process forms a linear loops. These feedback loops allow the output of the network to influence its
combiner.The total input signal uku_k to the neuron is given by:where mm subsequent inputs. This characteristic gives recurrent networks a type of
is the number of inputs.3. Activation FunctionThe output of the neuron is "memory" and enables them to process sequential data, making them ideal
determined by an activation function that limits the amplitude of the output for tasks involving time-dependent information.Recurrent Network with
signal.The purpose of the activation function is to "squash" or restrict the .
output to a specified range, often between 0 and 1, or between -1 and
1.BiasThe neuron also includes a bias term bkb_k, which is an external
parameter that shifts the input to the activation function.The bias can either
increase or decrease the net input, depending on
whether it is positive or negative.where vkv_k is the
adjusted input (also known as the induced local field), and the final output
is:where f(⋅)f(\cdot) represents the activation function
Rosenblatt’s Perceptron/The Perceptron Convergence TheoremTo Bayes Classifier is a probabilistic approach used to minimize the average
derive the error-correction learning algorithm for the perceptron, we begin risk of classification decisions. It utilizes the concept of prior probabilities,
by working with a modified signal-flow graph modelThis model is conditional probabilities, and associated costs of misclassification to make
equivalent to the one but with the bias b(n)b(n) treated as a synaptic weight decisions. For a two-class classification problem (classes c1c_1 and c2c_2),
driven by a fixed input the Bayes classifier computes the average risk rr, which involves integrating
over the probability distributions for each class.

The terms pX(x∣c1)p_X(x|c_1) and pX(x∣c2)p_X(x|c_2) represent the


conditional probability density functions for the two classes, while p1p_1
and p2p_2 represent the prior probabilities of the two classes. The costs
cijc_{ij} describe the penalty associated with misclassification..The decision
boundary derived from the Bayes classifier is linear in a Gaussian
environment because the likelihood ratio test simplifies to a linear function.
Specifically, the decision rule takes the form:

where Λ(x)\Lambda(x) is the likelihood


ratio. This test assigns xx to class c1c_1 if
the likelihood ratio is greater than the
threshold, and to class c2c_2 otherwise.

Advantages of Neural NetworksAdaptive Learning – Neural networks can


learn from data and improve performance over time.Fault Tolerance – Due
to distributed processing, neural networks can function even if some
connections or nodes fail.Nonlinear Processing – They can model complex,
nonlinear relationships that traditional algorithms struggle with.Pattern
Recognition – Neural networks excel at recognizing patterns in data, making
them ideal for image and speech recognition.Generalization – Once trained,
they can apply learned knowledge to new, unseen data.Self-Organization –
Neural networks can automatically organize and classify information
without explicit programming.Parallel Processing – They process multiple
computations simultaneously, making them efficient for large-scale
problems.
Linearly Separability of XOR Problem and Its SolutionIs
XOR Linearly Separable?The XOR (Exclusive OR) problem
is not linearly separable. This means that no single straight
line can divide the input space into two distinct classes (0 and
1).Example Justification:The XOR function operates as
follows:(0,0) → 0 (0,1) → 1 (1,0) → 1 (1,1) → 0If plotted
on a 2D plane, the outputs for (0,0) and (1,1) belong to one
class, while (0,1) and (1,0) belong to another. These points are
positioned diagonally, making it impossible to separate them
with a straight line.Base Model for Solving XORA Single-
Layer Perceptron cannot solve the XOR problem because it
uses a linear decision boundary. However, a Multilayer
Perceptron (MLP) with at least one hidden layer can handle
the XOR function by introducing non-linearity.How
MultiLayerPerpetron Solves the XOR Problem: Network
Architecture:Input Layer: Two neurons (x₁, x₂) representing
binary inputs.Hidden Layer: Two neurons that enable non-
linear feature transformations.Output Layer: One neuron that
classifies the output.Working Mechanism:The hidden neurons
transform the input space by introducing non-linearity.The
output neuron combines the results from the hidden layer to
generate the final decision.The non-linear activation function
(sigmoid/ ReLU) helps separate the XOR outputs correctly.
learning.Recommended parameters: 𝑎=1.7159, 𝑏=23Target
Valuesshould match the range of the activation
function.For tanh, set targets close to its limits (e.g.,
±1.7159) to prevent saturation and improve
convergence.Normalizing InputsRemove mean and
standardize inputs to have a variance of 1.Use Principal
Component Analysis (PCA) to decorrelate inputs.Scale
inputs so neurons learn at the same rate, improving
efficiency.Weight InitializationAvoid extreme weight
values to prevent neuron saturation or slow learning.For
tanh, initialize weights with variance 1/n, where n is the
Heuristics for Improving Back-Propagation number of inputs.Learning from HintsUse prior knowledge
PerformanceStochastic vs. Batch (e.g., symmetries, constraints) to guide learning.Proper initialization based
UpdateStochastic Update: Updates weights on known properties can speed up convergence.Learning Rate
after each training example, making learning AdjustmentLater layers should have smaller learning rates as gradients
faster and efficient for large datasets.Batch diminish.Neurons with more inputs should have lower learning rates to
Update: Updates weights after processing the balance learning across the network.
entire dataset but can be computationally feedback path. This closed-loop structure allows the system to adjust
intensive.Maximizing Information dynamically based on its own outputs, making it more adaptable to changing
ContentSelect training examples that cause the largest errors to drive more inputs. Feedback is crucial in recurrent neural networks (RNNs), where past
effective learning.Use diverse examples to explore more of the weight outputs influence future computations, enabling the modeling of sequential
space.Shuffle data between epochs to prevent bias and improve data and temporal dependencies.
generalization.Activation FunctionThe tanh function is preferred as it
outputs values between -1 and 1, has a steep slope near zero, and enhances

You might also like