00005187-Deep Learning
00005187-Deep Learning
LEARNING
Presented By .
1. Shashikiran Taur
2. Shubham Gautam
3. Rupesh Bavakar
Guided by.
4. Sneha Dubal
- Pranit Sir
AJENDA
▪ BSF (Binary Step Function)
▪ Chain Rule
▪ Optimizer
BSF (Binary Step Function)
➢ In deep learning, a binary step
function is a type of activation
function used in artificial neural
1 networks.
➢ It is a simple non-linear
0
function that maps the input to
either 0 or 1 based on a
threshold.
The binary step function was one of the earliest activation functions
used in artificial neural networks, dating back to the 1950s and 1960s.
Student Criterion (OR GATE)
Threshold >= 1
Hidden Layer:
The hidden layer presents in-between
input and output layers. It performs all
the calculations to find hidden features
and patterns.
Output Layer:
Backward Propagation
The input goes through a series of
transformations using the hidden layer,
which finally results in output that is
conveyed using this layer.
Weights & Bias
Weights:
Weight are values that are assigned to the connections between
neurons in the network. Each connection between two neurons has
a weight associated with it that determines the strength of the
connection.
Bias:
Bias is used for shifting the activation function towards left or right,
you can compare this to y-intercept in the line equation.
Chain Rule Matrix 1 Matrix 2 Matrix 3
❖ Gradient Descent:
This is a simple optimization algorithm that adjusts the weights and
biases of the network by moving in the opposite direction of the
gradient of the loss function.
Convergence Zone
Backward
Forward Propagation Propagation
When to Stop:
1. Early Stoppage:
epoch = 10
epoch = 5
W new = W old
Disadvantages Of Gradient Descent
▪ Weight Updation Takes Time.
▪ Computationally Expensive.
▪ Takes Time to converge.