Lecture 08 ML
Lecture 08 ML
Mohamed Yasser
1) Training model
2) Computational complexity
3) Gradient descent
4) Batch gradient descent
Topics 5) Stochastic gradient
descent
6) Mini batch gradient
descent
7) Polynomial Regression
7
Formula
9
Types of gradient descent
10
Batch gradient descent
In Batch Gradient Descent, the entire
training dataset is used to compute the
gradient of the cost function with respect
to the model parameters in each iteration.
The algorithm calculates the average
gradient for the entire dataset and then
updates the model parameters.
11
Stochastic gradient descent
In Stochastic Gradient Descent, the model
parameters are updated after computing
the gradient of the cost function with
respect to the parameters for each
training example individually. The key
difference from Batch Gradient Descent is
that it uses only one training example at a
time for the update.
13
Mini batch gradient descent
Mini-Batch Gradient Descent is a
compromise between Batch Gradient
Descent and Stochastic Gradient Descent.
Instead of using the entire training
dataset (Batch GD) or just one example
(SGD), Mini-Batch GD processes a small
random subset of the training data (a
mini-batch) in each iteration.
15
Polynomial regression
17
Thank you
20