An Introduction To Mathematics Behind Neural Networks
An Introduction To Mathematics Behind Neural Networks
An Introduction To Mathematics Behind Neural Networks
The row vectors of the inputs and weights are x = [x₁, x₂, … ,
xₙ] and w =[w₁, w₂, … , wₙ] respectively and their dot
product is given by
Step 2 : In order to find the best weights and bias for our
Perceptron, we need to know how the cost function changes
in relation to weights and bias. This is done with the help
the gradients (rate of change) — how one quantity changes
in relation to another quantity. In our case, we need to find
the gradient of the cost function with respect to the weights
and bias.
Let’s calculate the gradient of cost function C with respect to
the weight wᵢ using partial derivation. Since the cost function
is not directly related to the weight wᵢ, let’s use the chain
rule.
Let’s start with the gradient of the Cost function (C) with
respect to the predicted value ( ŷ )
Now let’s find the the gradient of the predicted value with
respect to the z. This will be a bit lengthy.
Therefore we get,
Conclusion