Activation+Functions
Activation+Functions
Without activation function, weight and bias would only have a linear transformation, or neural network is just a
linear regression model, a linear equation is polynomial of one degree only which is simple to solve but limited in
But opposite to that, the addition of activation function to neural network executes the non-linear transformation
to input and make it capable to solve complex problems such as language translations and image classifications.
In addition to that, Activation functions are differentiable due to which they can easily implement back
propagations, optimized strategy while performing backpropagations to measure gradient loss functions in the
neural networks.
Different Activation Functions
Sigmoid Activation Function
The main reason why we use sigmoid function is because it exists between (0 to 1).
Therefore, it is especially used for models where we have to predict the probability as
sigmoid is the right choice.The function is differentiable.That means, we can find the
slope of the sigmoid curve at any two points.The function is monotonic but function’s
derivative is not.The logistic sigmoid function can cause a neural network to get stuck