Mod 2.3 - Activation Function, Loss Functions
Mod 2.3 - Activation Function, Loss Functions
The function doesn't do anything to the weighted sum of the input, it simply
spits out the value it was given.
Because of its limited power, this does not allow the model to create
complex mappings between the network’s inputs and outputs.
.
This function takes any real value as input and outputs values in the range
of 0 to 1.
The larger the input (more positive), the closer the output value will be to
1.0, whereas the smaller the input (more negative), the closer the output will
be to 0.0, as shown below.
Sigmoid/Logistic Activation Function
Reasons
It is commonly used for models where we have to predict the
probability as an output. Since probability of anything exists only
between the range of 0 and 1, sigmoid is the right choice because of
its range.
The function is differentiable and provides a smooth gradient, i.e.,
preventing jumps in output values. This is represented by an S-shape
of the sigmoid activation function.
ReLU Function
ReLU stands for Rectified Linear Unit.
The main catch here is that the ReLU function does not activate all the neurons
at the same time.
The neurons will only be deactivated if the output of the linear transformation is
less than 0.
ReLU Activation Function
This is where loss functions come in. Loss functions are one of the
most important aspects of neural networks, as they (along with the
optimization functions) are directly responsible for fitting the model
to the given training data
Loss Functions Overview
One of the most popular loss functions, MSE finds the average of the
squared differences between the target and the predicted outputs
MAE finds the average of the absolute differences between the target
and the predicted outputs.