Chapter 6 Deep Learning Knowledge
Chapter 6 Deep Learning Knowledge
▪ The sigmoid function limits the node’s outputs to the unity regardless
of the input’s magnitude. In contrast, the ReLU function does not
exert such limits.
▪ Isn’t it interesting that such a simple change resulted in a drastic
improvement of the learning performance of the deep neural
network? Another element that we need for the back-propagation
algorithm is the derivative of the ReLU function.
▪ By the definition of the ReLU function, its derivative is given as: