Week 8
Week 8
1. What are the challenges associated with using the Tanh(x) activation function?
2. Which of the following problems makes training a neural network harder while using
sigmoid as the activation function?
(a) Not-continuous at 0
(b) Not-differentiable at 0
(c) Saturation
(d) Computationally expensive
4. We have observed that the sigmoid neuron has become saturated. What might be
the possible output values at this neuron?
(a) 0.0666
(b) 0.589
(c) 0.9734
(d) 0.498
(e) 1
6. Which of the following are common issues caused by saturating neurons in deep
networks?
7. Given a neuron initialized with weights w1 = 0.9, w2 = 1.7, and inputs x1 = 0.4,
x2 = −0.7, calculate the output of a ReLU neuron.
Correct Answer: 0
Solution: The weighted sum is 0.9 × 0.4 + 1.7 × (−0.7) = 0.36 − 1.19 = −0.83. ReLU
outputs the max of 0 and the input, so the result is max(0, −0.83) = 0.
8. Which of the following is incorrect with respect to the batch normalization process
in neural networks?
(a) We normalize the output produced at each layer before feeding it into the next
layer
(b) Batch normalization leads to a better initialization of weights.
(c) Backpropagation can be used after batch normalization
(d) Variance and mean are not learnable parameters.
10. How can you tell if your network is suffering from the Dead ReLU problem?