Understanding ResNet
Understanding ResNet
• Vanishing/exploding gradients
• Degradation problem: accuracy gets worse with deeper
networks
• Image/graph showing performance drop beyond certain
layers
Deep Convolutional Neural Network
• The problem with these deep neural network were as you increase
the layer we start seeing degradation problem. Or to put it in another
word as we increase depth of the network the accuracy gets
saturated and starts degrading rapidly. In a deep neural network as
we perform back-propogation, repeated mulitplication for finding
optimal solution makes gradient very small which result in
degradation. This problem is often called vanishing
gradient/exploding gradient.
What is ResNet?
• Introduced by Microsoft Research in 2015
• Won ImageNet 2015
• Concept: Residual Learning
• Allows training of networks with 100+ layers
Residual Block Explained
• ResNet solve this degradation problem, is by skipping
connection or layer. Skipping connection means,
consider input x and this input is passed through stack
of neural network layers and produce f(x) and this f(x) is
then added to original input x.So our ouput will be:
• Equation: F(x) + x
ResNet Architecture
ResNet Architecture Variants
• ResNet-18, ResNet-34, ResNet-50, ResNet-101, ResNet-
152
• Difference in depth and use of bottleneck blocks
• Table comparing number of layers and parameters
ResNet Architecture Table
Applications of ResNet
• Image classification
• Object detection (used in Faster R-CNN, Mask R-CNN)
• Medical imaging, facial recognition, etc.
Why ResNet Works?
• Solves vanishing gradient with skip connections
• Enables extremely deep networks to converge
• Simpler training, better generalization