ITNN Week3
ITNN Week3
POST
GRADUATE
PROGRAM
AIML ARTIFICIAL INTELLIGENCE & MACHINE
LEARNING
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
1
Brain Stormer
Q1. Back Propagation is a learning technique that adjusts weights in neural network by propagating weight
changes.
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
Week 3: Introduction to neural networks and deep learning
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
3
Learning Objective
❖ Types of Optimizers
❖ Weight initialization
❖ Regularization
❖ Drop out
❖ Batch Normalization
❖ Types of neural networks
❖ Case study
❖ Questions
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
4
Different types of optimizers
1. SGD with Momentum
This method computes gradient by exponentially weighted averages, hence it takes less time to converge compared to normal
stochastic gradient descent
4. ADAM
ADAM proposes the characteristics of both SGD with Momentum and RMSprop
References
Paperspace, medium, Medium
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
5
Weight Initialization
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
6
Why Initialize Weights
The aim of weight initialization is to prevent layer activation outputs from exploding or vanishing during the course of a
forward pass through a deep neural network. If either occurs, loss gradients will either be too large or too small to flow
backwards beneficially, and the network will take longer to converge, if it is even able to do so at all.
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
7
What Happends When W=0 Init Is Used
Output Layer
Input Layer
Hidden Layer
The method of setting W=0 serves almost no purpose as it causes neurons to perform the same calculation
in each iterations and produces same outputs. neurons will learn same features in each iterations.
This problem is known as network failing to break symmetry.
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
8
Initialization Techniques
● Zero initialization
● Random initialization
● Xavier initialization
● He initialization
● Kaiming initialization
● And many more
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
9
Regularization
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
10
Data Augmentation
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
11
Where We Should Do Data Augmentation
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
12
Data Augmentation Pipeline
Load image and label
“Dog ”
Compute
loss
CNN
Transformimage
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
13
Data Augmentation Techniques
● Horizontal flips
● Rotation
● Crop/scale
● Color jitter
● Other creative techniques
○ Random mix/combinations of :
■ translation (what about a pure ConvNet?)
■ Rotation
■ Stretching
■ Shearing
■ lens distortions
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
14
Dropout
Dropout is a regularization method that approximates training a large number of neural networks with different
architectures in parallel.
During training, some number of layer outputs are randomly ignored or “dropped out.” This has the effect of making the
layer look-like and be treated-like a layer with a different number of nodes and connectivity to the prior layer. In effect,
each update to a layer during training is performed with a different “view” of the configured layer.
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
15
Dropout
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
16
Dropout – How It Works
has an ear X
has a tail
is furry X cat
score
has claws
mischievous X
look
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
17
Dropout
Another interpretation:
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
18
Batch Normalization
•Due to this normalization “layers” between each fully
connected layers, the range of input distribution of each layer
stays the same, no matter the changes in the previous layer
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
19
Types of Neural Network
Feed Forward
Neural
Network
Convolutional
Neural
Network
Recurrent
Neural
Network
LSTM – Long
Short-Term
Memory
Sequence to
Sequence
Models
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited
20
Thank you! :)
Questions are always welcome
Proprietary content. ©Great Learning. All Rights Reserved. Unauthorized use or distribution prohibited