Deep Learning L3
Deep Learning L3
Image Augmentation
1. Geometric transformations: randomly flip, crop, rotate,
stretch, and zoom images. You need to be careful
about applying multiple transformations on the same
images, as this can reduce model performance.
2. Color space transformations: randomly change RGB
color channels, contrast, and brightness.
3. Kernel filters: randomly change the sharpness or
blurring of the image.
4. Random erasing: delete some part of the initial image.
5. Mixing images: blending and mixing multiple images.
Recurrent Neural Networks
CNN
RNNs
1. Vanishing Gradient
2. Exploding Gradient
Vanishing Gradient
Where the contribution from the earlier steps becomes insignificant in the
gradient descent step.
Exploding Gradients
The main advantage of RNN over ANN is that RNN can model sequence of data
(i.e. time series) so that each sample can be assumed to be dependent on
previous ones
Recurrent neural network are even used with convolutional layers to extend the
effective pixel neighborhood.
Disadvantages of Recurrent Neural Network
The key to LSTMs is the cell state, the horizontal line running through the top of the
diagram. The cell state is kind of like a conveyor belt. It runs straight down the
entire chain, with only some minor linear interactions. It’s very easy for information
to just flow along it unchanged.
LSTM
The LSTM does have the ability to remove or add
information to the cell state, carefully regulated by
structures called gates.