LECTURE#9 EE258 F22 Part2 Draft v1
LECTURE#9 EE258 F22 Part2 Draft v1
LECTURE#9 EE258 F22 Part2 Draft v1
(Part- 2)
Based on Chapter 7 of Deep Learning textbook
0
What is Regularization?
• How to make an algorithm perform well not only on the
training data but also on new inputs?
• Strategies designed to reduce test error = REGULARIZATION
optimal
/
[ huge GAP --~-→ overfitting
1
Regularization Methods
LI :rlw)=HwH ,
/
• 7.4 Dataset Augmentation ☆ regularization parameter
• 7.5 Noise Robustness
• 7.6 Semi-supervised Learning
• 7.7 Multitask Learning
• 7.8 Early Stopping
• 7.9 Parameter Sharing (will be discussed in CNN)
• 7.11 Bagging and Ensemble Models
• 7.12 Dropout
2
Dataset Augmentation
• It is always better to train a model using more data, but data
is in general limited.
" "
3
Data Augmentation for Object Recognition
Object Recognition Problem: detect the animal
Inputs
Rotate
180o
Outputs
Lion Images are from mygreatlearning.com 4
Data Augmentation for Speech Recognition
Add Noise
" "
sound of off
5
Noise Robustness
-
lecture
• Adding Noise to Inputs
– Discussed in data augmentation ¥wi¥s
– It was shown that this is equivalent to penalty on the norm under
certain conditions (Bishop 95)
Remen n→
• Adding Noise to Weights
-
Update egn
: w→cn 1)
+ = ch
217J
- +
Gaussian
noise
#
6
Semi-supervised Learning
In
y
X Unsupervised Supervised
Learning (PCA, Learning
Better representation of inputs
Kmeans)
(a version of X)
• Semi-supervised learning:
Project Ii
human or animalca
Task-
specific
Trade-offs:
– Space for storing the parameters
– Some training data is used for validation
• Retrain again with all the data and use early stopping point from previous
training
• Continue training with all the data check if error decreases below the
early stopping point
9
Parameter Sharing
• Force constraints on the parameter è be equal to each other
U2
èREDUCES THE NUMBER OF PARAMETERS
④ 3)
WL
→
v3
v3
10
Bagging (an ensemble model)
of
• Bagging:
"
– Regression
n=3
[ →
^
2
Egging )
f
i – Classification
↑ "
% (F) Majority
=
Rule
{ Fit :*}
,
BOOTSTRAPPING
-
11
Other Ensemble Models
Combine predictions
I see
prey
.
slide )
12
Dropout
• Definition: A practical way of implementing a version of
bagging in NN for a large set of models sharing parameters
REMOVE
NON-OUTPUT
UNITS