Unit-2 L1
Unit-2 L1
Regularization
• Regularization: Regularization: Overview, Parameter Penalties,
Norm Penalties as Constrained Optimization, Regularization and
Underconstrained Problems, Data Augmentation, Noise Robustness,
Batch Normalization, Semi-Supervised Learning, Multi-Task
Learning, Early Stopping, Parameter Tying and Parameter Sharing,
SparseRepresentations, Bagging, Dropout. Tuning Neural Networks,
Hyperparameters
Regularization
• As we increase capacity, training error decreases, but the gap between training
and generalization error increases. Eventually, the size of this gap outweighs
the decrease in training error, and we enter the overfitting regime, where
capacity is too large, above the optimal capacity.
Regularization (Overview)
• There are three situations when a deep learning model is trained
1. First regime is,excluded the true data generating process—corresponding to
underfitting and inducing bias, or
2. Matched the true data generating process (Just fit) (desired), or
3. Included the generating process but also many other possible generating
processes—the overfitting regime where variance rather than bias dominates the
estimation error.
• The goal of regularization is to take a model from the third regime into the
second regime.
• The best-fitting model (in the sense of minimizing generalization error) is a
large model that has been regularized appropriately.
• In regularization technique, we reduce the magnitude of the
features by keeping the same number of features.”
Overfitting occurs when our deep learning model tries to cover all
the data points or more than the required data points present in the
given dataset. The overfitted model has low bias and high variance.