18 - Computational Complexity
18 - Computational Complexity
Model Size
FLOPs
FLOPS
MACs
Inference Time
Learnable Parameters
weights and biases in the model
Model parameters influence the model's ability to generalize and make accurate
predictions
When the model parameters are set optimally, the model fits the training data well
and generalizes to unseen data effectively
If the parameters are poorly chosen, the model may overfit or underfit
Data availability
More data the more complex a model to use
Insufficient data with a complex model lead to overfitting
Model Complexity
Simple problems addressed by less complex models
Complex problems requires large number of parameter
Computational Resources
Training models with a large number of parameters computationally
expensive
Having limited computational resources, need to use smaller models
If all weights are stored with 32-bit numbers, total storage will be about
61M × 4 Bytes (32 bits) = 224 MB (224 × 10° Bytes)
If all weights are stored with 8-bit numbers, total storage will be about
61M × 1 Byte (8 bits) = 61 MB
• MACs = 96 X 3 X 11 X 11 X 55 X 55
=105,415,200
g = number of groups
Pooling
-subsampling layers
Separable Convolutions
-don't change the depth, reducing the number of FLOPs
-a pointwise convolution is a 1x1 convolution
Model Pruning
-redundant network parameters are removed
Weight Sharing
share the weights between neuron
so we have less of them to store
Thank You