Deep Vs Shallow Neural Networks
Deep Vs Shallow Neural Networks
shallow
neural networks
Presented by:
P.Narmatha shree
Final year CSE
Shallow neural
networkS
Shallow neural networks are comparatively simpler and
easier to train. They are suitable for straightforward tasks
such as data fitting, classification, and pattern
recognition. Examples include logistic regression,
support vector machine, random forest, and cluster
analysis. Shallow networks can be trained quickly with
fewer parameters, making them faster to train and
requiring less computational resources.
deep neural
network
Due to their fewer parameters and DNNs, with their large number of
simpler architecture, SNNs have a parameters and multiple layers, are
lower risk of overfitting. more prone to overfitting. The high
Overfitting occurs when a model capacity of DNNs allows them to fit the
learns the noise in the training training data very closely, which can
data rather than the underlying lead to overfitting if not managed
properly. Techniques such as
pattern, leading to poor
regularization, dropout, and early
generalization to new data. SNNs
stopping are often used to mitigate
are less likely to overfit as they
overfitting in DNNs.
have limited capacity to memorize
the training data.
Data Requirements
Shallow Neural Networks (SNNs): Deep Neural Networks (DNNs):
SNNs generally require less data DNNs require large amounts of data to
to train effectively. Their simpler train effectively. The multiple layers
architecture means they need and vast number of parameters mean
fewer examples to learn the that DNNs need extensive datasets to
patterns and relationships in the learn and generalize well. In many
data. However, this also limits cases, the performance of a DNN
their ability to handle complex improves as the size of the training
tasks that require a deeper data increases.
understanding of the data.
Parameter Count
Shallow Neural Networks (SNNs): Deep Neural Networks (DNNs):
Shallow Neural network with few Deep Neural network with many layers
layers (usually 1 hidden layer). (multiple hidden layers).
Fewer parameters counts in the shallow Many more parameters counts in the deep neural
neural networks. networks.
Requires less computational resources. Requires more computational resources (e.g., GPUs).