Rank of Neural Network Architectures
Rank of Neural Network Architectures
1 Perceptron 1958
5 Backpropagation 1986
1980s (re-popularized
7 Autoencoders
in 2006)
1990s (popularized in
11 Bayesian Neural Networks
2018)
1990s (popularized in
12 Spiking Neural Networks (SNNs)
2018)
1993 (popularized in
13 Siamese Networks
2015)
Rank Architecture Founding Date
2009 (popularized in
19 Graph Neural Networks (GNNs)
2017)
2016 (popularized in
28 HyperNetworks
2019)
30 Transformers 2017
Self-Attention Generative
33 2018
Adversarial Networks (SAGANs)
Various (popularized
38 Reinforcement Learning Networks
in 2015 with DQNs)
This table organizing the neural network architectures
based on their popularity:
Convolutional
Widely used in image and video
Neural
1 recognition, classification, and
Networks
segmentation.
(CNNs)
Recurrent
Popular for sequence data, including time
Neural
3 series prediction and language modeling
Networks
(especially LSTMs and GRUs).
(RNNs)
Generative
Adversarial Known for generating realistic images,
4
Networks videos, and other types of data.
(GANs)
Residual
Essential for training very deep neural
5 Networks
networks.
(ResNets)
Densely
Connected Popular for their efficiency in learning
6
Networks features with dense connections.
(DenseNets)
9
Rank Architecture Popularity Characteristics
Feedforward
Fundamental architecture, widely used as
Neural
10 the building block for more complex
Networks
networks.
(FNNs)
Neural
Becoming popular for automating the
13 Architecture
design of neural network architectures.
Search (NAS)
Neural ODEs
(Ordinary Emerging popularity for modeling
15
Differential continuous-time dynamics.
Equations)
Spiking Neural
Popular in neuromorphic computing for
17 Networks
low-power, real-time processing tasks.
(SNNs)
Self-Attention
Generative
Enhanced version of GANs with self-
19 Adversarial
attention mechanisms.
Networks
(SAGANs)
Spatial
Useful for enhancing the spatial
Transformer
22 manipulation capabilities of neural
Networks
networks.
(STNs)
Radial Basis
Function Used in specific applications like function
23
Networks approximation and time-series prediction.
(RBFNs)
Extreme
Learning Known for fast training, but less popular
24
Machines in mainstream deep learning.
(ELMs)
Echo State
Used in reservoir computing, with niche
25 Networks
applications.
(ESNs)
Liquid State
Similar to ESNs, used in specific areas of
26 Machines
reservoir computing.
(LSMs)
Hierarchical
Inspired by biological brains, used in
30 Temporal
niche applications.
Memory (HTM)
Adaptive
Computation Niche use cases where dynamic
33
Time Networks computation is beneficial.
(ACT)
Hierarchical
Research interest in tasks requiring
34 Attentive
hierarchical attention over sequences.
Memory (HAM)
Feedforward
Neural Simple and computationally efficient,
1
Networks especially for smaller datasets.
(FNNs)
Radial Basis
Efficient for specific tasks like function
Function
2 approximation and time-series prediction,
Networks
but limited in scalability.
(RBFNs)
Convolutional
Neural Efficient for image processing tasks due to
4
Networks parameter sharing and local connectivity.
(CNNs)
Neural ODEs
(Ordinary Efficient for modeling continuous-time
8
Differential dynamics with fewer parameters.
Equations)
Rank Architecture Efficiency Characteristics
Densely
Efficient in feature reuse, reducing the
Connected
9 number of parameters compared to other
Networks
deep networks.
(DenseNets)
Extreme
Learning Fast training times due to random hidden
10
Machines nodes, but less flexible.
(ELMs)
Graph Neural
Efficient for processing graph-structured
13 Networks
data by leveraging sparsity.
(GNNs)
Residual
Efficient in training very deep networks by
14 Networks
mitigating the vanishing gradient problem.
(ResNets)
Recurrent
Efficient for sequence data, with LSTMs
Neural
17 and GRUs being more efficient than vanilla
Networks
RNNs.
(RNNs)
18
Rank Architecture Efficiency Characteristics
Echo State
Efficient in reservoir computing with fixed
Networks
recurrent layers.
(ESNs)
Liquid State
Similar efficiency to ESNs in reservoir
19 Machines
computing applications.
(LSMs)
Generative
Efficient in generating high-quality
Adversarial
22 samples, but require careful tuning and
Networks
can be computationally expensive.
(GANs)
Self-Attention
Generative More efficient than standard GANs in
23 Adversarial capturing long-range dependencies but
Networks computationally demanding.
(SAGANs)
Spiking Neural
Efficient for low-power and real-time
26 Networks
processing, but challenging to train.
(SNNs)
Dynamic
Efficient in adaptive scenarios, but with
27 Neural
variable computational costs.
Networks
Rank Architecture Efficiency Characteristics
Adaptive
Computation Efficient for tasks requiring variable
30
Time Networks computation, but complex to implement.
(ACT)
Spatial
Efficient in enhancing spatial
Transformer
32 manipulation, but with added
Networks
computational cost.
(STNs)
Deep Belief
Historically significant but less efficient
34 Networks
compared to modern architectures.
(DBNs)