Ann Project Assignment
Ann Project Assignment
Supervisor: DR KS OJO
PROJECT ASSIGNMENT
22/11/2024
Artificial Neural Networks (ANNs) are designed to mimic the way the human brain
processes information. They use interconnected nodes (neurons) and layers to simulate the
brain's neural network. Here is a bit more detail on how this works:
1. Neurons: The basic units of a neural network, like neurons in the brain. Each neuron
receives input, processes it, and passes the output to the next layer.
o Hidden Layers: Intermediate layers that process the data through weighted
connections.
Learning Process
generate an output.
2. Loss Calculation: The difference between the predicted output and the actual output
is measured.
Brief history
Although the study of the human brain is thousands of years old. The first step towards
neural networks took place in 1943, when Warren McCulloch, a neurophysiologist, and a young
mathematician, Walter Pitts, wrote a paper on how neurons might work. They modelled a simple neural
In 1949, Donald Hebb reinforced the concept of neurons in his book, The Organization of Behaviour. It
pointed out that neural pathways are strengthened each time they are used.
In the 1950s, Nathanial Rochester from the IBM research laboratories led the first effort to simulate a
neural network.
In 1956 the Dartmouth Summer Research Project on Artificial Intelligence provided a boost to both
artificial intelligence and neural networks. This stimulated research in AI and in the much lower-level
In 1957, John von Neumann suggested imitating simple neuron functions by using telegraph relays or
vacuum tubes.
In 1958, Frank Rosenblatt, a neurobiologist of Cornell, began work on the Perceptron. He was intrigued
with the operation of the eye of a fly. Much of the processing which tells a fly to flee is done in its eye.
The Perceptron, which resulted from this research, was built in hardware and is the oldest neural
network still in use today. A single-layer perceptron was found to be useful in classifying a continuous-
valued set of inputs into one of two classes. The perceptron computes a weighted sum of the inputs,
subtracts a threshold, and passes one of two possible values out as the result.
In 1959, Bernard Widrow and Marcian Hoff of Stanford developed models they called ADALINE and
MADALINE. These models were named for their use of Multiple Adaptive Linear Elements.
MADALINE was the first neural network to be applied to a real-world problem. It is an adaptive filter
which eliminates echoes on phone lines. This neural network is still in commercial use.
Adetayo 3
Marvin Minsky & Seymour Papert proved the Perceptron to be limited in their book, Perceptrons
Progress on neural network research halted due fear, unfulfilled claims, etc. until 1981. This caused
respected voices to critique the neural network research. The result was to halt much of the funding. This
1982 — John Hopfield presented a paper to the national Academy of Sciences. His approach to create
1982- US-Japan Joint Conference on Cooperative/ Competitive Neural Networks at which Japan
announced their Fifth-Generation effort resulted US worrying about being left behind. Soon funding
1985 — American Institute of Physics began what has become an annual meeting — Neural Networks
for Computing. By 1987, the Institute of Electrical and Electronic Engineer’s (IEEE) first International
In 1997 — A recurrent neural network framework, Long Short-Term Memory (LSTM) was proposed
several other steps have been taken to get us to where we are now; today, neural networks discussions
are prevalent; the future is here! Currently most neural network development is simply proving that the
principal works. This research is developing neural networks that, due to processing limitations, take
weeks to learn.
Reservoir computing
network theory. Wolfgang Maass independently proposed it, Thomas Nat schlager, and
Henry Markram in 2002, and by Herbert Jaeger in 2001. The concept involves using a fixed,
random, and recurrent neural network called the reservoir, which transforms input signals
Adetayo 4
into a high-dimensional dynamic state. The key advantage is that only the output layer is
Reservoir Computing has its roots in earlier neural network architectures like
Liquid State Machines (LSMs) and Echo State Networks (ESNs). These models
demonstrated that randomly connected recurrent neural networks could be used for tasks such
as interval and speech discrimination. The framework has since been applied to various
incorporates physical laws described by partial differential equations into the training
process. This approach ensures that the network's predictions adhere to known physical
principles, making them highly accurate for scientific and engineering applications.
The concept of PINNs emerged from the need to solve complex physical problems
where traditional numerical methods fall short. By embedding physical laws into the neural
network, PINNs can enhance the information content of the available data, facilitating the
learning algorithm to capture the right solution even with a low amount of training examples.
This approach has been particularly useful in fields like fluid dynamics, where the Navier-
Both Reservoir Computing and PINNs represent significant advancements in the field
simplifying the training process and leveraging the dynamic properties of the reservoir. This
makes it particularly effective for tasks involving temporal patterns and dynamic systems.
Adetayo 5
On the other hand, PINNs enhance the robustness and accuracy of ANNs by
incorporating physical laws into the training process. This integration allows ANNs to solve
complex physical problems more effectively, bridging the gap between data-driven and
physics-based modeling.
The ability of a neural network to ‘think’ has revolutionized computing as we know it.
These smart solutions are capable of interpreting data and accounting for context.
Four critical steps that neural networks take to operate effectively are:
computer is shown an unfamiliar pattern, it will associate the pattern with the closest
• Prediction, or the production of expected results using a relevant input, even when all
Neural networks require high throughput to carry out these functions accurately in
near real-time. This is achieved by deploying numerous processors to operate parallel to each
The neural networking process begins with the first tier receiving the raw input data.
You can compare this to the optic nerves of a human being receiving visual inputs. After that,
each consecutive tier gets the results from the preceding one. This goes on until the final tier
Every individual processing node contains its database, including all its past learnings
and the rules that it was either programmed with originally or developed over time. These
The learning process (also known as training) begins once a neural network is
structured for a specific application. Training can take either a supervised approach or an
unsupervised approach. In the former, the network is provided with correct outputs either
through the delivery of the desired input and output combination or the manual assessment of
network performance. On the other hand, unsupervised training occurs when the network
allows machine learning algorithms to be modified as they learn from their training and
subsequent operations. Learning models are fundamentally centered around the weightage of
input streams, wherein, each node assigns a weight to the input data it receives from its
preceding nodes. Inputs that prove instrumental to deriving the correct answers are given
Apart from adaptability, neural networks leverage numerous principles to define their
Fuzzy logic, gradient-based training, Bayesian methods, and genetic algorithms all
play a role in the decision-making process at the node level. This helps individual nodes
decide what should be sent ahead to the next tier based on the inputs received from the
preceding tier.
Basic rules on object relationships can also help ensure higher quality data modeling.
For instance, a facial recognition neural network can be instructed ‘teeth are always below
the nose’ or ‘ears are on each side of a face’. Adding such rules manually can help decrease
training time and aid in the creation of a more efficient neural network model.
Adetayo 7
However, the addition of rules is not always a good thing. Doing so can also lead to
incorrect assumptions when the algorithm tries to solve problems unrelated to the rules.
Preloading the wrong ruleset can lead to the creation of neural networks that provide
While neural networking, and especially unsupervised learning, still have a long way
than we think. It is a fact that the connections within a neural network are nowhere as
numerous or efficient as those in the human brain. However, Moore’s Law, which states that
the average processing power of computers is expected to double every two years, is still
flourishing. This trend gives our expectations from AI and neural networks a definitive
direction.
Sure! Let's dive into the fascinating world of artificial neural networks (ANNs)
recognition tasks.
o Peculiarities: Primarily used for image and video recognition. They use
of features.
be complex to design.
series or natural language. They have connections that form directed cycles,
o Advantages: Effective for sequential data, good for tasks like language
problem. They have a more complex architecture that includes memory cells
that compete against each other. The generator creates data, and the
o Advantages: Capable of generating realistic data, useful for tasks like image
The exploration of dynamical systems using ANNs began in earnest with the advent of
Reservoir Computing (2000s), including Echo State Networks (ESNs) and Liquid State
With the resurgence of ANNs in the 2010s, thanks to increased computational power and
better algorithms, ANNs were increasingly applied to complex nonlinear dynamical systems.
Techniques like Long Short-Term Memory (LSTM) and Transformer networks now
1. Universal Approximation:
ANNs can approximate any continuous function, making them ideal for modeling the
2. Data-Driven Modeling:
3. Flexibility:
ANNs can handle time-series data, spatial data, or any combination thereof, enabling
1. Scalability:
2. Nonlinear Modeling:
dynamical systems.
3. Generalization:
o Once trained, ANNs can predict future states of a dynamical system with high
accuracy.
4. Automation:
models.
Adetayo 11
1. Data Dependency:
2. Interpretability:
physical insights.
3. Overfitting:
o Poor generalization may occur if the network is too complex relative to the
available data.
4. Computational Cost:
resource intensive.
ANNs are used to predict temperature, rainfall, and other meteorological phenomena.
2. Fluid Dynamics:
dynamics.
o ANNs help design control systems for robots, including path optimization and
stability.
Adetayo 12
5. Biological Systems:
Future Directions
Combine data-driven ANN methods with physical laws for more accurate and
• Hybrid Models:
f(x) = x
• Advantages:
o Simple to implement.
• Disadvantages:
a. Sigmoid Function
• Definition:
1
f(x)= 1+𝑒 −𝑥
• Range: (0,1)
• Advantages:
• Disadvantages:
• Definition:
𝑒 𝑥 −𝑒 −𝑥
f(x)=tanh(x) = 𝑒 𝑥 +𝑒 −𝑥
• Range: (−1,1)
• Advantages:
• Disadvantages:
• Definition:
f(x)=max (0, x)
Adetayo 14
• Range: [0, ∞)
• Advantages:
• Disadvantages:
o Can suffer from the dying ReLU problem, where neurons output zero
d. Leaky ReLU
• Definition:
• Range: (−∞,∞)
• Advantages:
o Addresses the dying ReLU problem by allowing small gradients for negative
inputs.
• Disadvantages:
• Definition:
• Range: (−α, ∞)
• Advantages:
• Disadvantages:
f. SoftMax Function
• Definition:
𝒆−𝒙𝒊
f(x)i=∑𝑵 𝒙𝒋
, where xi is the i-th output and N is the total number of outputs.
𝒋=𝟏 𝒆
• Advantages:
• Disadvantages:
g. Swish Function
• Definition:
𝟏
f(x)=x.sigmoid(x) = x.𝟏+𝒆−𝒙
• Range: (−∞,∞)
• Advantages:
• Disadvantages:
o Computationally intensive.
• Input/Hidden Layers:
o Use ReLU or its variants (Leaky ReLU, ELU) for fast training and improved
performance.
• Output Layers:
References
Brief History of Neural Networks. Although the study of the human brain… | by Kate