Notes Deep Learning
Notes Deep Learning
Computer Vision:
In computer vision, Deep learning models can enable machines to identify and
understand visual data. Some of the main applications of deep learning in
computer vision include:
• Object detection and recognition: Deep learning model can be used to
identify and locate objects within images and videos, making it possible for
machines to perform tasks such as self-driving cars, surveillance, and robotics.
• Image classification: Deep learning models can be used to classify images into
categories such as animals, plants, and buildings. This is used in applications
such as medical imaging, quality control, and image retrieval.
• Image segmentation: Deep learning models can be used for image
segmentation into different regions, making it possible to identify specific
features within images.
P Jyothi,Asst. Prof., CSE Dept.
Applications of Deep Learning Conti..
Reinforcement learning:
In reinforcement learning, deep learning works as training agents to take action
in an environment to maximize a reward. Some of the main applications of deep
learning in reinforcement learning include:
• Game playing: Deep reinforcement learning models have been able to beat
human experts at games such as Go, Chess, and Atari.
• Robotics: Deep reinforcement learning models can be used to train robots to
perform complex tasks such as grasping objects, navigation, and manipulation.
• Control systems: Deep reinforcement learning models can be used to control
complex systems such as power grids, traffic management, and supply chain
optimization.
The history of deep learning can be traced back to 1943, when Walter Pitts and
Warren McCulloch created a computer model based on the neural networks of the
human brain.
The 1960s:
Henry J. Kelley is given credit for developing the basics of a continuous Back
Propagation Model in 1960. In 1962, a simpler version based only on the chain rule
was developed by Stuart Dreyfus. While the concept of back propagation (the
backward propagation of errors for purposes of training) did exist in the early 1960s,
it was clumsy and inefficient, and would not become useful until 1985.
The 1970s
The first “convolutional neural networks” were used by Kunihiko Fukushima.
Fukushima designed neural networks with multiple pooling and convolutional layers.
In 1979, he developed an artificial neural network, called Neocognitron, which used a
hierarchical, multilayered design. This design allowed the computer “learn” to
recognize visual patterns.
The networks resembled modern versions, but were trained with a reinforcement
strategy of recurring activation in multiple layers, which gained strength over time.
Additionally, Fukushima’s design allowed important features to be adjusted manually
by increasing the “weight” of certain connections.
The 1989:
Yann LeCun provided the first practical demonstration of backpropagation at Bell
Labs. He combined convolutional neural networks with back propagation onto read
“handwritten” digits. This system was eventually used to read the numbers of
handwritten checks.
The1995:
Dana Cortes and Vladimir Vapnik developed the support vector machine (a system
for mapping and recognizing similar data). LSTM (long short-term memory) for
recurrent neural networks was developed in 1997, by Sepp Hochreiter and Juergen
Schmidhuber.
The 1999:
The next significant evolutionary step for deep learning took place in
1999, when computers started becoming faster at processing data
and GPU (graphics processing units) were developed. Faster
processing, with GPUs processing pictures, increased computational
speeds by 1000 times over a 10 year span.
During this time, neural networks began to compete with support
vector machines. While a neural network could be slow compared to a
support vector machine, neural networks offered better results using
the same data. Neural networks also have the advantage of
continuing to improve as more training data is added.
The 2000-2010:
Around the year 2000, The Vanishing Gradient Problem appeared. It was discovered
“features” (lessons) formed in lower layers were not being learned by the upper
layers, because no learning signal reached these layers. This was not a fundamental
problem for all neural networks, just the ones with gradient-based learning methods.
The source of the problem turned out to be certain activation functions. A number of
activation functions condensed their input, in turn reducing the output range in a
somewhat chaotic fashion. This produced large areas of input mapped over an
extremely small range. In these areas of input, a large change will be reduced to a
small change in the output, resulting in a vanishing gradient. Two solutions used to
solve this problem were layer-by-layer pre-training and the development of long
short-term memory.
2011-2020:
By 2011, the speed of GPUs had increased significantly, making
it possible to train convolutional neural networks “without” the
layer-by-layer pre-training. With the increased computing
speed, it became obvious deep learning had significant
advantages in terms of efficiency and speed.
One example is AlexNet, a convolutional neural network whose
architecture won several international competitions during
2011 and 2012. Rectified linear units were used to enhance the
speed and dropout.
Deep Learning algorithms are becoming more widely used in every industry sector
from online retail to photography; some use cases are more popular and have
attracted extra attention of global media than others. Some widely publicized Deep
Learning applications include:
• Speech recognition used by Amazon Alexa, Google, Apple Siri, or Microsoft Cortana.
• Image recognition used for analyzing documents and pictures residing on large
databases.
• Natural Language Processing (NLP) used for negative sampling, sentiment analysis,
machine translation, or contextual entity linking.
• Automated drug discovery and toxicology used for drug design and development
work, as well as for predictive diagnosis of diseases.
• CRM activities used for automated marketing practices.
• Recommendation engines used in a variety of applications.
• Predictions in gene ontology and gene-function relationships.
• Health predictions based on data collected from wearables and EMRs.
The output signal (s) is typically a nonlinear function flx) of the activation
value x. The following equations describe the operation of an MP model:
Activation:
Output Signal:
Three commonly used nonlinear functions (binary, ramp and sigmoid) are
shown in Figure 1.3, although only the binary function was used in the
original MP model. Networks consisting of MP neurons with binary (on-off)
output signals can be configured to perform several logical functions
The desired or target output (b) is compared with the actual binary output
(s), and the error (6) is used to adjust the weights. The following equations
describe the operation of the perceptron model of a neuron:
where is the
learning rate
parameter.
TLU Implementation
Having established the theoretical base, the next step is to describe and implement
the training phase of the model. Basically, the implementation is based on the
following steps:
1. Identify inputs and the corresponding representation
2. Identify the free parameters in the problem
3. Specify the learning rule
4. Adjust the free parameters for optimisation
5. Evaluate the model