Lecture 4
Lecture 4
AML-3104
Deep Learning
Conventional Machine Learning
• Process natural data in raw form
• Constructing features by hand
• Requires domain expertise
• Difficult and time consuming
Three-Layer Arbitrary
(Complexity A B
B
Limited by No. A
of Nodes) B A
5
Deep Learning
• Layers are not handcrafted
• Features are learned from raw data via a general-purpose learning
algorithm
https://devblogs.nvidia.com/parallelforall/accelerate-machine-learning-cudnn-deep-neural-network-library/
Applications of Deep Learning
• Domains in science, business and government
• Beat current records in image and speech recognition
• Beaten other machine-learning techniques at
• Predicting activity of potential drug molecules
• Analyzing particle accelerator data
• Reconstructing brain circuits
• Produced promising results in natural language understanding
• Topic classification
• Sentiment analysis
• Question answering
Overview
• Supervised Learning
• Backpropagation to Train Multilayer Architectures
• Convolution Neural Networks
• Image Understanding with Deep Convolution Networks
• Distributed Representation and Language Processing
• Recurrent Neural Networks
• Future of Deep Learning
Supervised Learning
• Most common form of machine learning
• Data set
• Labeling
• Training on data set (tuning parameters, gradient descent)
• Testing
•Objective function: measures error between output scores and the
desired pattern of scores
•Modifies internal adjustable parameters (weights) to reduce error
Back-Propagation
10
Supervised Learning
•Objective Function → “Hilly landscape” in high dimensional space of
weight values
•Computes a gradient vector
•Indicates how much the error would increase or decrease if the weight were
increased by a tiny amount
•Negative gradient vector indicates the direction of steepest descent in this
landscape
•Taking it closer to a minimum, where the output error is low on average
Gradient Descent Algorithm
• Gradient descent is an optimization algorithm used to find the
values of parameters of a function that minimizes a cost function.
• It is an iterative algorithm.
• We use gradient descent to update the parameters of the model.
Gradient
Descent
GRADIENT DESCENT
14
GRADIENT DESCENT
• https://miro.medium.com/max/1400/1*E-
5K5rHxCRTPrSWF60XLWw.gif
15
GRADIENT DESCENT
16
Gradient descent
25
Multi Layer Neural Network
Distort the input space to make the classes of data (ex: red and blue
lines) linearly separable
Illustrative example with only two input units, two hidden units and
one output unit
Feed Forward
• Trained by backpropagation
• problematic because the back propagated gradients either grow or shrink
at each time step
• over many time steps they typically explode or vanish
Recurrent Neural Networks
Future of Deep Learning
• Expect unsupervised learning to become more important
• Human and animal learning is largely unsupervised
• Future progress in vision
• Systems trained end-to-end
• Combine ConvNets with RNNS that use reinforcement learning.
• Natural language
• RNNs systems will become better when they learn strategies for selectively
attending to one part at a time
Discussion
• Deep Learning has already drastically improved the state-of-the-art
in
• image recognition
• speech recognition
• natural language understanding
• Deep Learning requires very little engineering by hand and thus has
the potential to be applied to many fields
Applications
36
Neural network for OCR
A
•feedforward network B
Hidden
Layer Output
Layer
Input
Layer
37
OCR for 8x10 characters
10 10 10
8 8 8
•NN are able to generalise
•learning involves generating a partitioning of the input space
•for single layer network input space must be linearly separable
•what is the dimension of this input space?
•how many points in the input space?
•this network is binary(uses binary values)
•networks may also be continuous
38
Engine management
39
ALVINN
Drives 70 mph on a public highway
30 outputs
for steering
30x32 weights
4 hidden
into one out of
units
four hidden
30x32 pixels unit
as inputs
40
Signature recognition
41
Sonar target recognition
42
Stock market prediction
43
Mortgage assessment
44
References
Y. LeCun, Y. Bengio, G. Hinton (2015). Deep Learning. Nature 521, 436-444.