DL Concepts 1 Overview
DL Concepts 1 Overview
Overview
1
Why Study Artificial Neural
Networks?
They are extremely powerful computational devices (Turing
equivalent, universal computers)
They are very noise tolerant – so they can cope with situations
where normal symbolic systems would have difficulty
Bipolar sigmoid
2
g ( x) 2 f ( x) 1 x
1
1 e
The McCulloch-Pitts Neuron
Equation
The output out of a McCulloch-Pitts neuron as a
function of its n inputs ini :
The Perceptron
We can connect any number of McCulloch-Pitts
neurons together in any way we like.
An arrangement of one input layer of McCulloch-
Pitts neurons feeding forward to one output
layer of McCulloch-Pitts neurons is known as a
Perceptron.
The Perceptron
Logic Gates with MP Neurons
We can use McCulloch-Pitts neurons to
implement the basic logic gates.
x1 θ=2
x1x2 y 2
0 0 0 y
0 1 1
x2 2
1 0 1
1 1 1
Implementation of Logical
NOT, AND, and OR
Logical AND
x1 θ=2
x1x2 y 1
0 0 0 y
0 1 0
x2 1
1 0 0
1 1 1
Implementation of Logical
NOT, AND, and OR
Logical NOT
x1 θ=2
x1y -1
0 1 y
1 0 1
2
bias
Implementation of Logical
NOT, AND, and OR
Logical AND NOT
x1 θ=2
x1x2 y 2
0 0 0 y
0 1 0
x2 -1
1 0 1
1 1 0
Neural Networks
Convolutional Networks: LeNet-5
Neural Networks that use convolution in place of general
matrix multiplication in at least one layer
Feature Map
Grayscale Image
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Convolution
Convolution
wT x
Figure: I. Kokkinos
Pooling
Pooling
max{a i }
Pooling
max{a i }
Pooling
max{a i }
Pooling
max{a i }
71
Success of Deep Learning
--Amazing results obtained with relatively
simple ideas
--Incremental algorithmic innovations
--Availability of large amount of perceptual
data
--Fast highly parallel computation hardware
NVDIA GPUs
--Software layers CUDA, Tensorflow, Keras
72
Steps of Deep Learning
--Define problem
--Metrics to measure success
--Define testing and validation process
--Vectorize data
--Define model
--Tune parameters, refine model
--Avoid overfitting
73
Deep Learning Architectures
--Dense, CNN, RNN
--Vector data: (Dense)
--Image Data: 2D convnets
--Sound Data:1d convnets (preferred)/RNN
--Text: 1d convnets (preferred)/RNN
--Timeseries: RNN (preferred)/1d convnets
--Sequence: RNN/1-d convnets
--Video : 3-d convnets, 2-d CNN + RNN
74
Limitations of Deep Learning
--Cannot read specs and design algorithm
--Cannot code
--Too complex to express certain things as
a chain of simple vector transformations
--Do not have capability to learn without
data
--Cannot plan
75
Future of Deep Learning
--Models based on richer primitives
--New forms of learning algorithms
--Automatic Deep Learning
--Greater systematic use of previously
learned features and architectures
76
Credits
Many Slides by
Subhendu Trivedi and Rishi Condor,
University of Chicago
Michael Scherger
Department of Computer Science
Kent State University
Reference Book 1
Deep Learning
by
Ian Goodfellow,
Yoshua Bengio,
Aaron Courville,
MIT Press, 2017.
ISBN-13: 978-0-262-03561-3.
Reference Book 2