1 IntroductionDL
1 IntroductionDL
Lecture 1
Introduction
Vasileios Belagiannis
Chair of Multimedia Communications and Signal Processing
Friedrich-Alexander-Universität Erlangen-Nürnberg
20.10.2023
• Course Topics.
• Course Logistics.
• Introduction to the topic.
The idea of the course is to work on the basics of deep learning where
the main focus will be on deep neural networks. We will learn how to
build, train and evaluate models. The course goes through the theory of
deep learning. Our reference will be mostly computer vision problems.
• Schedule: Lecture on Wednesdays at 08:15 and Exercise on Fridays
at 10:15.
• Course Material: StudOn.
• Exercises: Jupyter Notebook / Python (numpy, pytorch, matplotlib).
• Background: advanced mathematics and programming will be useful
for following the lecture.
Picture from Wikipedia, Max Schmachtenberg, Christian Bizer, Anja Jentzsch and Richard Cyganiak
AI ML DL
0
02
30 years of adaptive neural networks: perceptron, madaline, and backpropagation [20].
,2
9]
[1
• There have been two AI winters, which are closely related to the
er
rm 18
***Not for sharing (LMS, Friedrich-Alexander-Universität Erlangen-Nürnberg)***
fo 0 12
ns , 2 ] 20
tra 18 [ 14 ], 07
nt h 20 [14 , 20
Introduction to Deep Learning, 01. Introduction
r , r 99
eu a 5]
l ve ee Be
N et. t [1 tske D ,1
e e u vs eep 1]
H GN , S w D [1
y lo ,
al eh et 97
VG vsk Sh T eN 19
neural network evolution (gray zones chronology).
e L
izh l., & r, 0],
Kr t . a er
o
ff ne [1
e nd a M
2010
io si H T 86
ng , O & LS 19
Be ton o ,
gi er ],
[9
in en hub n
H , B id io
ou m at
tt ch ag
Bo S op
pr
n, r & ck 82
2000
Cu ite Ba 19
Le chre s, ],
o m [9
ia
H ill on
W itr
& gn
to
n co
in eo
N 4 0
1990
H
t, a, 19 97
7
ar hi
m
], ], 1
elh us [8
m uk n n [7
Ru o F io
at tio 6
9
20.10.2023
hi
k ag ga 19
ni op pa ],
Neural Network Evolution
Ku pr ro [6
ck kp ns
1980
Ba ac tro
s, , B ep
bo aa rc
Vasileios Belagiannis
pp & 19
Se sky ],
1970
in [5
M n
ro
pt
ce
er
,P 9
tt 1 4
la 5 19
nb 19 ],
se ], [3 43
1960
Ro [4 ry
o 19
k RC he ],
an A T [2
Fr SN n it
y, bi
a Un
sk b c
in He gi
M , Lo
n bb ol
d
vi e sh
ar H re
1950 M ald
on Th
D s,
itt
-P
ch
lo
ul
cC
M
McCulloch-Pitts Neuron [2] (1943)
• Neurophysiologist Warren McCulloch & logician Walter Pitts.
• Input xi ∈ {0, 1} and binary output y ∈ {0, 1}.
• Excitatory or inhibitory input, represented by the weights/parameters
wi ∈ {−1, 1}.
• The mapping is represented by a threshold function f : RD → R,
where: (
0, if w · x ≤ T
f (x) = (1)
1, otherwise.
x0
w0
x1
w1 P
w2
x2
w3
x3
Homework
Implement the NOR gate. Design a McCulloch-Pitts for boolean input of 3 elements.
Set the weights to 1,1,-1 and compute the threshold for the NOR logic operation.
Weights
b
w0 Activation
x1
w1 P
w2
x2
wn
xn
Perceptron limitations:
• A single layer Perceptron cannot implement the XOR logical
function.
• A multi-layer Perceptron is capable of approximating XOR.
• The above observations were not clear in the book and resulted in a
lot of criticism for the neural networks. The book Perceptrons made
predictions that may have caused a decline in neural net research in
the 1970s and early 1980s [22].
• Hierarchical
multilayered
ANN.
• Partial shift
invariance.
• Precursor of
convolutional
neural
networks.
• Convolutional
neural networks
by Yann LeCun
(computer
scientist).
• Digit
Recognition
(32x32 image
input).
• Precursor of
deep neural
networks.
)
NL
W o
o
Pr Pr
(K
re re
Fi
90
Fi
72
0
0X 0 K4
i
Ph
0 K2 K4
K2 sla sla
on
la sla Te
Te
Xe
Te
s Te
103 70 70 .
0 69 69 70 Xeon Phi 7120 (KNC)
7 Ed
58 HD HD Hz 89
GFLOP/sec
HD G HD v4
90 70 9
9 v3 v3 69
20 D 7 9 9 99 -2
M H 26 26 E5
70 sla 5- 5-
48 0 Te E E
HD 05
C2
sla v2
60 Te 7
70 0 10 69 INTEL Xeon CPUs
38 06 C 0 -2
HD C1 s la 69 E5
Te 5-
2
2 sla E NVIDIA Tesla GPUs
10 Te
• Handwritten digits.
• 28x28 grayscale images.
• 60k training samples.
• 10k testing samples.
• 10 categories (classification).
• Standard for convolution neural
networks and ML approaches.
http://www.statmt.org/wmt14/translation-task.html
Important Takeaways:
• Industry outpaces academia on SOTA.
• No significant results improvement on major benchmarks. However
more complex benchmarks appeared.
• AI is both helpful and harmful for the environment.
• Generative models outperform humans in some cases.
• AI misuse increases compared to the past.
• More and more companies adopt AI / ML.
• Policy making for AI gets more attention.
Stanford, H. A. I. ”The AI Index Report: Measuring Trends in Artificial intelligence [Ebook]” (2023).
Yoav Shoham, Raymond Perrault, Erik Brynjolfsson, Jack Clark, James Manyika, Juan Carlos Niebles, Terah Lyons, John Etchemendy,
Barbara Grosz and Zoe Bauer, ”The AI Index 2018 Annual Report”, AI Index Steering Committee, Human-Centered AI Initiative, Stanford
University, Stanford, CA, December 2018.
Stanford, H. A. I. ”The AI Index Report: Measuring Trends in Artificial intelligence [Ebook]” (2023).
Stanford, H. A. I. ”The AI Index Report: Measuring Trends in Artificial intelligence [Ebook]” (2023).
Stanford, H. A. I. ”The AI Index Report: Measuring Trends in Artificial intelligence [Ebook]” (2023).
Stanford, H. A. I. ”The AI Index Report: Measuring Trends in Artificial intelligence [Ebook]” (2023).
Stanford, H. A. I. ”The AI Index Report: Measuring Trends in Artificial intelligence [Ebook]” (2023).
Stanford, H. A. I. ”The AI Index Report: Measuring Trends in Artificial intelligence [Ebook]” (2023).
Zhang, Daniel, Saurabh Mishra, Erik Brynjolfsson, John Etchemendy, Deep Ganguli, Barbara Grosz, Terah Lyons et al. ”The AI Index 2021
Annual Report.” arXiv preprint arXiv:2103.06312 (2021).
Zhang, Daniel, Saurabh Mishra, Erik Brynjolfsson, John Etchemendy, Deep Ganguli, Barbara Grosz, Terah Lyons et al. ”The AI Index 2021
Annual Report.” arXiv preprint arXiv:2103.06312 (2021).
Zhang, Daniel, Saurabh Mishra, Erik Brynjolfsson, John Etchemendy, Deep Ganguli, Barbara Grosz, Terah Lyons et al. ”The AI Index 2021
Annual Report.” arXiv preprint arXiv:2103.06312 (2021).
Yoav Shoham, Raymond Perrault, Erik Brynjolfsson, Jack Clark, James Manyika, Juan Carlos Niebles, Terah Lyons, John Etchemendy,
Barbara Grosz and Zoe Bauer, ”The AI Index 2018 Annual Report”, AI Index Steering Committee, Human-Centered AI Initiative, Stanford
University, Stanford, CA, December 2018.
Zhang, Daniel, Saurabh Mishra, Erik Brynjolfsson, John Etchemendy, Deep Ganguli, Barbara Grosz, Terah Lyons et al. ”The AI Index 2021
Annual Report.” arXiv preprint arXiv:2103.06312 (2021).
Zhang, Daniel, Saurabh Mishra, Erik Brynjolfsson, John Etchemendy, Deep Ganguli, Barbara Grosz, Terah Lyons et al. ”The AI Index 2021
Annual Report.” arXiv preprint arXiv:2103.06312 (2021).
[26] Hans Moravec. When will computer hardware match the human brain. Journal of evolution
and technology, 1(1):10, 1998.
[27] Alex Krizhevsky and Geoffrey Hinton. Learning multiple layers of features from tiny images.
Technical report, Citeseer, 2009.
[28] Benjamin Recht, Rebecca Roelofs, Ludwig Schmidt, and Vaishaal Shankar. Do cifar-10
classifiers generalize to cifar-10? arXiv preprint arXiv:1806.00451, 2018.
[29] Mark Everingham, SM Ali Eslami, Luc Van Gool, Christopher KI Williams, John Winn, and
Andrew Zisserman. The pascal visual object classes challenge: A retrospective. International
journal of computer vision, 111(1):98–136, 2015.
[30] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan,
Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In
European conference on computer vision, pages 740–755. Springer, 2014.
[31] Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. Squad: 100,000+
questions for machine comprehension of text. arXiv preprint arXiv:1606.05250, 2016.
[32] Samuel R Bowman, Gabor Angeli, Christopher Potts, and Christopher D Manning. A large
annotated corpus for learning natural language inference. arXiv preprint arXiv:1508.05326,
2015.