Artificial Neural Network Course Slides
Artificial Neural Network Course Slides
Lecture 1: Introduction
Spring 2025
2
Prerequisites
AIS301: Machine
Learning
4
Grading Policy
Refernece : https://www.deeplearningbook.org/
Textbook
• A great book!
8
Topics
Deep neural networks
How to train them
How to measure their performance
How to make that performance better
9
Warm-Up
https://play.blooket.com/play?id=3123283
2
What is AI?
•AI aims to simulate intelligent behavior.
3
What is Machine Learning?
4
5
What is Deep Learning?
•Deep Learning (DL) is a type of ML based on deep neural
networks.
Model Model
Real world input Real world output
input output
22
Simple example …
• Predict the height of a child given his/her age.
Model Model
Age of child Height of child
input output
23
What is a supervised learning model?
Regression
24
Figures from http://udlbook.com
What is a supervised learning model?
Deep learning
model
Deep learning
model
Deep learning
model
Deep learning
model
Deep learning
model
34
35
“Given a news article, I want to predict if it is political,
sports, or economical”.
What type of a problem is that?
➢ multivariate classification
➢ univariate regression
➢ binary classification
➢ multivariate regression
➢ multiclass classification
36
Other type of examples …
37
Translation
Deep learning
model
38
Figures from http://udlbook.com
Image captioning
Deep learning
model
39
Figures from http://udlbook.com
Image generation from text
Deep learning
model
40
Figures from http://udlbook.com
What do these examples
have in common?
41
What do these examples have in common?
• Very complex relationship between input and output
• Sometimes may be many possible valid answers
• But outputs (and sometimes inputs) obey rules
34
DeepCluster: Deep Clustering for Unsupervised Learning of Visual Features (Caron et al., 2018) 35
Figures from http://udlbook.com
Unsupervised Learning
• Learning about a dataset without labels
• e.g., clustering
• Generative models can create examples
• e.g., generative adversarial networks
36
Unsupervised Learning
• Learning about a dataset without labels
• e.g., clustering
• Generative models can create examples
• e.g., generative adversarial networks
• PGMs learn distribution over data
• e.g., variational autoencoders,
• e.g., normalizing flows,
• e.g., diffusion models
47
Figures from http://udlbook.com
Generative models
48
Figures from http://udlbook.com
Generative models
49
Figures from http://udlbook.com
Generative models
50
Figures from http://udlbook.com
Conditional synthesis
51
Figures from http://udlbook.com
Figures from http://udlbook.com 42
ChatGPT
53
54
45
Reinforcement learning
• Build an agent which lives in a world and can perform certain actions
at each time step.
• Goal: take actions to change the state so that you receive rewards
• You don’t receive any data – you have to explore the environment
yourself to gather data as you go.
46
Example: chess
• States are valid states of the chess board
• Actions at a given time are valid possible moves
• Positive rewards for taking pieces, negative rewards for losing
them
57
Figures from http://udlbook.com
Example: chess
• States are valid states of the chess board
• Actions at a given time are valid possible moves
• Positive rewards for taking pieces, negative rewards for losing
them
58
Figures from http://udlbook.com
Why is this difficult?
• Stochastic
• Make the same move twice, the opponent might not do the same thing
• Rewards also stochastic (opponent does or doesn’t take your piece)
• Temporal credit assignment problem
• Did we get the reward because of this move? Or because we made good
tactical decisions somewhere in the past?
• Exploration-exploitation trade-off
• If we found a good opening, should we use this?
• Or should we try other things, hoping for something better?
59
Landmarks in (Machine) Deep Learning
• 1958 Perceptron (Simple `neural’ model)
• 1986 Backpropagation (Practical Deep Neural networks)
• 1989 Convolutional networks (Supervised learning)
• 2012 AlexNet Image classification (Supervised learning)
• 2014 Generative adversarial networks (Unsupervised learning)
• 2014 Deep Q-Learning -- Atari games (Reinforcement learning)
• 2016 AlphaGo (Reinforcement learning)
• 2017 Machine translation (Supervised learning)
• 2019 Language models ((Un)supervised learning)
• 2022 Dall-E2 Image synthesis from text prompts ((Un)supervised learning)
• 2022 ChatGPT ((Un)supervised learning)
• 2023 GPT4 Multimodal model ((Un)supervised learning)
60
2018 Turing award winners
61
Where are we going?
• Supervised learning (overview with regression example)
• Shallow neural networks (a more flexible model)
• Deep neural networks (an even more flexible model)
• Loss functions (guiding the training)
• How to train neural networks (gradient descent and variants)
• How to measure performance of neural networks (generalization)