Lec1 Intro
Lec1 Intro
Introduction
Kaiming He
Fall 2024, EECS, MIT
The “GenAI” Era
Chatbot and natural language conversation
The “GenAI” Era
Text-to-image generation
Generated by Sora
The “GenAI” Era
AI assistant for code generation
The “GenAI” Era
Protein design and generation
Watson, et al. De novo design of protein structure and function with RFdiffusion, Nature 2023
The “GenAI” Era
Weather forecasting
Skilful precipitation nowcasting using deep generative models of radar, Nature 2021
Generative Models before the “GenAI” Era
2009, PatchMatch: Photoshop’s Content-aware Fill
PatchMatch: A Randomized Correspondence Algorithm for Structural Image Editing, SIGGRAPH 2009
Generative Models before the “GenAI” Era
1999, the Efros-Leung algorithm for texture synthesis
In today’s word: this is an Autoregressive model
generative
model
• “label” y ⇨ “sample” x
• many possible outputs
model
“dog”
y x
Discriminative vs. Generative models
discriminative generative
discriminative generative
constant for given x
• Generative models can be discriminative: Bayes’ rule
assuming known prior
discriminative generative
constant for given x
generative discriminative
constant for given y
discriminative models
x
Probability is part of the modeling
p
underfit
x
Probability is part of the modeling
p
overfit
x
Probability is part of the modeling
• To the extreme, using delta functions is
like sampling from training data
p
overfit
x
Generative models w/ probabilistic modeling
data
Generative models w/ probabilistic modeling
data
• This is already part
of the modeling
distribution
of data
Generative models w/ probabilistic modeling
data
estimated
distribution
of data
Generative models w/ probabilistic modeling
data
distribution
of data
distribution
of data
to approximate
simple distribution
data distribution
Learning to represent probability distributions
• Not all parts of distribution modeling is done by learning
Case study:
This dependency graph is
Autoregressive model designed (not learned).
Learning to represent probability distributions
• Not all parts of distribution modeling is done by learning
Case study:
noising
Diffusion model
Case study:
noising
Diffusion model
denoising
x: generated
3D structures
y: text prompt
Figure credit: Tang, et al. LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation. ECCV 2024
Case study: Formulating as p(x|y)
• Protein structure generation
x: generated
y: condition/constraint protein structures
(e.g., symmetry)
Watson, et al. De novo design of protein structure and function with RFdiffusion, Nature 2023
Case study: Formulating as p(x|y)
• Class-conditional image generation
“red fox” y: class label
x: generated image
Image generated by: Li, et al. Autoregressive Image Generation without Vector Quantization, 2024
Case study: Formulating as p(x|y)
• “Unconditional” image generation
y: an implicit condition
“images following CIFAR10 distribution”
Images generated by: Karras, et al. Elucidating the Design Space of Diffusion-Based Generative Models, NeurIPS 2022
Case study: Formulating as p(x|y)
• Classification (a generative perspective)
cat
bird
horse
dog
Case study: Formulating as p(x|y)
• Open-vocabulary recognition
bird
flamingo
red color
orange color
...... ...
Case study: Formulating as p(x|y)
• Image captioning
Chi, et al. Diffusion Policy: Visuomotor Policy Learning via Action Diffusion, RSS 2023
Formulating Real-world Problems as Generative Models
• Generative models are about
• Many problems can be formulated as generative models
• What’s x? What’s y?
• How to represent x, y, and their dependence?
About this course
This course will cover:
• How real-world problems are formulated as generative models?
• Probabilistic foundations and learning algorithms
• Challenges, opportunities, open questions