cs236_lecture3
cs236_lecture3
Stefano Ermon
Stanford University
Lecture 3
Training data on the left (Caltech 101 Silhouettes). Samples from the
model on the right.
Figure from Learning Deep Sigmoid Belief Networks with Data
Augmentation, 2015.
To improve model: use one layer neural network instead of logistic regression
hi = σ(Ai x<i + ci )
x̂i = p(xi |x1 , · · · , xi−1 ; Ai , ci , αi , bi ) = σ(αi hi + bi )
| {z }
parameters
x
For example h2 = σ . . .. .
.. x1 + .. h3 = σ .. .. ( x2 ) + ..
1
In numpy: np.exp(a)/np.sum(np.exp(a))
Stefano Ermon (AI Lab) Deep Generative Models Lecture 3 12 / 36
RNADE
x̂ i defines the mean and standard deviation of each of the K Gaussians (µji , σij ).
Can use exponential exp(·) to ensure non-negativity
Stefano Ermon (AI Lab) Deep Generative Models Lecture 3 14 / 36
Autoregressive models vs. autoencoders
e and d are constrained so that we don’t learn identity mappings. Hope that
e(x) is a meaningful, compressed representation of x (feature learning)
A vanilla autoencoder is not a generative model: it does not define a
distribution over x we can sample from to generate new data points.
Stefano Ermon (AI Lab) Deep Generative Models Lecture 3 15 / 36
Autoregressive autoencoders
Pros:
1 Can be applied to sequences of arbitrary length.
2 Very general: For every computable function, there exists a finite
RNN that can compute it
Cons:
1 Still requires an ordering
2 Sequential likelihood evaluation (very slow for training)
3 Sequential generation (unavoidable in an autoregressive model)
Train 3-layer RNN with 512 hidden nodes on all the works of Shakespeare.
Then sample from the model:
KING LEAR: O, if you were a feeble sight, the courtesy of your law,
Your sight and several breath, will wear the gods
With his heads, and my hands are wonder’d at the deeds,
So drop upon your lordship’s head, and your opinion
Shall be against your honour.
Naturalism and decision for the majority of Arab countries’ capitalide was
grounded by the Irish language by [[John Clair]], [[An Imperial Japanese
Revolt]], associated with Guangzham’s sovereignty. His generals were
the powerful ruler of the Portugal in the [[Protestant Immineners]], which
could be said to be directly in Cantonese Communication, which followed
a ceremony and set inspired prison, training. The emperor travelled
back to [[Antioch, Perth, October 25—21]] to note, the Kingdom of
Costa Rica, unsuccessful fashioned the [[Thrales]], [[Cynth’s Dajoard]],
known in western [[Scotland]], near Italy to the conquest of India with
the conflict.
Note: correct Markdown syntax. Opening and closing of brackets [[·]]
== See also ==
*[[Iender dome of the ED]]
*[[Anti-autism]]
== External links==
* [http://www.biblegateway.nih.gov/entrepre/ Website of the World
Festival. The labour of India-county defeats at the Ripper of California
Road.]
Train on data set of baby names. Then sample from the model:
Convolutions are natural for image data and easy to parallelize on modern
hardware.
Dilated convolutions increase the receptive field: kernel only touches the
signal at every 2d entries.