Deep Learning U5
Deep Learning U5
• Definition: Deep Generative Models are a class of machine learning models designed
to generate new data samples that resemble a given dataset. They learn the
underlying distribution of the data and can produce new instances with similar
properties.
• Types: Include Restricted Boltzmann Machines (RBMs), Deep Belief Networks (DBNs),
Variational Autoencoders (VAEs), and Generative Adversarial Networks (GANs).
• Definition: RBMs are a type of generative stochastic neural network that can learn a
probability distribution over its set of inputs.
• Architecture:
• Energy Function:
▪ WijW_{ij}: Weights
• Training:
• Definition: A Markov Chain Monte Carlo (MCMC) method used to approximate the
distribution of the data and train RBMs.
• Steps:
o Sample Hidden Units: Sample hidden units from their conditional distribution
given the visible units.
o Sample Visible Units: Sample visible units from their conditional distribution
given the hidden units.
• Architecture:
o Stacks of RBMs, with each layer providing the input to the next layer.
• Training:
o Layer-wise Pretraining: Train each RBM layer by layer, starting from the bottom.
▪ Greedy Algorithm: Train one layer at a time, freezing the weights of the
previous layers.
5. Markov Networks
• Definition: Also known as Markov Random Fields (MRFs), these are undirected
graphical models that represent the joint distribution of a set of random variables.
• Components:
6. Markov Chains
• Definition: A stochastic process that transitions from one state to another based on a
probabilistic rule. The next state depends only on the current state (Markov property).
• Properties:
o Memoryless: The next state depends only on the current state, not the history.
7. Auto-Regressive Models
• Definition: Models that predict the next value in a sequence based on previous values.
They are commonly used for time series analysis and sequence generation.
• Examples:
• Components:
▪ Loss Function:
minGmaxDV(D,G)=Ex∼pdata(x)[logD(x)]+Ez∼pz(z)[log(1−D(G(z)))]
\min_G \max_D V(D, G) = \mathbb{E}_{x \sim p_{\text{data}}(x)}[\log
D(x)] + \mathbb{E}_{z \sim p_z(z)}[\log(1 - D(G(z)))]
• Training:
o Convergence: The goal is for the generator to produce data that the
discriminator cannot distinguish from real data.
• Object Detection:
o Techniques: YOLO (You Only Look Once), Faster R-CNN, SSD (Single Shot
MultiBox Detector).
• Speech Recognition:
• Image Recognition:
o Models: ResNet, Inception, EfficientNet.
• Video Analysis:
• Medical Science:
o Techniques