AAI - Module 2 - Variational Autoencoders
AAI - Module 2 - Variational Autoencoders
● Encoder Network
● Converts input data into a probability distribution in latent space
● Maps high-dimensional data to lower-dimensional representations
● Produces mean (μ) and variance (σ) parameters of the latent distribution
● Latent Space
● Compressed, continuous representation of input data
● Usually follows a normal distribution
● Enables meaningful interpolation between data points
● Dimensionality is much lower than input space
● Decoder Network
● Reconstructs input data from latent space samples
● Maps latent representations back to original data space
● Learns to generate new data samples
● Properties
● Continuous and smooth
● Semantically meaningful
● Supports interpolation
● Enables feature disentanglement
The Latent Space Representation is depicted in the diagram given below:
● Inference Process
● Encode input to get latent parameters
● Sample from latent distribution
● Generate new samples via decoder
Key Advantages
1. Undercomplete Autoencoders
● Definition: Simplest form where hidden layers have fewer dimensions than input
layer 5
● Key Characteristics:
● Forces compressed data representation
● Acts as dimensionality reduction technique
● More powerful than PCA due to non-linear transformation capabilities 1
2. Sparse Autoencoders
3. Contractive Autoencoders
4. Denoising Autoencoders
graph TD
A[Input Data] --> B{Autoencoder Type}
B --> C[Undercomplete]
B --> D[Sparse]
B --> E[Contractive]
B --> F[Denoising]
B --> G[Variational]
Comparison of Architectures
Training Objectives:
1. Undercomplete: Minimize reconstruction error with bottleneck constraint 5
2. Sparse: Balance reconstruction with activation sparsity 7
3. Contractive: Maintain similar encodings for similar inputs 1
4. Denoising: Reconstruct clean data from noisy input 6
5. Variational: Balance reconstruction with distribution matching 1
Use Cases: