Unit-V DL
Unit-V DL
construction.
with PCA.
as the output.
encoder/decoder.
Architecture of Autoencoders
1. Encoder
2. Code
3. Decoder
error.
Regularized Autoencoders
Applications of RAE
Sparse Autoencoders
Sparse autoencoders (SAEs) impose a sparsity constraint: rather than
creating an information bottleneck by reducing the number of nodes in
each hidden layer, SAEs create a bottleneck by reducing the number of
nodes that can be activated at the same time.
Denoising Autoencoders
Denoising autoencoders are given partially corrupted input data and
trained to restore the original input by removing useless information
through dimensionality reduction.
Unlike most autoencoders, denoising autoencoders do not have the
ground truth data as its input. Instead, Gaussian noise is added to the
original data—for example, adding random static to an image—and the
denoising autoencoder (DAE) learns to filter it out. During model
training, the reconstruction error of the denoised output is not measured
against the corrupted input data, but against the original image.
In addition to preventing overfitting, this training technique also makes
denoising autoencoders very useful for cleaning up noisy or corrupted
image and audio files. Denoising autoencoders have also served as
foundational training paradigms for state-of-the-art image generation
models like Stable Diffusion.
Contractive Autoencoder
A Contractive Autoencoder (CAE) is a specific type of autoencoder used
in unsupervised machine learning. Autoencoders are neural
networks designed to learn efficient representations of the input data,
called encodings, by training the network to ignore insignificant data
(“noise”). These encodings can then be used for tasks such as
dimensionality reduction, feature learning, and more.
The "contractive" aspect of CAEs comes from the fact that they are
regularized to be insensitive to slight variations in the input data. This is
achieved by adding a penalty to the loss function during training, which
forces the model to learn a representation that is robust to small changes
or noise in the input. The penalty is typically the Frobenius norm of the
Jacobian matrix of the encoder activations with respect to the input and
encourages the learned representations to contract around the training
data.
The training process involves minimizing a loss function that has two
terms. The first term is the reconstruction loss, which measures the
difference between the original input and the reconstructed output. The
second term is the regularization term, which measures the sensitivity of
the encoded representations to the input. By penalizing the sensitivity, the
CAE learns to produce encodings that do not change much when the
input is perturbed slightly, leading to more robust features.
Stochastic Autoencoders
Instead of a deterministic mapping between input and latent space, VAEs
learn a probability distribution over the latent space.
This means that for the same input, the encoder can produce different
latent representations, leading to diverse outputs during decoding.
The encoder learns a mapping from the input data to a latent space, and
the decoder learns to reconstruct the input data from the latent space.
Key Concepts:
Encoder: The part of the network that maps the input to the latent space.
Decoder: The part of the network that reconstructs the input from the
latent space.
Generative Capabilities:
They can be used to generate new data samples that are similar to the
training data.
Diverse Outputs:
The stochastic nature allows for the generation of a variety of outputs
for the same input.
Probabilistic Representations:
They learn a probability distribution over the latent space, capturing the
uncertainty in the data.
Denoising and Feature Extraction:
Like regular autoencoders, they can be used for denoising and feature
extraction.
Boltzmann Machine
hidden or visible.
system.
Boltzmann Machine learns how the system works in its
When these RBMs are stacked on top of each other, they are
model.
learning algorithm.
RBM is undirected and has only two layers, Input layer, and
hidden layer
All visible nodes are connected to all the hidden nodes. RBM
has two layers, visible layer or input layer and hidden layer so
nodes.
Working of RBM
There are two other layers of bias units (hidden bias and
from autoencoders.
pass and the visible bias helps RBM to reconstruct the input
themselves.
activated or not.
number of columns.
The first hidden node will receive the vector multiplication of
opposite direction.
resembles the deep neural network but are not the same.
input layer.
learning.
Restricted Boltzmann Machines
backpropagation.
Working of DBN
The Greedy learning algorithm is used to pre-train DBN.
above.
from it.
layers.
shown in Figure.
DBM learns the features hierarchically from the raw data and
algorithm is fine-tuned.
parameters.
The stacked layer of restricted Boltzmann machine with
The data latent features are detected with the usage of DBM
representation.
Working of DBM
data.
new data.
training.
Key concepts in DBM
state of the network. States that are more likely have lower
this energy.
important.
detecting fakes.
independently.
3. The generator modifies some data attributes by adding noise
process is over.