Questions Bank On DL
Questions Bank On DL
Introduction to RNN
What is a Recurrent Neural Network (RNN), and how does it differ from traditional
feedforward neural networks?
Why are RNNs particularly suited for sequential data?
How do RNNs handle temporal dependencies in data?
Can you describe the general architecture of an RNN and explain the role of each
component (input, hidden state, output)?
How does information flow through an RNN at each time step?
What is the significance of the "feedback" loop in an RNN?
What are some common types of recursive neural networks (RNN variants)?
How do bidirectional RNNs improve the performance of a standard RNN?
What are attention mechanisms in RNNs, and how do they work?
What is Back Propagation Through Time (BPTT), and how does it work in training
RNNs?
What are the challenges associated with BPTT, particularly the vanishing gradient
problem?
How does BPTT differ from traditional backpropagation?
What is an Elman Neural Network, and how does it differ from a standard RNN?
How does the context layer in an Elman network contribute to its learning?
In what scenarios is an Elman neural network preferable over a standard RNN?
What is Long Short-Term Memory (LSTM), and how does it address the limitations
of traditional RNNs?
What are the primary components of an LSTM unit (forget gate, input gate, output
gate)?
How does the cell state in an LSTM function, and why is it important?
8. Traditional LSTM
What are the main differences between an LSTM and a standard RNN in terms of
architecture and functionality?
How does an LSTM prevent the vanishing gradient problem that standard RNNs
suffer from?
What is the role of the forget gate in an LSTM?
9. Modes of LSTM
What are the different modes in which an LSTM can be used, such as in classification
or regression tasks?
How do the output and hidden states interact during the execution of an LSTM?
How does an LSTM handle both short-term and long-term dependencies in a dataset?
Explain LSTM with simple example.
What are the key properties that make LSTM networks well-suited for sequential data
analysis?
How does an LSTM network’s architecture help it maintain memory across time
steps?
What makes LSTMs more robust than traditional RNNs for long-term sequence
processing?
Can you explain the internal architecture of an LSTM unit (input, forget, and output
gates)?
What are the mathematical equations governing the flow of data through an LSTM?
How do LSTMs compare with other types of RNNs in terms of computational
complexity?
What is an autoencoder, and what are its primary components (encoder and decoder)?
How does an autoencoder learn to compress and reconstruct data?
What is the primary objective of training an autoencoder model?
Explain autoencoder with simple example.
2. Undercomplete Autoencoders
3. Regularized Autoencoders
5. Denoising Autoencoders
What is a denoising autoencoder, and how does it handle noisy input data?
How does training a denoising autoencoder improve the robustness of the model?
Can you explain the concept of adding noise to the input during training in the context
of denoising autoencoders?
Explain denoising autoencoder with simple example.
6. Contractive Autoencoders
What is a contractive autoencoder, and how does it differ from a regular autoencoder?
How does the contractive penalty in the loss function of a contractive autoencoder
help in learning more robust features?
In which types of problems are contractive autoencoders particularly useful?
8. Applications of Autoencoders
What are some common applications of autoencoders in machine learning and deep
learning?
How can autoencoders be used in anomaly detection and outlier detection?
How do autoencoders play a role in dimensionality reduction and feature extraction?
What are some fast learning algorithms that can be used to train autoencoders
efficiently?
How do techniques like stochastic gradient descent (SGD) or its variants speed up the
training process for autoencoders?
Can you explain the importance of using a fast learning algorithm for large datasets or
complex autoencoder models?
What is a deep generative model, and how does it differ from discriminative models
in machine learning?
How do deep generative models learn the underlying distribution of the data?
Can you explain how deep generative models are used to generate new data that is
similar to the training data?
2. Boltzmann Machine
What is a Deep Boltzmann Machine (DBM), and how does it differ from a Deep
Belief Network?
How do the layers in a Deep Boltzmann Machine communicate with each other, and
what makes DBMs more flexible than DBNs?
What are the key advantages of using Deep Boltzmann Machines for generative
modeling?
6. Discriminator Network
7. Generator Network
8. Types of GANs
What are some common types or variants of GANs, and how do they differ from the
original GAN architecture (e.g., DCGAN, CycleGAN, WGAN)?
How does a Conditional GAN (CGAN) differ from a standard GAN, and what are its
applications?
What is a Wasserstein GAN (WGAN), and how does it address issues related to
training instability in GANs?
What are some real-world applications of GANs in industries like art, healthcare, and
finance?
How can GANs be applied in image generation and data augmentation?
Can GANs be used for text-to-image generation, and how do they work in this
domain?
How are GANs being used in fields like drug discovery and protein folding?
1. Introduction to Deep Reinforcement Learning
What is deep reinforcement learning (DRL), and how does it combine reinforcement
learning and deep learning techniques?
How does deep reinforcement learning differ from traditional reinforcement learning
in terms of complexity and model architecture?
What types of problems are best suited for deep reinforcement learning?
Can you explain how DRL can learn from interaction with an environment and make
decisions?
What is a Markov Decision Process (MDP), and what are its key components (states,
actions, rewards, and transition probabilities)?
How does the Markov property influence the design of an MDP?
What is the role of the reward function in an MDP, and how does it guide the agent's
behavior?
How does an agent use the MDP to make decisions and optimize its actions over
time?
What are some common challenges faced when applying reinforcement learning to
real-world problems?
How does the high variance in rewards or sparse rewards impact the training of
reinforcement learning models?
What are some difficulties associated with scaling reinforcement learning algorithms
to large, complex environments?
What are the challenges of ensuring stability and convergence in deep reinforcement
learning algorithms?
What are dynamic programming (DP) algorithms, and how do they relate to
reinforcement learning?
Can you explain the difference between policy evaluation and policy improvement in
the context of DP algorithms for RL?
What is value iteration and policy iteration in reinforcement learning, and how do
these techniques help in solving MDPs?
How do DP methods ensure the optimality of solutions, and why are they
computationally expensive?
What are some common limitations of deep learning models when applied to
reinforcement learning problems?
How does the need for large amounts of data and computational resources impact the
application of deep learning in reinforcement learning?
What are some of the challenges in training deep reinforcement learning agents with
high-dimensional input spaces (e.g., images)?
How does the risk of overfitting affect deep learning models in reinforcement
learning, and what techniques can mitigate this risk?