GenAI QBank Ans
GenAI QBank Ans
1. Define generative AI and explain its significance in modern machine learning. How does it differ from
traditional AI approaches? (Understanding)
Generative AI refers to a subset of artificial intelligence that focuses on creating new data or content rather than
merely analyzing or interpreting existing data. It uses machine learning models to generate outputs such as images,
text, music, videos, and other forms of content that resemble real-world data. Popular generative AI models include
Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), and transformer-based models like
GPT (Generative Pre-trained Transformer).
The evolution of generative AI has been marked by breakthroughs in algorithms, computational power, and theoretical advancements in
machine learning. Below is a timeline of the historical development and key milestones leading to the current state of generative AI:
3. Illustrate with examples how generative AI has impacted different industries such as healthcare,
entertainment, and finance. (Applying)
Generative AI has profoundly impacted various industries by enabling creativity, efficiency, and innovation. Here are specific
examples illustrating its influence across healthcare, entertainment, and finance:
1. Healthcare
● Drug Discovery:
○ Example: Companies like Insilico Medicine use generative AI models to design new molecular structures for
drug candidates, reducing time and cost in drug discovery.
○ Impact: AI-designed molecules like DSP-1181 for treating obsessive-compulsive disorder were created in
record time.
● Medical Imaging:
○ Example: GANs enhance low-quality MRI scans, allowing for sharper images without additional radiation
exposure.
○ Impact: Improves diagnostics for conditions like tumors or heart diseases.
● Medical Record Summarization:
○ Example: Generative AI models summarize patient histories for doctors, facilitating better decision-making
during consultations.
○ Impact: Reduces administrative burden and improves patient care efficiency.
2. Entertainment
3. Finance
Impact: Generative AI enhances fraud detection, portfolio management, and customer engagement.
4. Compare and contrast the pipeline of a traditional ML model with that of a generative AI model. What are
the key differences in their design and implementation? (Evaluating)
The pipeline of a traditional ML model and a generative AI model differ significantly in their goals, methodologies, and
architectures. Below is a detailed comparison and contrast of their design and implementation:
1. Objective
● Traditional ML Model:
○ Focuses on predictive tasks, such as classification, regression, or clustering.
○ Example: Predicting house prices or detecting spam emails.
● Generative AI Model:
○ Aims to create new data resembling the training data.
○ Example: Generating realistic images, text, or synthetic voices.
2. Pipeline Structure
3. Algorithm Design
● Traditional ML:
○ Primarily discriminative models that map inputs to outputs.
○ Example Algorithms: Logistic Regression, SVM, Random Forest, Neural Networks.
○ Loss Function: Single-objective (e.g., binary cross-entropy, RMSE).
● Generative AI:
○ Includes generative models like:
■ Variational Autoencoders (VAEs): Learn a probabilistic latent space.
■ Generative Adversarial Networks (GANs): Use a generator and discriminator in a competitive setup.
■ Transformer-based models (e.g., GPT): Sequence-to-sequence modeling for language tasks.
○ Loss Function: May involve multi-objective functions, e.g., adversarial loss (GANs) or KL divergence (VAEs).
4. Computational Complexity
● Traditional ML:
○ Models like Random Forest or SVM are computationally efficient for small to medium datasets.
○ Neural networks used in traditional tasks are simpler than generative ones.
● Generative AI:
○ High computational demands due to complex architectures like GANs or transformers.
○ Training generative models often requires extensive hardware resources (e.g., GPUs or TPUs).
5. Output
● Traditional ML:
○ Deterministic or probabilistic outputs, such as classification labels or regression values.
○ Example: "The likelihood of spam is 95%."
● Generative AI:
○ Stochastic and creative outputs that mimic human-generated content.
○ Example: A photorealistic image of a non-existent person or a coherent paragraph of text.
Data Dependence Relies on labeled data for Can work with unlabeled data
supervised tasks. (unsupervised or
semi-supervised).
5. Explain how data preprocessing and feature engineering differ between standard ML tasks and generative
AI tasks. (Understanding)
Data preprocessing and feature engineering are crucial components of machine learning workflows, but they differ significantly
in focus and approach between standard ML tasks and generative AI tasks due to the distinct objectives and methodologies of
these fields. Here's a breakdown of these differences:
1. Data Preprocessing
The goal of preprocessing is to prepare raw data into a format suitable for machine learning models.
Objective Ensure the data is clean, structured, Retain data fidelity to learn the true
and optimized for prediction tasks. underlying distribution.
Handling Missing Fill missing values using techniques Missing values are often retained unless they
Data like mean imputation or remove distort the underlying distribution.
incomplete records.
Scaling and Apply techniques like Min-Max Similar methods are used, but generative
Normalization Scaling or Standardization to ensure tasks may also require pixel value
feature comparability. normalization (e.g., [0, 1] for images).
Data Augmentation Rarely needed; primarily used in Commonly used to increase data diversity
imbalanced datasets for (e.g., image rotation, cropping, or noise
oversampling. addition) to improve model robustness.
Dimensionality Techniques like PCA or feature Retain full dimensionality to preserve data
Reduction selection to reduce complexity and richness, as generative tasks need detailed
improve computation. inputs to capture nuances.
2. Feature Engineering
Feature engineering involves creating, modifying, or selecting features to improve model performance.
Objective Derive relevant features that improve Not a primary focus, as generative
predictive accuracy. models often work directly with raw
data.
Manual Feature Extensive: domain expertise is used to Minimal: models like GANs, VAEs, or
Design create meaningful features (e.g., transformers learn features automatically
interaction terms, categorical encodings). from data.
Feature Selection Emphasis on selecting the most relevant Not typically required; generative
features to avoid overfitting and reduce models process all features to capture
computation. full distributions.
Data Convert data into structured tabular Represent data in its raw or minimally
Representation formats (e.g., one-hot encoding for processed form (e.g., raw pixel arrays for
categorical variables). images, tokenized text for language).
Key Differences
Category Standard ML Tasks Generative AI Tasks
Focus Optimize input data for predictive Preserve data richness for learning
accuracy. distributions.
Feature Dependence Relies heavily on engineered features to Relies on raw data; feature
capture relationships. extraction is model-driven.
Processing Goals Reduce complexity and improve Maintain maximum detail for
interpretability. realistic generation.
Preprocessing for Specific Use domain-specific techniques to Retain raw characteristics critical to
Domains simplify data. realistic outputs.
Examples
● Standard ML:
○ Predicting house prices: Normalize numerical features (e.g., square footage), encode categorical features (e.g.,
city), and drop irrelevant columns.
○ Importance of feature engineering: Create interaction terms like "rooms per square foot."
● Generative AI:
○ Generating images: Use raw pixel values normalized to [0, 1] or [-1, 1].
○ Importance of retaining detail: Ensure data augmentation does not distort original content (e.g., noise for
denoising autoencoders).
6. Discuss the role of loss functions in ML and generative AI pipelines. How do the objectives of these loss
functions differ and why? (Analyzing)
Loss functions are critical in machine learning (ML) and generative AI pipelines because they quantify the difference between
the model's predictions and the desired outcomes, guiding the optimization process during training. While both types of pipelines
rely on loss functions to improve performance, their objectives and implementations differ due to the nature of their tasks. Here’s
an analysis:
Objective:
In standard ML tasks, the loss function measures the model's error in predicting outputs. The primary goal is to minimize this
error to improve predictive accuracy.
● Regression Tasks:
○ Mean Squared Error (MSE): Penalizes large errors, encouraging precise predictions.
○ Mean Absolute Error (MAE): Treats all errors equally, providing robustness to outliers.
● Classification Tasks:
○ Cross-Entropy Loss: Measures the difference between predicted probabilities and true class labels, commonly
used in multi-class classification.
○ Hinge Loss: Used for SVMs to maximize the margin between classes.
Focus:
Objective:
In generative AI tasks, loss functions are designed to help the model learn the data distribution and generate outputs that closely
resemble the original data. This often requires optimizing multiple objectives simultaneously.
Focus:
Primary Goal Minimize prediction error for a specific Learn and replicate the underlying data
task. distribution.
Error Focus Focuses on deterministic error between Focuses on probabilistic differences and
predicted and actual values. visual/perceptual quality.
Evaluation Metrics like accuracy, R², Metrics like FID (Fréchet Inception Distance)
Metrics precision-recall. or BLEU score.
4. Examples of Differences in Practice
1. Describe the architecture of a GAN. How do the generator and discriminator interact during training?
(Understanding)
A Generative Adversarial Network (GAN) is a deep learning architecture consisting of two neural networks: a
generator and a discriminator, which work in opposition to each other during training. This adversarial setup
enables the generator to create increasingly realistic outputs over time.
b. Discriminator (D):
● Objective: Distinguish between real data from the training set and fake data produced by the generator.
● Input: Data samples (both real and generated).
● Output: A probability score indicating whether the input is real (close to 1) or fake (close to 0).
● Design:
○ Often uses convolutional layers for feature extraction (in the case of images).
○ Acts as a binary classifier trained to maximize its discrimination accuracy.
1. Discriminator Training:
○ The discriminator is trained on:
■ Real samples from the dataset, labeled as 1.
■ Fake samples generated by the generator, labeled as 0.
○ The discriminator’s loss is computed using binary cross-entropy:
○ Here, D(x) is the probability that x is real, and G(z) is the generator’s output.
2. Generator Training:
○ The generator aims to produce data that fools the discriminator, making D(G(z)) approach 1.
○ The generator’s loss is also computed using binary cross-entropy but with reversed labels:
3. Adversarial Process:
○ The generator and discriminator are trained alternately:
■ The discriminator learns to better classify real vs. fake data.
■ The generator improves to produce more realistic data, minimizing the discriminator’s
ability to differentiate.
4. Optimization:
○ Both networks are optimized using gradient-based methods (e.g., SGD, Adam).
○ The overall GAN objective is expressed as:
3. Training Workflow
1. Step 1: Train the discriminator:
○ Feed real samples x and fake samples G(z).
○ Update discriminator weights to maximize logD(x)+log(1−D(G(z)))
2. Step 2: Train the generator:
○ Generate fake samples G(z).
○ Update generator weights to minimize log(1−D(G(z))), or equivalently maximize logD(G(z))
3. Repeat: Alternate between the two steps for a set number of epochs.
4. Interaction Dynamics
● Adversarial Learning:
○ The discriminator pushes the generator to improve by identifying flaws in generated data.
○ The generator adapts to produce data that increasingly resembles real samples.
● Equilibrium:
○ Ideally, the generator produces outputs that the discriminator cannot distinguish from real data.
○ At this point, D(x)=D(G(z))=0.5, indicating equal likelihood of real and fake samples.
5. Challenges in Training
● Mode Collapse: The generator produces a limited variety of outputs.
● Non-convergence: The adversarial game may fail to stabilize.
● Vanishing Gradients: If the discriminator becomes too powerful, the generator receives minimal feedback.
6. Applications
● Image generation (e.g., StyleGAN).
● Data augmentation for medical images.
● Text-to-image generation (e.g., DALL-E).
● Audio synthesis (e.g., WaveGAN).
2. What are the main challenges associated with training GANs, and what strategies can be
employed to address these challenges? (Evaluating)
1. Mode Collapse:
GANs may produce limited diversity in generated data, where the generator maps multiple inputs to
the same output.
2. Vanishing Gradients:
During training, the discriminator may become too powerful, resulting in negligible gradients for
the generator, hindering its learning.
3. Training Instability:
GANs involve simultaneous optimization of two neural networks (generator and discriminator),
making the process unstable due to non-convex loss functions and adversarial interplay.
4. Evaluation Metrics:
Measuring the quality of GAN outputs is non-trivial, as there are no universally agreed-upon
quantitative metrics for evaluating image realism and diversity.
5. Resource Intensiveness:
GAN training requires significant computational resources and time, especially for high-resolution
data or complex architectures.
1. Feature Matching:
The generator’s loss can include a term encouraging it to produce data matching the intermediate
feature representations learned by the discriminator, reducing mode collapse.
2. Batch Normalization and Gradient Penalty:
These techniques stabilize training by ensuring smooth gradient propagation and penalizing extreme
gradients in the loss function.
3. Label Smoothing:
Smoothing the discriminator's labels (e.g., using 0.9 for "real" instead of 1.0) prevents it from
becoming overly confident and aids generator learning.
4. Improved Architectures:
Models like Wasserstein GANs (WGAN) introduce alternative loss functions and weight clipping to
enhance stability.
5. Regularization:
Techniques like adding noise to inputs, dropout, or using spectral normalization in the discriminator
improve robustness and convergence.
6. Progressive Growing:
Gradually increasing the resolution of the generated data helps stabilize training, especially for
high-resolution GANs like StyleGAN.
3. Compare and contrast different types of GANs, such as DCGANs, WGANs, and StyleGANs. How
do their architectures and applications differ? (Analyzing)
● Architecture:
○ Uses convolutional layers in both the generator and discriminator to learn spatial hierarchies.
○ Employs batch normalization for stable training.
○ Uses Leaky ReLU activation in the discriminator and ReLU in the generator.
● Advantages:
○ Suitable for generating high-quality images from simple datasets.
○ Introduced best practices for GAN design, such as avoiding fully connected layers.
● Limitations:
○ Struggles with mode collapse and instability during training.
● Applications:
○ Image generation, super-resolution, and style transfer.
● Architecture:
○ Replaces the discriminator with a "critic" that estimates the Wasserstein distance between
real and generated data distributions.
○ Uses weight clipping or gradient penalty for enforcing Lipschitz continuity.
● Advantages:
○ Improves training stability by addressing vanishing gradients.
○ Provides a meaningful loss metric correlating with the quality of generated samples.
● Limitations:
○ Weight clipping can sometimes lead to capacity issues. Gradient penalty variants
(WGAN-GP) alleviate this but increase computational cost.
● Applications:
○ Scenarios requiring stable GAN training, such as scientific data synthesis or generating
realistic images.
3. StyleGAN:
● Architecture:
○ Introduces a novel style-based generator architecture.
○ Employs a mapping network and adaptive instance normalization (AdaIN) to control style
features at different levels.
○ Uses progressive growing for high-resolution image synthesis.
● Advantages:
○ Capable of generating high-resolution, photorealistic images.
○ Allows disentangled control over image attributes (e.g., facial expressions, lighting).
● Limitations:
○ Computationally intensive.
● Applications:
○ Face generation, virtual avatar creation, and art synthesis.
Key Differences
Feature DCGAN WGAN StyleGAN
Conclusion
DCGAN laid the groundwork for convolutional GANs. WGAN improved stability and training dynamics,
while StyleGAN represents the state-of-the-art in generating controllable, high-resolution images. Each
type of GAN serves specific purposes, determined by the complexity of the task and the quality
requirements.
4. Define and differentiate between generative and discriminative models in machine learning. How
do their goals and methodologies vary?
Generative vs. Discriminative Models in Machine Learning
Generative Models
● Definition:
Models that learn to generate the underlying data distribution by modeling the joint probability
P(X,Y)P(X, Y)P(X,Y), where XXX represents features and YYY represents labels.
● Goal:
○ To understand the data distribution and generate new data samples similar to the training
data.
○ Predict P(Y∣X)P(Y|X)P(Y∣X) indirectly by leveraging the joint probability P(X,Y)P(X,
Y)P(X,Y).
● Methodology:
○ Focuses on learning how data is generated.
○ Examples: Gaussian Mixture Models (GMM), Variational Autoencoders (VAE), Generative
Adversarial Networks (GANs).
● Applications:
○ Data augmentation, image generation, speech synthesis, and anomaly detection.
Discriminative Models
● Definition:
Models that directly estimate the conditional probability P(Y∣X)P(Y|X)P(Y∣X), focusing on the
decision boundary between different classes.
● Goal:
○ To classify data points or make predictions without understanding the data distribution.
● Methodology:
○ Focuses on maximizing the separation between classes by learning features that differentiate
labels.
○ Examples: Logistic Regression, Support Vector Machines (SVM), Random Forest, Neural
Networks (for classification tasks).
● Applications:
○ Classification, regression, and other supervised learning tasks.
Key Differences
Aspect Generative Models Discriminative Models
Summary
Generative models aim to understand and replicate data, making them suitable for creative tasks, while
discriminative models are optimized for accuracy in prediction and classification. Their methodologies and
goals reflect their differing priorities in machine learning.
1. Explain the core principles of VAEs. How do VAEs generate new data samples from learned latent
representations? (Understanding)
5. Objective Function:
○ Combines two components:
■ Reconstruction Loss: Ensures the decoded output resembles the input (e.g., MSE for
images).
■ KL Divergence: Regularizes q(z∣x) to stay close to the prior p(z), often a standard normal
distributions
How VAEs Generate New Data Samples
Main Points:
2. Discuss the role of the encoder and decoder in a VAE. How does the variational aspect influence
the generation process? (Analyzing)
3. Compare VAEs with GANs in terms of their generative capabilities and common applications.
(Evaluating)
4. Compare and contrast the architectures of GANs and VAEs. How do their approaches to data
generation differ? (Evaluating)
Latent Space Does not explicitly define a latent Defines a latent space with a
space or prior distribution. prior (e.g.,
N(0,1)\mathcal{N}(0,1)N(0,1))
and samples from it.
Training Can be unstable and prone to More stable but may not
Stability mode collapse. capture all data complexities
as well as GANs.
1. Describe the Transformer architecture and its key components. How does the Transformer's
design address limitations found in earlier neural network models? (Understanding)
2. Explain the concept of self-attention in Transformers. How does it contribute to the model's
performance in handling sequential data? (Understanding)
3. Discuss the significance of positional encoding in Transformers. How does it enable the model to
understand the order of input data? (Analyzing)
4. Detail the process of multi-head attention in Transformers. How does it enhance the model’s
ability to capture different aspects of the input data? (Understanding)
5. Compare the attention mechanism used in Transformers with that of RNNs and LSTMs. What
are the computational benefits and limitations of each approach? (Evaluating)
6. Explain the concept of scaled dot-product attention. How does scaling affect the performance and
stability of the attention mechanism in Transformers? (Understanding)
7. Discuss the significance of positional encoding in transformers. (Understanding) 8. Define
self-attention (Understanding)
9. Discuss the impact of Transformers on natural language processing tasks. How have
Transformers changed the landscape of NLP research and applications? (Analyzing)
1. Define and differentiate between generative and discriminative models in machine learning. How
do their goals and methodologies vary? (Understanding)
2. Discuss the advantages and limitations of using generative models versus discriminative models
for tasks like image generation and classification. (Evaluating)
3. How do generative models approach the creation of new data, and how does this differ from the
approach taken by discriminative models in data classification? (Applying)
4. Outline the key components of a typical generative AI architecture. How do these components
work together to generate new data? (Understanding)
5. Compare the architectural differences between GANs, VAEs, and Transformers in the context of
generative AI. What are the strengths and weaknesses of each? (Evaluating)
6. Discuss the role of latent space in generative AI models. How is it used in both GANs and VAEs to
generate new samples? (Analyzing)
Word Embeddings
1. Explain the concept of word embeddings. How do they transform text into numerical vector
representations? (Understanding)
2. Discuss the applications of word embeddings in NLP. How do they improve tasks such as
sentiment analysis and machine translation? (Applying)
3. Using the sentence "The cat sat on the mat," discuss how the self- attention mechanism in
transformers works. Describe how self-attention focuses on the sentence to capture contextual
information effectively.
4. Compare different word embedding models like Word2Vec, GloVe, and FastText. What are their
unique features and use cases? (Analyzing)
5. Describe the training process for Word2Vec. How does it learn word representations from large
text corpora? (Understanding)
6. Explain the key differences between continuous bag of words (CBOW) and skip-gram models in
Word2Vec. (Understanding)
7. What is the process of transforming text into numerical representations using word embedding’s,
and how does it differ from traditional one-hot encoding? Provide an example. 8. Enlist the various
word embedding models used in NLP. Discuss any one in short 9. Discuss the role of context
windows in word embedding models. How do they influence the quality of word embeddings?
(Analyzing)
10. Discuss the process of transforming text into numerical vector representations through word
embedding’s. How do these embedding’s enhance sentiment analysis in social media data? 11. For
example, consider the following user reviews: "I loved the movie! It was fantastic and exciting."
"The movie was boring." Discuss how word embedding’s represent the words in these reviews
12. Provide an example of a sentiment analysis task using few-shot prompting. 13. Outline the major
steps involved in training a large language model. What are the key challenges associated with each
step?
1. Outline the major steps involved in training a large language model. What are the key challenges
associated with each step? (Understanding)
2. Discuss the impact of dataset size and diversity on the performance of an LLM. How do these
factors influence the model’s generalization capabilities? (Analyzing)
3. Explain the concept of fine-tuning in the context of LLMs. How does it differ from pre-training
and what are its benefits? (Understanding)
4. Provide an example of a tool or system that leverages Generative AI for video creation. How does
it integrate with other technologies and workflows to produce final video outputs? 5. Compare the
training process of a large language model with traditional machine learning models. How does
pre-training influence the effectiveness of LLMs?
6. Discuss the challenges associated with one-shot learning
7. In the context of sentiment analysis, analyze the benefits of using models that have been
pre-trained on large-scale dataset. How does this pre-training process contribute to improved model
performance?
8. Develop a step wise plan to fine-tune a pre-trained language model for a question-answering
system
9. What are the common evaluation metrics used for classification tasks in NLP? How do these
metrics measure model performance? (Understanding)
10. Consider a scenario where a large language model (LLM) is deployed in a customer support
system for an online retail company. The LLM is responsible for answering customer inquiries,
processing returns, and providing product recommendations.
11. Discuss what specific measures would you implement to ensure ethical use, prevent bias in
responses, and protect customer privacy while using the LLM?
12. Discuss the evaluation metrics for summarization tasks. What criteria are used to assess the
quality and relevance of summaries? (Analyzing)
13. Explain the evaluation metrics for question answering and text generation tasks. How do metrics
like BLEU, ROUGE, and F1-score differ in their application? (Understanding) 14. Describe the
concept of pre-training in the context of machine learning. How does it contribute to model
performance? (Understanding)
15. Discuss the benefits of transfer learning. How does leveraging pre-trained models enhance the
training of new models for specific tasks? (Analyzing)
16. Explain how transfer learning is applied in the context of LLMs. What are the advantages of
using pre-trained models for downstream tasks? (Understanding)
17. Detail the architecture of a typical Generative Adversarial Network (GAN). What roles do the
generator and discriminator play? (Understanding)
18. Discuss common training challenges associated with GANs, such as mode collapse and
convergence issues. What strategies can be used to overcome these challenges? (Analyzing) 19.
Explain how variations in GAN architecture, like Conditional GANs and Progressive Growing
GANs, address specific limitations or enhance performance. (Understanding) 20. Walk through a
basic TensorFlow code example for training a GAN to generate images. What are the main
components and their functions in image processing tasks? (Applying) 21. Discuss the role of loss
functions in the TensorFlow GAN tutorial. How do they influence the training process and output
quality? (Analyzing)
22. Identify the evaluation metrics commonly used for tasks such as classification, summarization,
and question answering. How do these metrics help in assessing the performance of LLMs? Justify
with your answer
23. Explain how data preprocessing and augmentation are handled in TensorFlow GAN examples.
Why are these steps critical for successful image generation? (Understanding)
Stable Diffusion
1. What is Stable Diffusion, and how does it fit into the landscape of generative models?
(Understanding)
2. Discuss the primary applications of Stable Diffusion. How does it compare with other image
generation models such as GANs and VAEs? (Analyzing)
3. Explain the significance of diffusion models in generating high-quality images. Why is Stable
Diffusion considered a breakthrough in this area? (Understanding)
4. Identify and describe the main components of the stable diffusion model. How does each
component contribute to the overall image generation process? (Understanding) 5. Explain the role
of the noise scheduler in stable diffusion. How does it influence the quality and characteristics of
generated images? (Understanding)
6. Discuss the function of the denoising network within the Stable Diffusion architecture. How does
it interact with the diffusion process to generate final images? (Analyzing)
7. Describe the training process for Stable Diffusion. What are the key stages and components
involved in training the model? (Understanding)
8. Discuss the types of data required for training Stable Diffusion. How does the choice of training
data affect the performance and versatility of the model? (Analyzing)
9. Explain the role of loss functions and optimization techniques in training Stable Diffusion. How
do these elements impact the model’s ability to generate high-quality images? (Understanding) 10.
Explain the process of generating images from noise in Stable Diffusion. How does the model
convert random noise into coherent images? (Understanding)
11. Discuss how prompts are used during the inference phase of Stable Diffusion. How does the
model interpret and integrate prompts to guide the image generation process? (Analyzing) 12.
Describe the iterative process of diffusion and denoising in stable diffusion. How does this process
contribute to refining and enhancing the generated images? (Understanding) 13. Identify and
explain common methods used to enhance the performance of stable diffusion. What techniques are
employed to improve image quality and generation speed? (Analyzing) 14. Discuss the tools and
frameworks commonly used for implementing and experimenting with stable diffusion. How do
these tools facilitate the development and deployment of diffusion models? (Analyzing)
15. Explain how stable diffusion can be integrated with other technologies or tools (e.g., APIs, user
interfaces) for practical applications. (Understanding)
16. Identify and describe the main components of the Stable Diffusion model. How does each
component contribute to the overall image generation process?
17. Compare different versions of stable diffusion. What are the key differences and improvements
between versions? (Analyzing)
18. Discuss the advancements introduced in newer versions of Stable Diffusion. How do these
advancements impact the model’s capabilities and performance? (Analyzing) 19. Explain the
reasons behind the development of various versions of Stable Diffusion. What specific challenges or
limitations were addressed in each version? (Understanding) 20. Describe some advanced techniques
used to enhance Stable Diffusion’s image generation capabilities. How do these techniques address
specific challenges in the diffusion process? (Analyzing)
21. Compare different versions of stable diffusion. What are the key differences and improvements
between versions?
22. Discuss the role of hyperparameter tuning in optimizing stable diffusion. What are the key
hyperparameters, and how do they affect the model’s performance? (Evaluating)
23. Discuss the tools and frameworks commonly used for implementing and experimenting with
stable diffusion
24. Explain how stable diffusion can be customized for specific applications or use cases. What are
some examples of advanced modifications orDiscuss the process of generating images from noise
using Stable Diffusion. extensions of the base model? (Applying)
25. Evaluate the benefits and limitations of using Stable Diffusion for content generation compared
to other generative models
Generative AI Applications
1. Discuss the impact of Generative AI on the healthcare industry. How are generative models used
for drug discovery, personalized medicine, and medical imaging? (Analyzing)
2. Explain how Generative AI is transforming the entertainment industry. What are some
applications in content creation, including music and video games? (Analyzing)
3. Explore the use of Generative AI in the fashion industry. How are generative models utilized for
design, trend prediction, and virtual try-ons? (Analyzing)
4. Analyze the role of Generative AI in the automotive industry. What are its applications in
autonomous driving, vehicle design, and simulation? (Analyzing)
1. How can LLMs be utilized for analyzing financial disclosures such as 10-K reports? Provide a
case study example where LLMs have been used to extract insights or perform analysis on financial
statements. (Applying)
2. Discuss the role of Generative AI in fundamental analysis of stocks. How do LLMs assist in
predicting stock performance and generating financial forecasts? (Analyzing)
3. Explain the challenges and benefits of using LLMs for Q&A on financial disclosures. How do
these models handle complex financial jargon and large volumes of data? (Evaluating) 4. What is
LangChain, and how does it facilitate the creation of domain-specific chatbots? Describe its core
features and benefits. (Understanding)
5. What is LangChain, and how does it facilitate the creation of domain-specific Chatbots? Describe
its core features and benefits
6. Walk through the process of building a domain-specific chatbot using LangChain. What are the
key steps involved, and how does LangChain integrate with various data sources? (Applying)
7. Discuss the importance of fine-tuning and customization when creating a domain-specific chatbot.
How does LangChain support these modifications? (Analyzing)
8. Explain how Generative AI can be used to generate news articles based on a given headline. What
are the key components and techniques involved in this process? (Understanding)
9. Discuss the challenges of generating coherent and accurate news articles using AI. How do
generative models handle factual accuracy and stylistic consistency? (Analyzing)
10. Provide an example of a system that generates news articles from headlines. How does it ensure
that the generated content is relevant and high-quality? (Applying)
11. Describe the process of summarizing speeches using Generative AI. What techniques are
employed to extract and condense key points from spoken content? (Understanding)
12. Discuss the challenges in summarizing speeches compared to other forms of text. How does
Generative AI address issues such as context retention and summarization accuracy? (Analyzing)
13. Provide a case study where AI has been used to summarize speeches or lectures. What were the
outcomes and limitations of the approach? (Evaluating)
14. Explain the concept of few-shot prompting in the context of sentiment analysis. How does it
allow models to perform sentiment classification with limited examples? (Understanding)
15. Discuss the benefits and limitations of using few-shot prompting for sentiment analysis. How
does it compare to traditional supervised learning approaches? (Evaluating)
16. Provide an example of a sentiment analysis task using few-shot prompting. How did the model
perform, and what were the key factors influencing its accuracy? (Applying)
17. Describe the role of Generative AI in video creation. What are some of the key techniques used
to generate or enhance video content? (Understanding)
18. Discuss the process of generating video content using Generative AI. What are the challenges
associated with maintaining coherence and visual quality in AI-generated videos? (Analyzing)
19. Provide an example of a tool or system that leverages Generative AI for video creation. How
does it integrate with other technologies and workflows to produce final video outputs? (Applying)
20. What is LangChain, and how does it facilitate the creation of domain-specific Chatbots?
Describe its core features and benefits (Applying)
Unit 4
1. What is prompt engineering, and why is it important in the context of generative AI chatbots?
2. Explain the role of training data in the architecture of large language models (LLMs) used in
chatbots.
3. Describe the basic architecture of a convolutional neural network (CNN).
4. What are retrieval-augmented generation (RAG) systems, and how do they differ from standard
LLMs?
5. What is fine-tuning in the context of LLMs?
6. How does a Recurrent Neural Network (RNN) handle sequential data?
7. Discuss how Generative AI can be used to generate news articles based on a given headline. What
are the key components and techniques involved in this process?
8.Analyse the important key considerations when choosing training data for fine-tuning an LLM for
medical diagnostics?
9. Given a scenario where a chatbot is underperforming, analyze how prompt engineering can be
used to improve its responses.
8. Evaluate the impact of training data quality on the performance of BARD or ChatGPT..
10. Apply your knowledge of CNNs to design a basic image classifier using TensorFlow. What are
the key components of your model?
11. Evaluate the effectiveness of fine-tuning LLMs for domain-specific tasks. What are the pros and
cons?
12. Analyze how RAG systems can enhance the quality of information provided by a chatbot
compared to traditional LLMs.
13. How would you optimize an LLM's performance if it's generating incomplete or irrelevant
responses?
14. Apply the concept of training data preprocessing to improve the accuracy of an RNN model.
14. Evaluate the impact of training data quality on the performance of BARD or ChatGPT.
15. How can Variational Autoencoders (VAEs) be used for data generation? Provide a real-world
application.
16. You are tasked with improving a generative AI Chatbots ability to engage users in casual
conversation. Develop a set of prompts that could be used to enhance its conversational flow.
17. Compare the approaches of using CNNs and RNNs for different types of data. What are the
benefits and limitations of each?
18. How would you assess the performance of an LLM before and after fine-tuning? What metrics
would you consider?
19. Analyze the trade-offs between using pre-trained models versus training an LLM from scratch.
20. How can you use TensorFlow to implement a basic VAE for anomaly detection? Provide an
overview of the process.
21. Evaluate the challenges and solutions in optimizing the performance of RAG systems.
22. Discuss the role of architectural choices (e.g., transformer layers, attention mechanisms) in
determining the performance of LLMs.
23. Apply prompt engineering techniques to design a set of prompts that improve the performance
of a customer service chatbot.
24. Apply your understanding of neural networks to design an RNN that predicts stock prices based
on sequential historical data.
25. What are the key considerations when choosing training data for fine-tuning an LLM for
medical diagnostics?
26. Discuss the concept of Retrieval-Augmented Generation (RAG) systems. How do they enhance
the capabilities of large language models in providing accurate and relevant responses
25. Design a prompt engineering strategy for developing a chatbot that provides mental health
support.
26. Create a tutorial code example using TensorFlow that demonstrates the use of RNNs for text
generation.
27. How would you build a RAG system that leverages a knowledge base for technical support
chatbots? Outline the key components.
28. Design a step-by-step process for fine-tuning an LLM for a legal document review application.
29. Develop a set of metrics for evaluating the performance of generative AI chatbots in customer
service settings.
30. Create a prototype architecture for an LLM-based chatbot that integrates both
retrieval-augmented generation and real-time user feedback for continuous learning.
31. Discuss the concept of Retrieval-Augmented Generation (RAG) systems. How do they enhance
the capabilities of large language models in providing accurate and relevant responses?
32. Analyse the key considerations when choosing training data for fine-tuning an LLM for medical
diagnostics?
Unit 5
1. What is Stable Diffusion, and how does it differ from traditional generative models?
7. How can Stable Diffusion be applied to generate realistic images from textual descriptions?
Provide an example.
8. Analyze the importance of training data quality when training a Stable Diffusion model.
9. Compare the different versions of Stable Diffusion. What improvements or changes were made in
each iteration?
10. Evaluate the benefits and limitations of using Stable Diffusion for content generation compared
to other generative models.
11. How would you optimize the performance of a Stable Diffusion model for faster image
generation during inference?
12. Apply the principles of Stable Diffusion to create a basic workflow for generating images using a
specific prompt.
13. What are some advanced techniques to improve the output quality of images generated by Stable
Diffusion?
14. Analyze the impact of adjusting noise levels during the image generation process in Stable
Diffusion.
15. How can prompt engineering enhance the results produced by Stable Diffusion models? 16.
Evaluate the challenges involved in training a Stable Diffusion model from scratch. What are some
strategies to overcome these?
17. Given a use case for generating medical imaging data, how would Stable Diffusion be applied,
and what considerations must be taken into account?
18. Explain the steps needed to deploy a Stable Diffusion model for real-time image generation in a
web application.
19. How can Stable Diffusion be integrated into existing creative workflows (e.g., digital art, game
design)?
20. Analyze the advantages of using different noise schedules during the training of Stable Diffusion
models.
21. Evaluate the role of GPUs and TPUs in the training and inference processes of Stable Diffusion.
22. How can Stable Diffusion be fine-tuned for a specific domain (e.g., anime-style art, architectural
designs)?
23. Compare the output quality of images generated by different versions of Stable Diffusion. What
factors contribute to the differences?
24. What are the ethical considerations when using Stable Diffusion for image generation?
25. Design a detailed step-by-step guide for training a stable diffusion model using custom datasets.
26. Create a tutorial on how to implement stable diffusion inference to generate images from
prompts using popular tools and libraries.
27. Propose a method to improve the stability and quality of images generated by a Stable diffusion
model during inference.
28. Develop an architecture for a web-based application that uses stable diffusion to convert textual
descriptions into images.
29. Design a custom noise scheduling technique to enhance the image generation process of stable
diffusion. Explain how it improves the results.
30. Create a comprehensive strategy for fine-tuning a pre-trained stable diffusion model for
generating product images in an e-commerce platform.
Unit 6
1. What are some key applications of Generative AI in the finance industry?
Fraud Detection: Identifying unusual patterns or activities to prevent financial fraud.
Risk Management: Generating synthetic data for stress testing and risk assessment.
Algorithmic Trading: Enhancing strategies by generating market scenarios and predictions.
Customer Support: Creating chatbots for personalized financial assistance.
Document Processing: Automating extraction, summarization, and analysis of financial reports.
Credit Scoring: Generating insights for predicting creditworthiness.
Portfolio Optimization: Simulating market conditions for investment strategies.
2. Describe how generative AI can be used for content creation, specifically in text generation.
Generative AI enables text generation by leveraging advanced language models to create human-like and
contextually relevant text. Key applications include:
1. Content Writing: Automatically generating articles, blogs, and social media posts tailored to
specific topics or audiences.
2. Summarization: Creating concise summaries of lengthy documents, reports, or articles.
3. Translation: Producing high-quality translations while preserving the original context and tone.
4. Creative Writing: Generating stories, poems, or scripts for entertainment or educational purposes.
5. Marketing Copy: Crafting product descriptions, advertisements, and personalized emails for
campaigns.
6. Personalized Messages: Drafting tailored communications for individual users in customer support
or engagement.
7. Academic Assistance: Producing explanations, research overviews, and study guides based on user
inputs.
Generative AI models like GPT-4 achieve this by analyzing prompts, predicting the next word sequence,
and ensuring coherence and fluency in the output.
3. What is LangChain, and how does it facilitate the creation of domain-specific chatbots?
LangChain is a framework designed for developing applications that integrate language models like GPT.
It simplifies building domain-specific chatbots by providing tools to manage and interact with external data
sources and knowledge bases.
LangChain thus enables the creation of chatbots tailored to specific industries or use cases, such as finance,
healthcare, or customer support, by combining large language model capabilities with domain-relevant
data.
Few-shot prompting in sentiment analysis refers to using a pre-trained generative language model (like
GPT) to perform sentiment analysis with minimal task-specific training data. Instead of retraining the
model, a few examples (prompts) are provided in the input to guide the model's understanding of the task.
1. Pre-Trained Model: A language model, pre-trained on vast datasets, is used without fine-tuning.
2. Task Instructions: The prompt includes instructions explaining the sentiment analysis task.
3. Few Examples: A small number (1-5) of labeled examples (e.g., sentences with their sentiment
labels) are included in the prompt.
4. Inference: The model uses the provided examples to infer the sentiment of new, unseen text.
Example:
Prompt:
"The task is to identify the sentiment of a sentence as Positive, Negative, or Neutral.
Example 1: 'The movie was amazing!' -> Positive
Example 2: 'I didn't like the food.' -> Negative
Example 3: 'The weather is okay.' -> Neutral
Sentence: 'The product quality is excellent!' ->"
Benefits:
Generative AI models used for video creation consist of several primary components that work together to
generate high-quality and coherent video content. These components include:
1. Input Encoder:
○ Processes input data (e.g., text descriptions, images, or audio) and converts it into a feature
representation.
○ Examples: Text encoders for captions or image encoders for reference frames.
2. Latent Space Representation:
○ Encodes video-specific attributes like motion, style, and scene details into a compact,
learnable format.
○ Enables consistent and high-quality video synthesis across frames.
3. Temporal Modeling Component:
○ Captures motion dynamics and ensures temporal coherence between frames.
○ Often implemented using architectures like Recurrent Neural Networks (RNNs),
Transformers, or 3D Convolutional Networks.
4. Generator:
○ Produces the video frames based on latent representations and temporal modeling.
○ Typically implemented using GANs (e.g., VideoGAN), Variational Autoencoders (VAEs),
or diffusion models.
5. Discriminator (for GAN-based models):
○ Evaluates the generated video for realism, ensuring visual and temporal consistency.
○ Trained adversarially alongside the generator.
6. Loss Functions:
○ Guides the model during training, focusing on factors like realism, coherence, and alignment
with input prompts.
○ Examples: Adversarial loss, perceptual loss, temporal consistency loss.
7. Post-Processing Module:
○ Refines and enhances generated video frames, improving quality and adding details like
color correction or noise reduction.
8. Audio Integration Module (optional):
○ Synchronizes audio with video for applications like animated storytelling or music videos.
These components work together, enabling generative AI models to create high-quality, contextually
relevant videos tailored to specific inputs or use cases.
6. How can LLMs assist in fundamental analysis of financial disclosures, such as 10-K filings?
Large Language Models (LLMs) can significantly assist in the fundamental analysis of financial
disclosures, like 10-K filings, by leveraging their natural language understanding and data processing
capabilities. Here's how:
1. Text Summarization:
○ Extract key insights and summarize lengthy sections like risk factors, management
discussion, and financial highlights.
○ Saves analysts time by providing concise overviews.
2. Sentiment Analysis:
○ Evaluate the tone of disclosures to gauge management sentiment (e.g., optimistic or
cautious).
○ Detect subtle changes in language over time to assess risk or performance outlook.
3. Key Entity Extraction:
○ Identify and extract financial metrics, ratios, or industry-specific terms from filings.
○ Highlight entities like competitors, market trends, or legal risks.
4. Comparison and Trend Analysis:
○ Compare disclosures year-over-year or across companies to identify significant changes or
trends.
○ Detect deviations from industry norms or company benchmarks.
5. Question Answering:
○ Provide specific answers to queries like "What are the company's main risks?" or
"Summarize the revenue breakdown by segment."
6. Risk Identification:
○ Highlight potential red flags, such as debt concerns, legal issues, or supply chain risks
mentioned in the filings.
7. Automated Compliance Checks:
○ Ensure that disclosures meet regulatory requirements by identifying missing or inconsistent
information.
8. Thematic Analysis:
○ Group content into themes (e.g., ESG factors, market strategy) to assess alignment with
company goals and investor interests.
LLMs like GPT can process vast amounts of data, uncover patterns, and deliver actionable insights, making
fundamental analysis faster, more efficient, and more comprehensive.
7. Analyze a case study where LLMs were used to automate Q&A on financial disclosures. What
were the key findings?
Overview:
A financial organization implemented a Large Language Model (LLM), such as GPT-4, to automate
question-answering (Q&A) on financial disclosures like 10-K and 10-Q filings. The goal was to improve
efficiency, accuracy, and accessibility of critical information for stakeholders.
1. Data Preparation:
○ Financial disclosures were preprocessed into machine-readable formats.
○ A knowledge base was created using document embeddings for efficient retrieval.
2. Model Deployment:
○ A Retrieval-Augmented Generation (RAG) system was used:
■ Retrieval Component: Extracted relevant text from filings based on user queries.
■ Generative Component: The LLM generated concise, natural-language answers.
3. Training and Fine-Tuning:
○ Domain-specific fine-tuning was performed using historical disclosures and financial Q&A
datasets.
Key Findings:
1. Improved Efficiency:
○ Analysts spent 60% less time retrieving information compared to manual searches.
○ Automated responses were generated in seconds, reducing turnaround time significantly.
2. High Accuracy:
○ The LLM achieved an accuracy of ~85-90% for domain-specific questions, verified against
expert-curated answers.
○ Consistently retrieved the most relevant sections of disclosures.
3. Enhanced User Experience:
○ Users appreciated the conversational interface, which allowed them to ask questions in
natural language.
○ Example: "What are the company’s main risks this year?"
4. Scalability:
○ The system handled thousands of queries simultaneously, making it viable for large-scale
deployments.
5. Challenges Identified:
○ Complexity of Legal Language: Some nuances in legal and financial terminology required
further model refinement.
○ Model Interpretability: The model occasionally generated ambiguous responses,
necessitating human oversight.
○ Data Freshness: Ensuring the LLM operated on up-to-date filings was critical.
6. Cost Savings:
○ Significant reduction in operational costs by automating repetitive Q&A tasks.
Conclusion:
LLMs proved to be a transformative tool for automating Q&A on financial disclosures, enhancing both
efficiency and accessibility. However, fine-tuning for domain specificity, continuous monitoring for
accuracy, and handling legal language nuances remain essential for optimal performance.
8. How would you implement a generative AI model to generate news articles from given headlines?
Outline the steps.
● Goal: Generate coherent and contextually relevant news articles from given headlines.
● Requirements: High-quality, informative, and grammatically correct content adhering to journalistic
standards.
● Data Sources: Gather a large dataset of news articles and their corresponding headlines from reliable
sources (e.g., news APIs, public datasets).
● Preprocessing:
○ Clean the data (remove HTML tags, special characters, etc.).
○ Tokenize and normalize text.
○ Pair headlines with their articles for supervised learning.
● Use a pre-trained Large Language Model (LLM) such as GPT, T5, or BART for text generation.
● Fine-tune the model for headline-to-article generation.
● Develop a user interface where users input headlines and receive generated articles.
● Pipeline:
○ Input headline → Preprocessing → LLM Inference → Postprocessing → Output article.
8. Post Processing
9. Iterative Improvements
● Regularly update the model with newer data to improve relevance and accuracy.
● Address issues like hallucination or bias using reinforcement learning or additional fine-tuning.
● Ensure generated articles are factually accurate and free from misinformation.
● Implement safeguards to avoid misuse, like generating false or misleading content.
1. Contextual Understanding:
○ Generative AI:
■ LLMs like GPT can understand the context and nuances of speeches, making
summaries more comprehensive and aligned with the speaker's intent.
○ Traditional Methods:
■ Often rely on predefined rules or keyword-based approaches, leading to less nuanced
summaries.
2. Language Fluency:
○ Generative AI:
■ Produces summaries in coherent and natural-sounding language, suitable for diverse
audiences.
○ Traditional Methods:
■ May generate fragmented or awkwardly phrased outputs.
3. Adaptability:
○ Generative AI:
■ Easily adapts to different speech styles, topics, and domains through fine-tuning or
prompt engineering.
○ Traditional Methods:
■ Requires significant manual reprogramming for domain-specific applications.
4. Summarization Styles:
○ Generative AI:
■ Supports abstractive summarization, creating summaries in novel phrasing while
preserving meaning.
○ Traditional Methods:
■ Primarily extractive, limiting summaries to exact text snippets without paraphrasing.
5. Handling Ambiguity:
○ Generative AI:
■ Resolves ambiguous or incomplete speech content using contextual inference.
○ Traditional Methods:
■ Struggle with ambiguity, often resulting in incomplete or misleading summaries.
1. Factual Accuracy:
○ Generative AI models may "hallucinate" or introduce inaccuracies in summaries.
○ Traditional methods, being extractive, are less prone to such errors.
2. Dependence on Training Data:
○ Generative AI requires large, high-quality datasets to ensure effective summarization.
○ Traditional methods are less data-intensive.
3. Computational Requirements:
○ Generative AI models are resource-heavy, requiring GPUs/TPUs for training and inference.
○ Traditional methods are computationally lightweight.
Effectiveness in Practice
10. Discuss the challenges faced when training generative AI models for financial disclosures and
how they can be addressed.
11. How can sentiment analysis with few-shot prompting be applied in social media monitoring for
financial trends?
12. Apply your understanding of LangChain to design a domain-specific chatbot for a legal firm.
What features would you include?
13. Compare the output of different generative AI models in generating financial reports. What
factors influence the quality of the results?
14. Evaluate the potential risks and ethical considerations associated with using generative AI for
content creation in journalism.
15. How can generative AI enhance customer service in finance through chatbots? Provide an
example.
16. Analyze the impact of generative AI on improving the efficiency of fundamental analysis in
finance.
17. Propose methods to measure the accuracy and relevance of news articles generated by a
generative AI model.
18. How would you assess the performance of a generative AI system in creating video content for
marketing?
19. Discuss the role of feedback loops in improving the performance of generative AI applications in
sentiment analysis.
20. Evaluate the effectiveness of few-shot prompting in understanding complex financial jargon
during sentiment analysis.
21. How can generative AI models be adapted to accommodate various financial regulatory
requirements in their applications?
22. Analyze the relationship between user engagement and the quality of content generated by
domain-specific chatbots.
23. How can the output from generative AI in video creation be validated for accuracy and
reliability?
IMP QUESTIONS
1. GAN and its architecture
Definition:
Generative Adversarial Networks (GANs) are a class of generative models that consist of two neural
networks, a Generator (G) and a Discriminator (D), trained in a competitive setting. GANs aim to
generate realistic data samples by modeling the underlying data distribution.
Architecture of GANs
1. Generator (G):
○ Purpose: Generates synthetic data samples resembling the real data.
○ Input: A random noise vector zzz sampled from a latent space (e.g., Gaussian or uniform
distribution).
○ Output: A data sample G(z)G(z)G(z) (e.g., an image, audio, or text).
○ Structure: Typically uses transposed convolutional layers or dense layers to upsample the
input noise vector into the desired data format.
2. Discriminator (D):
○ Purpose: Classifies inputs as "real" (from the dataset) or "fake" (generated by GGG).
○ Input: Either real data XXX or synthetic data G(z)G(z)G(z).
○ Output: A probability score indicating whether the input is real or fake.
○ Structure: Consists of convolutional layers or dense layers to downsample and extract
features from the input.
Training Process
1. Discriminator's Objective:
○ Maximizes its ability to correctly classify real and fake samples.
○ Loss: LD=−EX∼Pdata[logD(X)]−Ez∼Pz[log(1−D(G(z)))].\mathcal{L}_D =
-\mathbb{E}_{X \sim P_{\text{data}}}[\log D(X)] - \mathbb{E}_{z \sim P_z}[\log(1 -
D(G(z)))].LD=−EX∼Pdata[logD(X)]−Ez∼Pz[log(1−D(G(z)))].
2. Generator's Objective:
○ Tries to "fool" the discriminator into classifying generated samples as real.
○ Loss: LG=−Ez∼Pz[logD(G(z))].\mathcal{L}_G = -\mathbb{E}_{z \sim P_z}[\log
D(G(z))].LG=−Ez∼Pz[logD(G(z))].
3. Adversarial Training:
○ The generator and discriminator are trained alternately.
○ The discriminator improves in distinguishing real from fake samples, while the generator
improves in producing realistic data.
1. Adversarial Framework:
○ Training is a zero-sum game where improvements in one network challenge the other.
2. Latent Space Mapping:
○ Random noise vectors from the latent space are transformed into realistic data by the
generator.
3. Unsupervised Learning:
○ GANs do not require labeled data, relying instead on learning the data distribution.
Applications of GANs
1. Image Generation:
○ Creating realistic images (e.g., faces, art, or scenes).
2. Data Augmentation:
○ Generating synthetic data for training machine learning models.
3. Style Transfer:
○ Modifying the style of an image while preserving its content.
4. Video and Audio Synthesis:
○ Generating lifelike animations, music, or speech.
GANs are powerful but challenging to train due to issues like mode collapse and instability. Proper
architectural design and regularization techniques can mitigate these challenges.