Generative AI
Generative AI
An Overview
Understanding Recurrent Neural Networks (RNNs)
RNNs are a type of neural network.
They are designed to process
sequential data.
These architectures were widely
used for NLP tasks, speech
processing, and time series.
Challenge-?
The Rise of Transformers: Self-Attention
In 2017, researchers at Google
published a paper that proposed a
novel neural network architecture
for sequence modeling known as
Transformer.
Outperformed recurrent neural
networks (RNNs) on machine
translation tasks, both in terms of
translation quality and training cost.
A Timeline of Large Language Models
2022: ChatGPT
Generative Pre-trained Transformer 2.
2024: Meta's Llama 3, Claude 3, and Q2, and Mistral's Mixtral 8x7B
Larger and more powerful model.
2025: DeepSeek-R1
Multimodality: Text, Image, Video
Diving into ChatGPT
Generative Pre-trained Transformer
Next word prediction LLM is pre-trained on massive Encoder-decoder architecture
amount of text