Large Large Models
Large Large Models
com/in/thota-adinarayana/
1
Introduction to Large Language Models
2 https://www.linkedin.com/in/thota-adinarayana/
Introduction to Large Language Models
• Level of linguistic knowledge
Not applicable
Not applicable
Great. LLMs are robust to typos, coinage, cacography
These examples are copied from Stanford CS224N/Ling284 slides (author: John Hewitt).
They are actually examples for masked language modeling which is a bit different from how GPT is pre-trained. 25
From GPT-3 to ChatGPT:
Learn human intents behind their language
OpenAI
6 “Aligning language models to follow instructions” 2022
Follow Instructions & Align with Human Preference
Human-in-
the-loop!
(Discuss more
later)
Hugging
9 face “Large Language Models: A New Moore's Law?” 2021
Introduction to Large Language Models
• In-context learning
• No parameter update
• Wrap “training” samples in the prompt
11
Introduction to Large Language Models
• Emergence abilities
• An ability is emergent if it is not present in smaller models but is present in
larger models.
• In-context learning ability is one of them.
• Scaling to improve unlock abilities.
37
Yao Fu “How does GPT Obtain its Ability? Tracing Emergent Abilities of Language Models to their Sources”
Unix Family Tree
Ecosystem Graphs
Put LLMs into a Larger System
• Analogy: operating system (OS)
• How do we interact with OS?
• How do we interact with LLMs?
16
Put LLMs into a Larger System
• Analogy: operating system (OS)
• How do we interact with OS?
• How do we interact with LLMs?
17
Put LLMs into a Larger System
AutoGPT demo 42
Put LLMs into a Larger System
LLM functions as a controller
and can use tool on its own.
19 https://openai.com/blog/chatgpt-plugins
LLM as a Controller
I’m inspired
20 by https://oval.cs.stanford.edu/ to add this illustration.
LLM as a Controller: Challenges
• How to design the interaction interface between LLMs and other
components (e.g., external databases, API schemas)?
• Desiderata:
robustness, unambiguity, privacy-protecting, easy-to-build for non-AI developers
I’m inspired
21 by https://oval.cs.stanford.edu/ to add this illustration.
Bring Human into the Loop
• Returning to the OS analogy
• What’s special with LLMs?
• LLMs can learn from the human-
model interaction and evolve.
22
Bring Human into the Loop
Core challenges:
• How can we let human easily provide
feedback?
• Exploiting cheap labor is unethical and
infeasible to collecting domain-specific
feedbacks.
• I think research from the HCI side is
important.
• How can we let the LLM take
feedback?
• Current approach: RLHF
• What’s next? (distinct challenges exist)
24
25