LLM - A Introduction To Generative AI
LLM - A Introduction To Generative AI
Introduction to Generative AI
AI and Society
- Generative AI
It covers practical content on how to use generative AI to generate answers, explain issues
clearly and comprehensively, develop new ideas and implementation plans, create realistic
content, solve problems creatively, and get things done quickly. The concluding section
Skills for explains how generative AI will evolve in the near future, towards AGI. Hence, it is key to
Future Jobs learn how to control and exploit this new kind of AI power.
• how to use generative AI to generate insights, analytics, explanations and also creative ideas that can be used to
solve problems.
• how to use generative AI to automate tasks, which can free up time for other activities.
• use generative AI to generate realistic content, which can be used to create prototypes or presentations.
• POE.COM –
uses
multiple LLM
in 1 site.
• https://poe.com/
History of GPTs
Inventing Transformers created the base
• Transformers are a type of AI that have been shown to be very
effective for natural language processing tasks, such as text
generation, translation, and question answering. They quickly
became new AI models that are super large and can thus
understand and generate text in a human-like way.
• Its attention mechanism allows them to learn long-range
dependencies between words in a sentence. This makes them very
well-suited for tasks that require understanding the context of a
sentence, such as text generation.
• First introduced in the paper "Attention is All You Need" by
Vaswani et al. (2017). Since then, they have become the baseline
architecture for models such as GPT-3, T5 etc. These models have Replacing older AI methods for:
achieved impressive results on a variety of tasks, including • Forecasting
generating realistic and coherent text, translating languages, and • NLP data recognition from
answering questions in an informative way.
relationship to entities and
• Previous AI image, audio
recognition.
Transformers: Pros and Cons
Applicable to all GPTs
• Advantages of transformers:
• Can learn long-range dependencies between words in a sentence.
• Very efficient, do its work in parallel paths.
• Can be fine-tuned for specific tasks, making them very versatile.
Disadvantages of transformers:
• Computationally expensive to train.
• Difficult to understand and interpret. This is known as Explainability.
• Not always able to generate text that is logically correct or factually
accurate.
Definitions
Foundation Models
Pre-trained base models, trained on a massive dataset of text and code, e.g. everything is on the Internet.
This allows them to learn the statistical relationships between words and concepts, images etc and can
thus do super human tasks.
Token: A chunk of text that the model reads or generates. A token is typically not a word; it
could be a smaller unit, like a character or a part of a word, or a larger one like a whole
phrase. Measurement for output, e.g. 50 tokens /sec. Roughly, one token = 0.7 x word.
Definitions
Finetuning
Adjusting the parameters of a pre-trained language model to improve its performance on a specific
task, just by updating a small subset.
Temperature
Parameter that controls the level of creativity in the text that the model generates. A higher
temperature results in more creative text, while a lower temperature results in more accurate text.
Prompting (Prompt Engineering)
Using large language models to generate creative text formats by providing them with a prompt
that describes the desired output. E.g. Prompt: Write a poem about a flower that is dying.
Context (size is important, how many Kbytes e.g. 2K or 8K)
refers to the information that is provided to the model as input, which it uses to generate its
output. Context can include a wide range of information, such as the text of a conversation, the
user’s previous interactions with the model, or any other relevant information that can help the
model understand the user’s intent and generate an appropriate response.
Narrow AI
Conventional AI that does just one task. But now, new models can do multiple tasks.
Or
• AGI – Artificial General Intelligence Going to end
the world…
• https://youtu.be/Mqg3aTGNxZ0 It
Media:
6. Automated article generation for regular events like sports matches
Text Summarization
7. Creating newsletters
8. Summarizing internal company documents
• Types of models: quantised (faster like got compressed, via TheBloke in different bits)
• Names: animals e.g. llama, alpaca, dolphin, orca etc.
Open Source versus Closed Source LLM
• Open source and closed source refer to the availability of the source code of a software
program. Open source software is software whose source code is made available to the
public, allowing anyone to view, modify, and distribute the code. Closed source software,
on the other hand, is software whose source code is not made available to the public and is
typically owned by a company or individual.
• In the context of large language models (LLMs), an open source LLM would be one whose
code and training data are publicly available, allowing anyone to view, modify, and use the
model. A closed source LLM would be one whose code and training data are not publicly
available and are typically owned by a company or individual.
• The commercial usage of open source and closed source LLMs can vary. Open source LLMs
can be used by anyone for any purpose, including commercial purposes, as long as the
terms of the open source license are followed. Closed source LLMs, on the other hand, are
typically only available for commercial use through licensing agreements with the owner of
the model.
• E.g. Meta’s LLama2 is free until customer number reaches 700 million accounts (for
companies).
Debate about open source developing faster
• According to a leaked memo from a Google employee, the company’s
advantage in artificial intelligence is shrinking quickly due to the rapid
developments being made by open source developers. The memo warns
that while Google and other big tech companies have been focused on their
rivalry with each other, they have neglected the growing community of
programmers and companies that rely on publicly available AI code and
models to create smaller but more efficient projects.
• The Google employee wrote that open source developers are “lapping” big
tech companies in terms of the quality of their AI models, and that the gap is
closing astonishingly quickly.
• Some experts and industry analysts agree with this assessment, with OpenAI
founder Andrej Karpathy writing on Twitter that the surge of small AI
companies competing with tech giants is starting to massively shake up the
industry.
• Leaked Google engineer memo warns that Big Tech could lose AI race to the
little guys (nbcnews.com)
28
Robots are coming
• LLMs can be used to give robots the ability to understand natural language
instructions, which can make them more versatile and easier to use.
• Latest developments in robots that are driven by LLMs:
• Google AI researchers have unveiled RoboCat, a self learning artificial intelligence model that is
built on the Gato general agent. It is a self-improving AI agent that learns to perform a variety
of tasks across different arms, and then self-generates new training data to improve its
technique. Can pick up a new task with as few as 100 demonstrations because it draws from a
large and diverse dataset. https://www.youtube.com/watch?v=RrfD1NxFN-o
• OpenAI has revealed their self learning agent that used video data from humans to
learn various Minecraft tasks.
• https://www.youtube.com/watch?v=HaaZ8ss-HP4 Deepmind Mobile Aloha doing housework.
• https://www.youtube.com/watch?v=gyURDZB7imo Tesla Optimus folding clothes
• https://www.facebook.com/parentips/videos/teach-kids-to-fold-clothes-with-a-diy-folding-
board/801417770748603/ Human folding clothes
• A team of researchers at the University of California, Berkeley has developed a robot that can
learn to perform tasks by watching humans. The robot, called REAPR, uses an LLM to learn the
sequences of actions that humans take to perform tasks. REAPR is still under development, but
it has the potential to make robots more autonomous and adaptable.
• A company called Intuition Robotics has developed a robot called Moxi that can provide
companionship and assistance to elderly people. Moxi uses an LLM to understand natural
language instructions and to generate conversation. Moxi is still under development, but it has
the potential to make a positive impact on the lives of elderly people.
Robots are coming
• https://www.youtube.com/watch?v=6OB7r4gUh74
• DARPA Subterranean Challenge Final Event Wrap-Up, Sep 21, 2021
• The Subterranean (SubT) Challenge was a Defense Advanced Research
Projects Agency (DARPA) program to develop innovative technologies that
can augment underground operations.
• Explored new approaches to rapidly map, navigate, search, and
exploit complex underground environments such as human-made
tunnel systems, urban underground, and natural cave networks.
These difficult settings present challenges for military and civilian first
responders; hazards vary drastically across terrain that can change
over time, making it too risky for personnel to enter.
Conclusion
• IMF job impact 40% https://www.imf.org/en/Blogs/Articles/2024/01/14/ai-will-transform-the-global-
economy-lets-make-sure-it-benefits-humanity
• Massive job loss is not immediate.
• When will it start… already started, and will accelerate quickly.
• Covid crisis will pressure all companies to go digital much faster, e.g. cosmetics manufacturing, schools,
banking services.
• Is it old wine in new bottle; like global warming, globalisation etc, that we have experienced before?
• How will AI make its greatest impact?
• Changes to world order, nature of our economy, how we will live, our hopes and aspiration for the future etc.
• Who will benefit most?
• The various Internet giants
• You! As you have seen the future.