Thanks
- To the host for the great venue!
- To our sponsors
Administrivia
- Fire escapes
- Toilets
- Cleaning up after ourselves
- WiFi
Demystifying AI
How generative AI works and how that informs how to best use it
How does a chatbot generate text?
Where do LLM weights come from?
- Weights start as random numbers
- Weights are updated by a process called
pre-training:
- Uses millions of training examples from
general sources, like the internet and books
- E.g. Leave-one-word-out examples:
The Beatles were a ____ from England
→ band
- For each example, weights are updated
slightly to get closer to the right answer - like tuning a
dial
- This trains the LLM in language
patterns AND general knowledge
- Extra training based on human
feedback makes chatbots respond better to
questions and instructions
- E.g. Chatbot gives 5 answers and human picks the
best
You now understand what GPT stands for!
Generative
|
The model generates text (or other content)
|
Pre-trained
|
Pre-trained on many examples from general sources
|
Transformer
|
The kind of neural network architecture used
|
What does this teach us about using GenAI?
- The generation process is not magic, just
simple maths!
- We know how AI works, but no-one
truly knows why it works
- The billions of calculations across neurons are too
complex
- AI doesn’t “think” with logic -
it’s predicting the best answer based on its training sources
- People say AI sometimes “hallucinates”
wrong answers - but every response is generated the same way
- It generates answers that sound right -
and it’s surprising so many actually are right!
- It doesn’t train during your chats
- It “remembers” by including chat history in the
prompt
- Even “memory” features just include past chats in
the prompt
- Public chatbots may use your chats to train later
models
GenAI Strengths and Limitations
Key Strengths
- Generates text that looks similar to what a human would write
- Reads lots of text really fast
- Surprisingly good responses to a wide variety of tasks
Key Limitations
- Everything is a hallucination - but it’s correct surprisingly
often
- It might not always “pay attention” to everything in the prompt
- It might not consistently give the same answer
- Different models (or model versions) can give very different
responses to the same prompt
Homework
- TODO
- To learn more about how LLMs work, watch these videos from
3Blue1Brown: