Lecture 25: Course Revision
| Module | Topic |
|---|---|
| 0 | Orientation |
| 1 | How AI systems are designed |
| 2 | Language and perception |
| 3 | Retrieval, generation, pipelines |
| 4 | Data ethics and bias |
| 5 | Policy, governance, social impact |
| 6 | Applications, limits, and the future |
Next-token prediction explains hallucinations and creativity (L10) and sycophancy (L24). RLHF comes back in Lectures 4 and 24.
The proxy problem is the most recurring idea in this course: Goodhart’s Law (L5), measurement bias (L15), the Optum case (L16), engagement metrics (L21).
| Paradigm | How it learns | Example |
|---|---|---|
| Supervised | Labelled pairs | Spam filter |
| Unsupervised | Finds patterns | Customer segments |
| Reinforcement | Trial and error | AlphaGo, RLHF |
Reward hacking = Goodhart’s Law (L5) applied to training. RLHF returns in Lecture 24 as a key approach to alignment.
Goodhart’s Law returns in RLHF (L4), the Optum case (L16), and engagement metrics (L21). Subgroup analysis is formalised as disaggregated evaluation in Lecture 14.
Embeddings come back in Lecture 7 (images as vectors), Lecture 11 (RAG/cosine similarity), and Lecture 23 (deepfake detection).
Same embedding idea as Lecture 6, applied to images and audio. Same vision tech powers medical imaging and deepfakes (L23). Regulating dual-use models: Lecture 17.
CoT reduces hallucinations (L10). Prompt injection came back in the chatbot failures (L13) and in why we need regulation (L17).
Root cause: next-token prediction (L1-2). Fix: RAG (L11). Sycophancy connects to the alignment problem (L24).
Uses embeddings (L6) to fix hallucinations (L10). But if the documents are biased, the retrieval will be too (L16).
Concept drift relates to bias (L15): the world changes, the model doesn’t. The chatbot failures set legal precedents that motivated regulation (L17).
Disaggregated evaluation is how the Optum bias was found (L16). The EU AI Act (L17) now requires this documentation for high-risk systems.
Measurement bias = the proxy problem (L3). Textbook example: Optum (L16). Feedback loops relate to concept drift (L13). The impossibility theorem connects to alignment (L24).
Ties together the proxy problem (L3), measurement bias (L15), and Goodhart’s Law (L5). The bias was invisible until disaggregated evaluation (L14).
| Tier | Examples |
|---|---|
| Prohibited | Social scoring, real-time biometric ID |
| High-risk | Employment, credit, law enforcement |
| Limited | Chatbots (must disclose they are AI) |
| Minimal | Spam filters, game AI |
“High-risk” = the systems from Lectures 15-16. The Act requires documentation from Lecture 14. Brussels Effect connects to GDPR (L19).
Inference = the proxy problem (L3) applied to personal data. GDPR is the Brussels Effect (L17) in practice. Consent ties to Lecture 14.
Lecture 20: Labour
Lecture 21: Wellbeing
Attention economy = Goodhart’s Law (L5) in action: engagement is the proxy, wellbeing is what we care about. Skill compression relates to bias-variance (L4).
Powered by vision models (L7), detected via embeddings (L6), regulated by the EU AI Act (L17).
Ties together RLHF (L4), Goodhart’s Law (L5), bias (L15-16), and regulation (L17). This is the question the whole course builds toward.
1. Next-token prediction is the foundation
LLMs predict the next word, not the truth. This explains hallucinations, creativity, and sycophancy (Lectures 1, 2, 6, 10)
2. The proxy problem is everywhere
Labels, metrics, objectives, and engagement are all proxies. Optimising a proxy corrupts it (Lectures 3, 5, 15, 16, 21)
3. Embeddings are the universal language
Text, images, and audio all become vectors. RAG, semantic search, CLIP, and multimodal AI all rely on this (Lectures 6, 7, 11)
4. Bias enters at every stage
Data, labels, training, evaluation, deployment. Feedback loops make it self-reinforcing. The impossibility theorem means fairness requires a values choice (Lectures 3, 14, 15, 16)
5. Documentation is accountability
Datasheets, model cards, system cards. Disaggregated evaluation reveals hidden gaps (Lectures 14, 15, 16)
6. Regulation is catching up
EU AI Act, GDPR, the Brussels Effect. Market failures justify intervention (Lectures 17, 19)
If you understand these six threads, you understand the course!
How AI works: next-token prediction, transformers, embeddings, three learning paradigms, metrics and their limits
How AI handles information: tokens, prompting (PTCF, CoT), hallucinations, RAG, pipelines and drift
Ethics and bias: bias at every stage, the impossibility theorem, the Optum case, documentation as accountability
Policy and society: EU AI Act, GDPR, the privacy paradox, labour markets, the attention economy, environmental costs
Safety and the future: concrete safety problems, the alignment problem, multiple plausible futures
You now have the vocabulary to read AI claims critically and ask the right questions about data, bias, and harms. Good luck on the quiz! 🍀