Lecture 02: A Brief History of AI and the Recent Shift
Source: Wikipedia - Talos
Source: Wikipedia - Analytical Engine
Source: NASA
Source: Wikipedia - Turing Test
“Every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it.”
Source: IEEE Spectrum
Symbolic AI, 1956–1980s
Expert systems, 1970s–1980s
Shift to narrow domains with expert knowledge
Encode human expertise as if-then rules
Examples: MYCIN (bacterial infections, video here), DENDRAL (chemical analysis, video here)
Edward Feigenbaum: “The problem-solving power… is primarily a consequence of the specialist’s knowledge employed”
Both approaches relied on hand-crafted knowledge, not learning from data
Neither could learn or improve automatically
The first winter (1974–1980)
The second winter (1987–1993)
The pattern
Anything familiar about this pattern? 😅
Maybe this time is different? 🤷🏻♂️
Source: AI Mind
Source: Medium
“Now go out and gather some data, and see what it can do.”
Source: Google Research
Source: Medium
Source: Hivenet
A transformer has three main parts:
Source: Medium
Source: Google Cloud
Source: Jay Alammar
Source: Thomas Wiecki
Source: Mohamed Traore
Source: Transformer Explainer
| Model | Year | Parameters | Notable Achievement |
|---|---|---|---|
| GPT-1 | 2018 | 117M | Showed pre-training works |
| BERT | 2018 | 340M | Revolutionised NLP benchmarks |
| GPT-2 | 2019 | 1.5B | “Too dangerous to release” |
| GPT-3 | 2020 | 175B | Few-shot learning emergence |
| GPT-4 | 2023 | ~1.7T? | Multimodal, near-human reasoning |
| GPT-5 | 2025 | 635B? | More efficient, even better reasoning |
Source: Voronoi
Source: Tarun Sharma
What changed:
What stayed the same: