If you want to actually understand what’s happening inside ChatGPT (or any modern neural network) — not just use it, but get a real mental model of what the math is doing — Grant Sanderson’s Deep Learning series on 3Blue1Brown is the single best place I’ve found. Seven chapters, each one beautifully animated, walking from “what is a neural network, structurally” all the way to “how might an LLM actually store the fact that Michael Jordan plays basketball.”
I’ve found the early chapters (1–4) timeless — they teach the foundations any practitioner needs. Chapters 5–7 are newer (2024) and address what makes today’s LLMs work: transformers, attention, and the under-discussed role of MLPs in storing factual knowledge.
Watch them in order. Take notes. It’s worth it.
— Mark






