ML educator, 'Build a LLM from Scratch' author
Sebastian Raschka
Profile
Sebastian Raschka is the rare ML educator who actually makes you build the thing. His book Build a Large Language Model (From Scratch) walks you through coding a GPT-style transformer in PyTorch — tokenizer, attention heads, pretraining loop, fine-tuning — without pulling in Hugging Face or any other LLM library. The accompanying GitHub repo is one of the clearest teaching artifacts on the internet for anyone who wants to stop treating LLMs as magic boxes.
He’s a PhD statistician turned Staff Research Engineer at Lightning AI, and previously a professor at the University of Wisconsin–Madison. That dual background shows in his work: he explains the math without hiding behind it, and writes code that runs without hand-waving past the hard parts. Before the LLM book he co-authored the bestselling Machine Learning with PyTorch and Scikit-Learn, which for years has been the default “teach yourself ML properly” recommendation.
His newsletter Ahead of AI, read by 150k+ subscribers, is where he breaks down LLM research — LoRA, DPO, positional embeddings, KV caching, reasoning models — with annotated diagrams and minimum-viable code. No hype cycles, no AGI takes. Just “here’s the paper, here’s what’s actually new, here’s what it looks like in 40 lines of PyTorch.”
For a father-son duo learning AI together, Raschka is the bridge. If university gives you the theory and decades of engineering give you the instincts, Raschka’s from-scratch approach is where those meet: code-first, math-honest, no shortcuts.
Books
Build a Large Language Model (From Scratch) Code a GPT-style LLM in PyTorch from tokenizer to instruction fine-tuning, with no LLM libraries — the definitive hands-on book on how transformers actually work. Machine Learning with PyTorch and Scikit-Learn A comprehensive tour of classical ML and deep learning in Python, co-authored with Yuxi Liu and Vahid Mirjalili — still one of the best self-study ML textbooks. Machine Learning Q and AI Thirty essential questions and answers covering the modern ML and AI topics that interviews and real work actually hit — attention, LoRA, evaluation, and more.Key Articles & Papers
LLMs from Scratch (GitHub) Understanding and Coding the Self-Attention Mechanism Practical Tips for Finetuning LLMs Using LoRA Understanding Reasoning LLMs Understanding Large Language Models New LLM Pre-training and Post-training Paradigms Ahead of AI NewsletterSpotify Podcasts