Build GPT from Scratch

From raw text to a working GPT. You'll implement every component of a transformer-based language model from the ground up.

11
chapters
15
problems
~120hours

Prerequisites

  • Python proficiency
  • Linear algebra fundamentals
  • Basic PyTorch
  • Neural network basics

What You'll Build

  • The complete Transformer architecture, built layer by layer
  • Your own GPT model that generates text autoregressively
Phase 1Foundations
Phase 2The Transformer Core
Phase 3Project GPT
Loading...