Reiner Pope

Reiner Pope

CEO of MatX and former Google TPU architecture contributor. In [src-042], Pope explains LLM serving and training from first principles: roofline analysis, batch-size economics, KV-cache memory pressure, mixture-of-experts layout, and memory-tier trade-offs.

Key facts

  • Type: AI hardware / ML infrastructure engineer
  • Current role: CEO of MatX [src-042]
  • Previous work: TPU architecture and related systems work at Google [src-042]
  • Contribution in this source: Uses back-of-the-envelope hardware models to explain why model pricing, context limits, fast modes, and cache behavior look the way they do [src-042]

Related entities

Related concepts

Source references

  • [src-042] Dwarkesh Patel — “How GPT, Claude, and Gemini are actually trained and served – Reiner Pope” (2026-04-29)