Building AI infrastructure — from data pipelines to GPU kernels. I write about data engineering, LLMOps, GPU optimization, and the systems behind reinforcement learning.
Every LLM is built on five foundational pillars: Basics, Systems, Scaling Laws, Data, and Alignment. This post maps out what they are and why mastering them is the path to building real AI systems.
Test-time training lets models update their own weights during inference. Learn how TTT layers work, their GPU implications, and why this changes AI infrastructure.
RAG isn't magic - it's Extract, Transform, Load with vectors. I break down how your existing pipeline skills map directly to building production AI systems.