SYS JS.DEV
BUILD F3CRD9
DATE 2026.04.26
UTC 01:30 UTC
LOC NYC → STANFORD
STATUS OPEN TO ML/SYSTEMS ROLES

Currently Studying — ML Systems

Open reading list and learning targets for the next 6–12 months, oriented around training infrastructure, inference serving, evaluation tooling, and the systems craft underneath frontier-model work.

The active reading list. This page updates monthly. The aim is the engineering substrate underneath frontier-model work — training, inference, evals, infrastructure — not research-method-of-the-week pursuit.

Books in progress

Topics I’m going deep on

In rough priority order:

  1. Transformer internals. Attention math from scratch, KV cache, RoPE, MQA / GQA, FlashAttention.
  2. Distributed training. DDP, FSDP, ZeRO, pipeline + tensor parallelism. The actual primitives, not the framework wrapper.
  3. Inference serving. vLLM, TensorRT-LLM, Triton — paged attention, continuous batching, speculative decoding.
  4. RLHF and preference optimization. DPO, IPO, KTO. The simplifications matter more than the math.
  5. Eval design. LLM-as-judge, offline eval pipelines, regression-testing models — the part I have the most direct production analog for already.
  6. Agent frameworks and tool-use patterns. What’s actually load-bearing in production agents vs. what’s library scaffolding.
  7. Kubernetes and GPU orchestration. Closing the ECS-only gap.

What I’m tracking

Public research I read more or less weekly: Anthropic and OpenAI engineering posts, the SemiAnalysis archive, the more rigorous ML blogs (Lilian Weng, Jay Alammar), and the major training-infra papers as they ship.

Side work

A public-artifacts plan is in motion — the goal is to have a small but serious open-source artifact in the ML-systems space shipped by the time I’m in the Bay. Updates land on the writing index when there’s substance to share.