Cover image

Speculation, But With Standards: Training Draft Models That Actually Get Accepted

Opening — Why this matters now Speculative decoding has quietly become one of the most important efficiency tricks in large language model inference. It promises something deceptively simple: generate multiple tokens ahead of time with a cheap draft model, then let the expensive model verify them in parallel. Fewer forward passes, lower latency, higher throughput. ...

February 8, 2026 · 4 min · Zelina
Cover image

Speculate Smarter, Not Harder: Hierarchical Decoding Without Regret

Opening — Why this matters now LLM inference has quietly become the dominant cost center of modern AI systems. Training grabs headlines; inference drains budgets. As models scale into the tens of billions of parameters, every additional forward pass hurts — financially and operationally. Speculative decoding promised relief by letting small models run ahead and big models merely verify. But verification, ironically, became the bottleneck. ...

January 12, 2026 · 3 min · Zelina