Thinking in Branches: Why LLM Reasoning Needs an Algorithmic Theory
Opening — Why this matters now Enterprises are discovering a strange contradiction: Large Language Models can now solve competition-level math, yet still fail a moderately complex workflow audit if you ask for the answer once. But let them think longer—sampling, refining, verifying—and suddenly the same model performs far beyond its pass@1 accuracy. Welcome to the age of inference-time scaling, where raw model size is no longer the sole determinant of intelligence. Instead, we orchestrate multiple calls, combine imperfect ideas, and build pipelines that behave less like autocomplete engines and more like genuine problem solvers. ...