Opening — Why this matters now

Inference-time reasoning has quietly become the dominant performance lever for frontier language models. When benchmarks get hard, we don’t retrain—we let models think longer. More tokens, more scratchpad, more compute. The industry narrative is simple: reasoning scales, so accuracy scales.

This paper asks an uncomfortable question: how long must a model think, at minimum, as problems grow? And the answer, grounded in theory rather than vibes, is not encouraging.

Background — From clever prompts to communication limits

Chain-of-thought (CoT) reasoning works because it externalizes intermediate computation into tokens. Prior work argued that this helps models solve problems that exceed their single-step expressive capacity. But this paper builds on a sharper abstraction: the Bounded Attention Prefix Oracle (BAPO).

BAPO treats a transformer as a communication-constrained system. At every decoding step, only a limited amount of information can cross from earlier tokens to later ones—through prefix summaries and attention. CoT helps by spreading computation over many steps, but the bandwidth per step remains bounded.

This reframes reasoning as a token complexity problem: how many reasoning tokens are fundamentally required, regardless of prompting tricks?

Analysis — What the paper actually proves

The authors analyze three canonical problems known to be communication-heavy:

  • Binary Majority: deciding whether ones outnumber zeros.
  • Triplet Matching (MATCH3): detecting modular triplets.
  • Graph Reachability: determining if a path exists between two nodes.

Using the BAPO framework, they prove a clean result: all three require (\Omega(n)) reasoning tokens when input size is (n), assuming constant per-step bandwidth.

In plain terms: for these problems, no clever chain-of-thought compression can beat linear growth in reasoning length. If the input doubles, the minimal required thinking doubles too.

Why this matters computationally

Attention cost scales roughly quadratically with context length. So linear growth in reasoning tokens implies superlinear—often quadratic—inference cost. Token-efficient reasoning isn’t just nice to have; it’s the difference between viable systems and compute sinkholes.

Findings — Theory meets frontier models

The paper doesn’t stop at proofs. It tests frontier reasoning models on parametrically scaled instances.

Model Task Token Scaling Accuracy Trend
GPT-5.2 Majority / MATCH3 / Reachability ~Linear Near-perfect with sufficient budget
Gemini 2.5 Pro Same tasks Worse than linear (large constants) Degrades at larger (n)

When reasoning budgets are capped below the linear threshold, performance collapses to near-random guessing. No amount of prompt engineering rescues it.

The empirical takeaway is blunt: models already behave as the theory predicts. Reasoning tokens are not optional overhead—they are the algorithm.

Implications — Why this breaks the scaling fantasy

Three uncomfortable implications fall out:

  1. Inference-time scaling has hard limits Linear reasoning requirements translate into rapidly exploding costs. “Just think longer” does not scale indefinitely.

  2. Reasoning compression has a ceiling If a task’s minimal reasoning length is (\Omega(n)), any fixed-budget or constant-ratio compression will eventually fail.

  3. Architecture matters more than prompts The paper explicitly points toward alternatives: looped transformers, external tools, retrieval, or architectures that change the communication bottleneck itself.

In short, this is not a prompt problem. It’s a systems problem.

Conclusion — When thinking becomes the bottleneck

This paper does something rare in modern LLM research: it draws a firm line between possible and impossible. Some problems simply require long reasoning chains, and no amount of cleverness will compress them away.

If inference-time compute is becoming the dominant cost center—and it is—then understanding token complexity is no longer academic. It’s operational.

The era of unpriced thinking is over.

Cognaptus: Automate the Present, Incubate the Future.