Opening — Why this matters now

Reasoning models have learned how to think longer. Unfortunately, they have not learned when to stop.

Test-time scaling has become the industry’s favorite blunt instrument: allocate more tokens, get better answers—on average. But averages are a luxury in deployment. In production systems, every additional token is a cost, and every premature stop is a risk. The uncomfortable truth is that “adaptive reasoning” merely replaces one opaque knob (token limits) with another (confidence thresholds), without offering a principled way to tune either.

This paper proposes a reframing that is both cleaner and more honest: reasoning budgets are not a compute problem; they are a risk management problem.

Background — From overthinking to uncontrolled exits

Modern reasoning LLMs frequently overthink. Given a prompt, they emit long chains of thought even after the correct answer is already reachable. Prior work attempts to mitigate this by upper-threshold early stopping: monitor a confidence or uncertainty signal and halt once it exceeds a fixed cutoff.

The flaw is subtle but fatal. Thresholds are:

  • signal-dependent,
  • model-dependent,
  • dataset-dependent,
  • and largely uninterpretable.

A confidence value of 0.7 may be conservative for one model and reckless for another. As shown empirically in the paper, identical target error rates map to wildly different threshold values depending on the signal used. Adaptive reasoning, in its current form, optimizes intuition, not guarantees.

Analysis — Conformal risk control enters reasoning

The key move in this work is to treat early stopping as a statistical decision under uncertainty. Any stop introduces two possible errors:

  1. False positives — stopping because the model believes it is correct, but it is not.
  2. False negatives — stopping because the model appears stuck, even though further reasoning would succeed.

Rather than tuning thresholds directly, the framework asks the user to specify acceptable risk levels for these errors. Thresholds are then derived, not guessed.

Dual-threshold design

The paper introduces a symmetric early-exit mechanism:

Threshold What it controls Failure mode avoided
Upper threshold False positives Overthinking after convergence
Lower threshold False negatives Token burn on unsolvable problems

The lower threshold is the novel contribution. Instead of waiting forever for confidence to rise, the system monitors lack of progress. A parametric sigmoid function—growing stricter as tokens accumulate—halts reasoning when confidence stagnates.

This is a quiet but important shift: unsolvability is treated as a first-class outcome, not a timeout artifact.

Calibration via distribution-free guarantees

Thresholds are selected using a held-out validation set and conformal-style risk control. Crucially:

  • Empirical risk is adjusted with finite-sample corrections.
  • Thresholds are accepted only if guaranteed (with high probability) to respect the user’s risk tolerance.
  • Among all valid candidates, the most compute-efficient configuration is chosen.

This avoids the classic failure mode of validation overfitting, where a threshold looks safe offline and fails silently in deployment.

Findings — Efficiency without gambling

Across multiple models (Qwen, DeepSeek) and reasoning benchmarks (AIME, GPQA, MathVision), three consistent results emerge:

  1. Risk control works: realized test risk stays below the specified tolerance, unlike naive validation.
  2. Lower thresholds matter: when unsolvable instances are common, upper-threshold-only methods waste enormous compute.
  3. Signal ensembling helps: selecting the most efficient stopping signal per risk level yields better accuracy–token trade-offs than committing to a single heuristic.

A particularly telling result shows that when solvable and unsolvable problems are mixed, dual-threshold systems naturally split labor:

  • solvable instances exit via confidence,
  • unsolvable ones exit via pessimism.

This is not heuristic cleverness. It is what rational allocation under uncertainty looks like.

Implications — What this changes for agentic systems

For agent-based AI systems, this paper closes a conceptual gap.

Agents are expected to:

  • reason selectively,
  • respect budgets,
  • escalate when uncertain,
  • and justify abstention.

Risk-controlled early stopping provides a unifying interface: “Here is how wrong you are allowed to be.” Everything else follows mechanically.

More broadly, this work signals a maturation of reasoning research. We are moving from scaling behavior to governance of cognition—from “think harder” to “think responsibly.”

Conclusion — Stop guessing, start budgeting risk

The uncomfortable lesson of this paper is that adaptive reasoning without risk guarantees is just structured guesswork. Conformal Thinking replaces guesswork with accountability.

By translating compute budgets into statistical commitments, it offers a path toward reasoning systems that are not only powerful, but operable.

That, more than another million tokens, is what real-world AI has been waiting for.

Cognaptus: Automate the Present, Incubate the Future.