Opening — Why this matters now
The AI industry has quietly entered its barbell phase. On one end, closed-source giants wield compute-rich models that brute-force reasoning through sheer output length. On the other, open-source models aspire to the same depth but collide with the quadratic wall of long-context Transformers.
Into this tension steps a familiar trend: multi-agent reasoning systems. Instead of one monolithic brain grinding through 100,000 tokens, multiple agents collaborate—solve, check, correct, repeat. Elegant in theory, brittle in practice. Outside elite proprietary stacks, the Verifier and Corrector tend to behave more like well-meaning interns than rigorous mathematicians.
MarsRL proposes to fix this imbalance. And surprisingly, it succeeds.
Background — Context and prior art
Prior work on improving reasoning has largely revolved around two levers:
- RLVR (Reinforcement Learning with Verifiable Rewards) — reward the model when the answer is objectively checkable.
- Test-time scaling — let models think longer, generate more candidates, and stitch together coherent solutions.
This is how systems like o1 and DeepSeek R1 gained their reputation: not just cleverness, but persistent, evaluable iteration.
The Verifier–Corrector (V–C) framework, popularized by Huang & Yang’s IMO pipeline, embodies this idea. Instead of extending a single model’s output to 64k+ tokens, multiple agents take turns. A Solver proposes, a Verifier critiques, and a Corrector repairs.
But in open-source ecosystems, this structure collapses. The Verifier often hallucinates bugs; the Corrector over-edits; and the Solver receives noisy feedback. Classic multi-agent credit assignment failure.
Analysis — What MarsRL actually brings
MarsRL addresses two endemic issues:
1. Reward noise across agents
A multi-agent rollout typically produces a long trajectory:
Solver → Verifier → Corrector → Verifier → Corrector → …
If the final answer is correct, RL tends to shower the entire chain with positive reward—even if the Verifier made an obviously wrong judgment somewhere in the middle.
MarsRL fixes this by agent-specific verifiable rewards:
- Solver is judged only on its proposed solution.
- Corrector is judged only on its refinements.
- Verifier is rewarded/punished purely on whether its detection decision matches ground truth.
Granularity replaces ambiguity.
2. Training efficiency: the 300k-token elephant
A single trajectory can stretch to 320,000 tokens when multiple 64k-token agents fire sequentially.
MarsRL applies pipeline parallelism—a trick borrowed from training huge models. Once any agent produces its chunk, that output immediately enters the training queue. There is no need to wait for the full multi-step chain.
Combine this with segmented rollouts (16k blocks), and suddenly the latency curve bends sharply.
3. Grouped agentic sampling
GRPO’s group-relative advantage requires comparable samples. MarsRL maintains this logic across agents by grouping multiple responses per stage and ensuring the Corrector sees only verifier-flagged outputs.
The kicker: an adaptive sampling strategy—Verifiers train on incorrect solutions, Correctors train on correctly detected mistakes—yields superior convergence.
Findings — Results with visualization
MarsRL is applied to Qwen3‑30B‑A3B‑Thinking‑2507. The improvements are not subtle.
Table 1. End-to-end results
| Model | AIME2025 (Solver) | AIME2025 (Reasoning System) | BeyondAIME (Solver) | BeyondAIME (Reasoning System) |
|---|---|---|---|---|
| Qwen3‑30B‑A3B | 73.5 | 69.7 | 50.7 | 47.6 |
| Qwen3‑30B‑A3B‑Thinking‑2507 | 86.5 | 85.6 | 64.9 | 63.3 |
| MarsRL‑Trained (this work) | 91.1 | 93.3 | 70.2 | 73.8 |
A 93.3% score on AIME2025 is a clean leap—surpassing even the 235B parameter variant.
Table 2. Why Solvers improve even without being trained
| Training Variant | Solver Accuracy | Reasoning System Accuracy |
|---|---|---|
| MarsRL‑S (train Solver only) | 89.5 | 90.8 |
| MarsRL‑VC (train only Verifier & Corrector) | 90.4 | 91.7 |
A curious, almost counterintuitive result: training Verifier/Corrector improves the Solver more than training the Solver itself.
Why?
Because the model initially cannot verify or correct with depth. Once those two roles acquire long-form reasoning capability, their improvements generalize upstream to the Solver—likely due to shared backbone weights.
Table 3. Cross-Solver generalization
| Replaced Solver | Solver Alone | With MarsRL V–C Agents |
|---|---|---|
| Qwen3-A3B | 73.5 | 91.7 |
| Qwen3-A22B | 92.3 | 93.3 |
| DeepSeek V3.1-Think | 86.2 | 91.2 |
The trained Verifier+Corrector act as powerful “reasoning amplifiers” for entirely different Solvers.
Implications — What this means for business and AI operators
MarsRL is not just a research novelty. Its architecture hints at deeper shifts in enterprise AI:
1. Specialized agent roles outperform monolithic giants
If a 30B model with properly trained Verifier/Corrector can outperform a 235B model, the future tilts toward compositional intelligence rather than scale-for-scale’s-sake.
2. Traceability and auditability improve
Agent-specific rewards produce clear provenance:
- Where did reasoning fail?
- Was a correction justified?
- Did the Verifier invent a bug or catch a subtle one?
This structure aligns neatly with emerging AI assurance frameworks.
3. Faster reinforcement learning cycles for reasoning-intensive products
Pipeline-parallel RL unlocks long-horizon training without burning GPU months. Enterprises working on compliance automation, financial modeling, or scientific assistants can train deeper reasoning stacks using manageable compute.
4. Swappable reasoning components become real assets
MarsRL’s Verifier & Corrector are portable across Solvers.
Imagine:
- A finreg compliance agent plugging into any corporate LLM.
- A medical QA pipeline where Verifier modules are cross-model safety gates.
Agent roles become modular IP.
Conclusion — The shape of agentic RL to come
MarsRL is a reminder that reasoning is not merely a property of bigger models—it’s a property of better organized systems. By cleaning up reward attribution and compressing the training pipeline, MarsRL gives open-source models a ladder to climb toward deep, multi-step reasoning.
The message is subtle but important:
The next frontier in agentic AI is not longer outputs, but cleaner roles, clearer rewards, and faster feedback loops.
Cognaptus: Automate the Present, Incubate the Future.