Cover image

When Safety Stops Being a Turn-Based Game

Opening — Why this matters now LLM safety has quietly become an arms race with terrible reflexes. We discover a jailbreak. We patch it. A new jailbreak appears, usually crafted by another LLM that learned from the last patch. The cycle repeats, with each round producing models that are slightly safer and noticeably more brittle. Utility leaks away, refusal rates climb, and nobody is convinced the system would survive a genuinely adaptive adversary. ...

December 28, 2025 · 4 min · Zelina
Cover image

When Policies Read Each Other: Teaching Agents to Cooperate by Reading the Code

Opening — Why this matters now Multi-agent systems are finally leaving the toy world. Autonomous traders negotiate with other bots. Supply-chain agents coordinate across firms. AI copilots increasingly share environments with other AI copilots. And yet, most multi-agent reinforcement learning (MARL) systems are still stuck with a primitive handicap: agents cannot meaningfully understand what other agents are doing. ...

December 26, 2025 · 4 min · Zelina
Cover image

When One Clip Isn’t Enough: Teaching LLMs to Watch Long Videos Like Adults

Opening — Why this matters now Large language models have learned to see. Unfortunately, they still have the attention span of a distracted intern when the video runs longer than a minute. As multimodal LLMs expand their context windows and promise “end-to-end” video understanding, a hard reality remains: long videos are not just longer inputs—they are fundamentally different reasoning problems. Information is sparse, temporally distant, multimodal, and often only meaningful when grounded precisely in time and space. Compress everything up front, and you lose the evidence. Don’t compress, and you blow the context budget. ...

December 24, 2025 · 4 min · Zelina
Cover image

Policy Gradients Grow Up: Teaching RL to Think in Domains

Opening — Why this matters now Reinforcement learning keeps winning benchmarks, but keeps losing the same argument: it doesn’t generalize. Train it here, deploy it there, and watch confidence evaporate. Meanwhile, classical planning—decidedly uncool but stubbornly correct—has been quietly producing policies that provably work across arbitrarily large problem instances. This paper asks the uncomfortable question the RL community often dodges: can modern policy-gradient methods actually learn general policies, not just big ones? ...

December 23, 2025 · 4 min · Zelina
Cover image

When Benchmarks Rot: Why Static ‘Gold Labels’ Are a Clinical Liability

Opening — Why this matters now Clinical AI has entered an uncomfortable phase of maturity. Models are no longer failing loudly; they are failing quietly. They produce fluent answers, pass public benchmarks, and even outperform physicians on narrowly defined tasks — until you look closely at what those benchmarks are actually measuring. The paper at hand dissects one such case: MedCalc-Bench, the de‑facto evaluation standard for automated medical risk-score computation. The uncomfortable conclusion is simple: when benchmarks are treated as static truth, they slowly drift away from clinical reality — and when those same labels are reused as reinforcement-learning rewards, that drift actively teaches models the wrong thing. ...

December 23, 2025 · 4 min · Zelina
Cover image

Darwin, But Make It Neural: When Networks Learn to Mutate Themselves

Opening — Why this matters now Modern AI has become very good at climbing hills—provided the hill stays put and remains differentiable. But as soon as the terrain shifts, gradients stumble. Controllers break. Policies freeze. Re-training becomes ritualistic rather than intelligent. This paper asks a quietly radical question: what if adaptation itself lived inside the network? Not as a scheduler, not as a meta-optimizer bolted on top, but as part of the neural machinery that gets inherited, mutated, and selected. ...

December 21, 2025 · 3 min · Zelina
Cover image

When Rewards Learn to See: Teaching Humanoids What the Ground Looks Like

Opening — Why this matters now Humanoid robots can now run, jump, and occasionally impress investors. What they still struggle with is something more mundane: noticing the stairs before falling down them. For years, reinforcement learning (RL) has delivered impressive locomotion demos—mostly on flat floors. The uncomfortable truth is that many of these robots are, functionally speaking, blind. They walk well only because the ground behaves politely. Once the terrain becomes uneven, discontinuous, or adversarial, performance collapses. ...

December 21, 2025 · 4 min · Zelina
Cover image

Adversaries, Slices, and the Art of Teaching LLMs to Think

Opening — Why this matters now Large language models can already talk their way through Olympiad math, but they still stumble in embarrassingly human ways: a missed parity condition, a silent algebra slip, or a confident leap over an unproven claim. The industry’s usual fix—reward the final answer and hope the reasoning improves—has reached diminishing returns. Accuracy nudges upward, but reliability remains brittle. ...

December 19, 2025 · 4 min · Zelina
Cover image

Stepwise Think-Critique: Teaching LLMs to Doubt Themselves (Productively)

Opening — Why this matters now Large Language Models have learned how to think out loud. What they still struggle with is knowing when that thinking is wrong — while it is happening. In high‑stakes domains like mathematics, finance, or policy automation, delayed error detection is not a feature; it is a liability. Most modern reasoning pipelines still follow an awkward split: first generate reasoning, then verify it — often with a separate model. Humans do not work this way. We reason and judge simultaneously. This paper asks a simple but uncomfortable question: what if LLMs were trained to do the same? ...

December 18, 2025 · 4 min · Zelina
Cover image

Picking Less to Know More: When RAG Stops Ranking and Starts Thinking

Opening — Why this matters now Retrieval-Augmented Generation has a dirty secret: it keeps retrieving more context while quietly getting no smarter. As context windows balloon to 100K tokens and beyond, RAG systems dutifully shovel in passages—Top‑5, Top‑10, Top‑100—hoping recall will eventually rescue accuracy. It doesn’t. Accuracy plateaus. Costs rise. Attention diffuses. The model gets lost in its own evidence pile. ...

December 17, 2025 · 4 min · Zelina