Cover image

Black Boxes, White Coats: AI Epidemiology and the Art of Governing Without Understanding

Opening — Why this matters now We keep insisting that powerful AI systems must be understood before they can be trusted. That demand feels intuitively correct—and practically paralysing. Large language models now operate in medicine, finance, law, and public administration. Yet interpretability tools—SHAP, LIME, mechanistic circuit tracing—remain brittle, expensive, and increasingly disconnected from real-world deployment. The gap between how models actually behave and how we attempt to explain them is widening, not closing. ...

December 20, 2025 · 4 min · Zelina
Cover image

Don’t Tell the Robot What You Know

Opening — Why this matters now Large Language Models are very good at knowing. They are considerably worse at helping. As AI systems move from chat interfaces into robots, copilots, and assistive agents, collaboration becomes unavoidable. And collaboration exposes a deeply human cognitive failure that LLMs inherit wholesale: the curse of knowledge. When one agent knows more than another, it tends to communicate as if that knowledge were shared. ...

December 20, 2025 · 4 min · Zelina
Cover image

Let There Be Light (and Agents): Automating Quantum Experiments

Opening — Why this matters now Quantum optics sits at an awkward intersection: conceptually elegant, mathematically unforgiving, and operationally tedious. Designing even a “classic” experiment often means stitching together domain intuition, optical components, and simulation code—usually in tools that were never designed for conversational exploration. As AI agents move from text completion to task execution, the obvious question emerges: can they design experiments, not just describe them? ...

December 20, 2025 · 3 min · Zelina
Cover image

Memory Over Models: Letting Agents Grow Up Without Retraining

Opening — Why this matters now We are reaching the awkward teenage years of AI agents. LLMs can already do things: book hotels, navigate apps, coordinate workflows. But once deployed, most agents are frozen in time. Improving them usually means retraining or fine-tuning models—slow, expensive, and deeply incompatible with mobile and edge environments. The paper “Beyond Training: Enabling Self-Evolution of Agents with MOBIMEM” takes a blunt stance: continual agent improvement should not depend on continual model training. Instead, evolution should happen where operating systems have always handled adaptation best—memory. ...

December 20, 2025 · 4 min · Zelina
Cover image

Prompt-to-Parts: When Language Learns to Build

Opening — Why this matters now Text-to-image was a party trick. Text-to-3D became a demo. Text-to-something you can actually assemble is where the stakes quietly change. As generative AI spills into engineering, manufacturing, and robotics, the uncomfortable truth is this: most AI-generated objects are visually plausible but physically useless. They look right, but they don’t fit, don’t connect, and certainly don’t come with instructions a human can follow. ...

December 20, 2025 · 4 min · Zelina
Cover image

Stop or Strip? Teaching Disassembly When to Quit

Opening — Why this matters now Circular economy rhetoric is everywhere. Circular economy decision-making is not. Most end-of-life products still follow a depressingly simple rule: disassemble until it hurts, or stop when the operator gets tired. The idea that we might formally decide when to stop disassembling — based on value, cost, safety, and information — remains oddly underdeveloped. This gap is no longer academic. EV batteries, e‑waste, and regulated industrial equipment are forcing operators to choose between speed, safety, and sustainability under real constraints. ...

December 20, 2025 · 4 min · Zelina
Cover image

The Ethics of Not Knowing: When Uncertainty Becomes an Obligation

Opening — Why this matters now Modern systems act faster than their understanding. Algorithms trade in microseconds, clinical protocols scale across populations, and institutions make irreversible decisions under partial information. Yet our ethical vocabulary remains binary: act or abstain, know or don’t know, responsible or not. That binary is failing. The paper behind this article introduces a deceptively simple idea with uncomfortable implications: uncertainty does not reduce moral responsibility — it reallocates it. When confidence falls, duty does not disappear. It migrates. ...

December 20, 2025 · 4 min · Zelina
Cover image

Adversaries, Slices, and the Art of Teaching LLMs to Think

Opening — Why this matters now Large language models can already talk their way through Olympiad math, but they still stumble in embarrassingly human ways: a missed parity condition, a silent algebra slip, or a confident leap over an unproven claim. The industry’s usual fix—reward the final answer and hope the reasoning improves—has reached diminishing returns. Accuracy nudges upward, but reliability remains brittle. ...

December 19, 2025 · 4 min · Zelina
Cover image

AGI by Committee: Why the First General Intelligence Won’t Arrive Alone

Opening — Why this matters now For years, AGI safety discussions have revolved around a single, looming figure: the model. One system. One alignment problem. One decisive moment. That mental model is tidy — and increasingly wrong. The paper “Distributional AGI Safety” argues that AGI is far more likely to emerge not as a monolith, but as a collective outcome: a dense web of specialized, sub‑AGI agents coordinating, trading capabilities, and assembling intelligence the way markets assemble value. AGI, in this framing, is not a product launch. It is a phase transition. ...

December 19, 2025 · 4 min · Zelina
Cover image

CitySeeker: Lost in Translation, Found in the City

Opening — Why this matters now Urban navigation looks deceptively solved. We have GPS, street-view imagery, and multimodal models that can describe a scene better than most humans. And yet, when vision-language models (VLMs) are asked to actually navigate a city — not just caption it — performance collapses in subtle, embarrassing ways. The gap is no longer about perception quality. It is about cognition: remembering where you have been, knowing when you are wrong, and understanding implicit human intent. This is the exact gap CitySeeker is designed to expose. ...

December 19, 2025 · 3 min · Zelina