Cover image

When Agents Agree Too Much: Emergent Bias in Multi‑Agent AI Systems

Opening — Why this matters now Multi‑agent AI systems are having a moment. Debate, reflection, consensus — all the cognitive theater we associate with human committees is now being reenacted by clusters of large language models. In finance, that sounds reassuring. Multiple agents, multiple perspectives, fewer blind spots. Or so the story goes. This paper politely ruins that assumption. ...

December 21, 2025 · 4 min · Zelina
Cover image

When Rewards Learn to See: Teaching Humanoids What the Ground Looks Like

Opening — Why this matters now Humanoid robots can now run, jump, and occasionally impress investors. What they still struggle with is something more mundane: noticing the stairs before falling down them. For years, reinforcement learning (RL) has delivered impressive locomotion demos—mostly on flat floors. The uncomfortable truth is that many of these robots are, functionally speaking, blind. They walk well only because the ground behaves politely. Once the terrain becomes uneven, discontinuous, or adversarial, performance collapses. ...

December 21, 2025 · 4 min · Zelina
Cover image

When Tensors Meet Telemedicine: Diagnosing Leukemia at the Edge

Opening — Why this matters now Healthcare AI has a credibility problem. Models boast benchmark-breaking accuracy, yet quietly fall apart when moved from lab notebooks to hospital workflows. Latency, human-in-the-loop bottlenecks, and fragile classifiers all conspire against real-world deployment. Leukemia diagnosis—especially Acute Lymphocytic Leukemia (ALL)—sits right in the crosshairs of this tension: early detection saves lives, but manual microscopy is slow, subjective, and error-prone. ...

December 21, 2025 · 4 min · Zelina
Cover image

Black Boxes, White Coats: AI Epidemiology and the Art of Governing Without Understanding

Opening — Why this matters now We keep insisting that powerful AI systems must be understood before they can be trusted. That demand feels intuitively correct—and practically paralysing. Large language models now operate in medicine, finance, law, and public administration. Yet interpretability tools—SHAP, LIME, mechanistic circuit tracing—remain brittle, expensive, and increasingly disconnected from real-world deployment. The gap between how models actually behave and how we attempt to explain them is widening, not closing. ...

December 20, 2025 · 4 min · Zelina
Cover image

Don’t Tell the Robot What You Know

Opening — Why this matters now Large Language Models are very good at knowing. They are considerably worse at helping. As AI systems move from chat interfaces into robots, copilots, and assistive agents, collaboration becomes unavoidable. And collaboration exposes a deeply human cognitive failure that LLMs inherit wholesale: the curse of knowledge. When one agent knows more than another, it tends to communicate as if that knowledge were shared. ...

December 20, 2025 · 4 min · Zelina
Cover image

Let There Be Light (and Agents): Automating Quantum Experiments

Opening — Why this matters now Quantum optics sits at an awkward intersection: conceptually elegant, mathematically unforgiving, and operationally tedious. Designing even a “classic” experiment often means stitching together domain intuition, optical components, and simulation code—usually in tools that were never designed for conversational exploration. As AI agents move from text completion to task execution, the obvious question emerges: can they design experiments, not just describe them? ...

December 20, 2025 · 3 min · Zelina
Cover image

Memory Over Models: Letting Agents Grow Up Without Retraining

Opening — Why this matters now We are reaching the awkward teenage years of AI agents. LLMs can already do things: book hotels, navigate apps, coordinate workflows. But once deployed, most agents are frozen in time. Improving them usually means retraining or fine-tuning models—slow, expensive, and deeply incompatible with mobile and edge environments. The paper “Beyond Training: Enabling Self-Evolution of Agents with MOBIMEM” takes a blunt stance: continual agent improvement should not depend on continual model training. Instead, evolution should happen where operating systems have always handled adaptation best—memory. ...

December 20, 2025 · 4 min · Zelina
Cover image

Prompt-to-Parts: When Language Learns to Build

Opening — Why this matters now Text-to-image was a party trick. Text-to-3D became a demo. Text-to-something you can actually assemble is where the stakes quietly change. As generative AI spills into engineering, manufacturing, and robotics, the uncomfortable truth is this: most AI-generated objects are visually plausible but physically useless. They look right, but they don’t fit, don’t connect, and certainly don’t come with instructions a human can follow. ...

December 20, 2025 · 4 min · Zelina
Cover image

Stop or Strip? Teaching Disassembly When to Quit

Opening — Why this matters now Circular economy rhetoric is everywhere. Circular economy decision-making is not. Most end-of-life products still follow a depressingly simple rule: disassemble until it hurts, or stop when the operator gets tired. The idea that we might formally decide when to stop disassembling — based on value, cost, safety, and information — remains oddly underdeveloped. This gap is no longer academic. EV batteries, e‑waste, and regulated industrial equipment are forcing operators to choose between speed, safety, and sustainability under real constraints. ...

December 20, 2025 · 4 min · Zelina
Cover image

The Ethics of Not Knowing: When Uncertainty Becomes an Obligation

Opening — Why this matters now Modern systems act faster than their understanding. Algorithms trade in microseconds, clinical protocols scale across populations, and institutions make irreversible decisions under partial information. Yet our ethical vocabulary remains binary: act or abstain, know or don’t know, responsible or not. That binary is failing. The paper behind this article introduces a deceptively simple idea with uncomfortable implications: uncertainty does not reduce moral responsibility — it reallocates it. When confidence falls, duty does not disappear. It migrates. ...

December 20, 2025 · 4 min · Zelina