Cover image

Survival by Swiss Cheese: Why AI Doom Is a Layered Failure, Not a Single Bet

Opening — Why this matters now Ever since ChatGPT escaped the lab and wandered into daily life, arguments about AI existential risk have followed a predictable script. One side says doom is imminent. The other says it’s speculative hand-wringing. Both sides talk past each other. The paper behind this article does something refreshingly different. Instead of obsessing over how AI might kill us, it asks a sharper question: how exactly do we expect to survive? Not rhetorically — structurally. ...

January 17, 2026 · 5 min · Zelina
Cover image

When Robots Guess, People Bleed: Teaching AI to Say ‘This Is Ambiguous’

Opening — Why this matters now Embodied AI has become very good at doing things. What it remains surprisingly bad at is asking a far more basic question: “Should I be doing anything at all?” In safety‑critical environments—surgical robotics, industrial automation, AR‑assisted operations—this blind spot is not academic. A robot that confidently executes an ambiguous instruction is not intelligent; it is dangerous. The paper behind Ambi3D and AmbiVer confronts this neglected layer head‑on: before grounding, planning, or acting, an agent must determine whether an instruction is objectively unambiguous in the given 3D scene. ...

January 12, 2026 · 4 min · Zelina
Cover image

When the Tutor Is a Model: Learning Gains, Guardrails, and the Quiet Rise of AI Co‑Tutors

Opening — Why this matters now One‑to‑one tutoring is education’s gold standard—and its most stubborn bottleneck. Everyone agrees it works. Almost no one can afford it at scale. Into this gap steps generative AI, loudly promising democratized personalization and quietly raising fears about hallucinations, dependency, and cognitive atrophy. Most debates about AI tutors stall at ideology. This paper does something rarer: it runs an in‑classroom randomized controlled trial and reports what actually happened. No synthetic benchmarks. No speculative productivity math. Just UK teenagers, real maths problems, and an AI model forced to earn its keep under human supervision. fileciteturn0file0 ...

December 31, 2025 · 4 min · Zelina
Cover image

When Models Look Back: Memory, Leakage, and the Quiet Failure Modes of LLM Training

Opening — Why this matters now Large language models are getting better at many things—reasoning, coding, multi‑modal perception. But one capability remains quietly uncomfortable: remembering things they were never meant to remember. The paper underlying this article dissects memorization not as a moral failure or an anecdotal embarrassment, but as a structural property of modern LLM training. The uncomfortable conclusion is simple: memorization is not an edge case. It is a predictable outcome of how we scale data, objectives, and optimization. ...

December 30, 2025 · 3 min · Zelina
Cover image

When Safety Stops Being a Turn-Based Game

Opening — Why this matters now LLM safety has quietly become an arms race with terrible reflexes. We discover a jailbreak. We patch it. A new jailbreak appears, usually crafted by another LLM that learned from the last patch. The cycle repeats, with each round producing models that are slightly safer and noticeably more brittle. Utility leaks away, refusal rates climb, and nobody is convinced the system would survive a genuinely adaptive adversary. ...

December 28, 2025 · 4 min · Zelina
Cover image

Reading the Room? Apparently Not: When LLMs Miss Intent

Opening — Why this matters now Large Language Models are increasingly deployed in places where misunderstanding intent is not a harmless inconvenience, but a real risk. Mental‑health support, crisis hotlines, education, customer service, even compliance tooling—these systems are now expected to “understand” users well enough to respond safely. The uncomfortable reality: they don’t. The paper behind this article demonstrates something the AI safety community has been reluctant to confront head‑on: modern LLMs are remarkably good at sounding empathetic while being structurally incapable of grasping what users are actually trying to do. Worse, recent “reasoning‑enabled” models often amplify this failure instead of correcting it. fileciteturn0file0 ...

December 25, 2025 · 4 min · Zelina
Cover image

Don’t Tell the Robot What You Know

Opening — Why this matters now Large Language Models are very good at knowing. They are considerably worse at helping. As AI systems move from chat interfaces into robots, copilots, and assistive agents, collaboration becomes unavoidable. And collaboration exposes a deeply human cognitive failure that LLMs inherit wholesale: the curse of knowledge. When one agent knows more than another, it tends to communicate as if that knowledge were shared. ...

December 20, 2025 · 4 min · Zelina
Cover image

Safety Without Exploration: Teaching Robots Where Not to Die

Opening — Why this matters now Modern autonomy has a credibility problem. We train systems in silico, deploy them in the real world, and hope the edge cases are forgiving. They usually aren’t. For robots, vehicles, and embodied AI, one safety violation can be catastrophic — and yet most learning‑based methods still treat safety as an expectation, a probability, or worse, a regularization term. ...

December 12, 2025 · 4 min · Zelina
Cover image

Learning by X-ray: When Surgical Robots Teach Themselves to See in Shadows

Opening — Why this matters now Surgical robotics has long promised precision beyond human hands. Yet, the real constraint has never been mechanics — it’s perception. In high-stakes fields like spinal surgery, machines can move with submillimeter accuracy, but they can’t yet see through bone. That’s what makes the Johns Hopkins team’s new study, Investigating Robot Control Policy Learning for Autonomous X-ray-guided Spine Procedures, quietly radical. It explores whether imitation learning — the same family of algorithms used in self-driving cars and dexterous robotic arms — can enable a robot to navigate the human spine using only X-ray vision. ...

November 9, 2025 · 4 min · Zelina
Cover image

Agents, Automata, and the Memory of Thought

If you strip away the rhetoric about “thinking” machines and “cognitive” agents, most of today’s agentic AIs still boil down to something familiar from the 1950s: automata. That’s the thesis of Are Agents Just Automata? by Koohestani et al. (2025), a paper that reinterprets modern agentic AI through the lens of the Chomsky hierarchy—the foundational classification of computational systems by their memory architectures. It’s an argument that connects LLM-based agents not to psychology, but to formal language theory. And it’s surprisingly clarifying. ...

November 1, 2025 · 4 min · Zelina