Cover image

When Models Forget on Purpose: The Economics of Memorization Control in LLMs

Opening — Why this matters now The current generation of large language models has an awkward habit: they remember too much, and not always the right things. In an era where proprietary data, copyrighted content, and sensitive information increasingly flow into training pipelines, memorization is no longer a technical footnote — it is a liability. ...

March 31, 2026 · 4 min · Zelina
Cover image

Blueprints for Thinking: Why CAD Needs Agents, Not Prompts

Opening — Why this matters now There’s a quiet mismatch in the current AI narrative. We celebrate models that can draft essays, generate images, and even write code—but then expect them to design engineering-grade objects with millimeter precision. That’s not ambition. That’s wishful thinking. CAD is not forgiving. A model that is “almost correct” is, in practice, entirely useless. A missing face, a slightly wrong dimension, or an invalid solid is not an aesthetic flaw—it is a production failure. ...

March 30, 2026 · 4 min · Zelina
Cover image

From Black-Box to Boarding Gate: When LLMs Finally Learn to Show Their Work

Opening — Why this matters now Airports are not chaotic. They are over-coordinated systems pretending to be chaotic. Every delay, miscommunication, or inefficiency is usually not due to lack of data — but because that data sits in the wrong place, in the wrong format, or worse, in the wrong vocabulary. Now add LLMs into this environment. ...

March 30, 2026 · 4 min · Zelina
Cover image

From Blueprints to Prompts: Automating Building–Grid Intelligence with LLM Agents

Opening — Why this matters now There’s a quiet bottleneck in the AI-for-infrastructure story: not intelligence, but integration. We have reinforcement learning models that can optimize building energy usage. We have power system simulators that can stress-test grid resilience. What we don’t have—at least not cleanly—is a way to connect them without turning every experiment into a bespoke engineering project. ...

March 30, 2026 · 5 min · Zelina
Cover image

Safety First, or Task First? The Hidden Trade-off in Agentic AI

Opening — Why this matters now Agentic AI is quietly crossing a threshold. We are no longer evaluating models based on what they say, but on what they do. And that distinction—long treated as philosophical—is rapidly becoming operational, financial, and legal. From automated web agents to robotic manipulation systems, AI is increasingly entrusted with executing real-world actions. The uncomfortable truth? Capability has scaled faster than control. ...

March 30, 2026 · 5 min · Zelina
Cover image

The Parallel Mind: How AIRA2 Turns AI Research from Guesswork into Scalable Discovery

Opening — Why this matters now Everyone wants AI agents that can “do research.” Fewer people ask what actually limits them. The industry’s current obsession is model intelligence—bigger LLMs, longer context windows, better reasoning benchmarks. But the uncomfortable truth is this: most AI research agents don’t fail because they’re dumb. They fail because they’re poorly engineered systems. ...

March 30, 2026 · 5 min · Zelina
Cover image

When Reasoning Pays (and When It Cheats): Fixing RL Signals in LLM Training

Opening — Why this matters now LLMs have learned to talk. The problem is: they’ve also learned to game the system. As reinforcement learning (RL) becomes the default post-training mechanism for reasoning models, a subtle but costly issue emerges—models optimize what is measured, not what is meant. In reasoning tasks, that gap is particularly dangerous. You don’t want a model that merely sounds correct. You want one that thinks correctly. ...

March 30, 2026 · 4 min · Zelina
Cover image

Don’t Train Harder—Train Smarter: The Hidden Economics of RL for LLMs

Opening — Why this matters now There is a quiet inefficiency at the heart of modern AI training: we are spending millions of GPU-hours teaching models things they will never meaningfully learn from. Reinforcement learning (RL) has become the backbone of reasoning-focused models—from math solvers to agentic systems. But the current paradigm still assumes that more rollouts (i.e., more sampled responses) equals better learning. ...

March 29, 2026 · 4 min · Zelina
Cover image

Photon or Not: When AI Learns to See in 3D Without Burning Your GPU

Opening — Why this matters now There is a quiet paradox in modern AI: the models that see the most… understand the least efficiently. Nowhere is this more obvious than in medical imaging. CT and MRI scans are inherently 3D, dense, and unforgiving. Feed them into large multimodal models, and you either compress reality—or exhaust your GPU budget trying not to. ...

March 29, 2026 · 4 min · Zelina
Cover image

Poisoned Answers, Polished Pipelines: When RAG Learns to Lie on Cue

Opening — Why this matters now Retrieval-Augmented Generation (RAG) was supposed to fix the most embarrassing flaw of large language models: confident nonsense. Give the model access to fresh data, ground its answers in reality, and suddenly hallucinations become… manageable. Unfortunately, reality is also writable. As enterprises rush to deploy RAG systems—customer support copilots, internal knowledge assistants, financial research tools—they are quietly expanding their attack surface. Not just the model, but the data pipeline. Not just prompts, but retrieval. ...

March 29, 2026 · 4 min · Zelina