Cover image

Graph Expectations: Why Context Compression Needs Structure, Not Just Similarity

Opening — Why this matters now The AI industry has developed a charmingly expensive habit: when models struggle with long documents, we buy them larger windows and pretend the problem has been solved. It has not. Long-context LLMs are useful, but longer context is not the same as better context. A model can accept a very large input and still miss the crucial paragraph buried in the middle, over-attend to duplicated evidence, or lose the argumentative spine of a document. The result is familiar to anyone building AI tools for legal review, finance research, policy analysis, procurement, consulting, compliance, or enterprise knowledge work: the model has “read” everything, yet somehow understands the wrong thing. Very modern. Very expensive. ...

May 1, 2026 · 12 min · Zelina
Cover image

When AI Can Solve But Can't Search: The MathNet Equation

Opening — Why this matters now The AI industry enjoys announcing that models now perform at medal level on Olympiad mathematics. Impressive headlines. Elegant demos. Much applause. Then MATHNET arrives with the social grace of an auditor. This new benchmark shows that while leading models can often solve difficult mathematics, they are far worse at finding related problems, recognizing structural equivalence, or reliably using retrieved examples to improve reasoning. In practical terms: your AI intern may ace the exam, then fail to locate the right binder. ...

April 23, 2026 · 4 min · Zelina
Cover image

Epistemic Infrastructure: Why Your AI Knows Less Than It Thinks

Opening — Why this matters now The enterprise AI stack has a favorite illusion: if you retrieve the right documents, you will get the right answer. It’s a comforting belief—engineer better embeddings, expand context windows, sprinkle some graph retrieval, and the system will eventually behave. Except it doesn’t. The paper “Retrieval Is Not Enough: Why Organizational AI Needs Epistemic Infrastructure” fileciteturn0file0 argues something quietly inconvenient: the bottleneck is no longer retrieval fidelity—it’s epistemic fidelity. ...

April 14, 2026 · 5 min · Zelina
Cover image

From YouTube to Execution: How GUIDE Teaches AI Agents to Actually Use Software

Opening — Why this matters now Everyone is excited about AI agents that can “use a computer.” Few are impressed once they actually try. The failure mode is strangely consistent: the agent understands what you want, but fails somewhere embarrassingly practical—clicking the wrong menu, missing a button, or wandering into a dead-end workflow. This is not a capability problem. It’s a familiarity problem. ...

March 30, 2026 · 5 min · Zelina

Build a Small RAG Knowledge Tool

How to build a lightweight retrieval-augmented knowledge tool with grounded answers, source citations, narrow scope, and a realistic MVP.

March 16, 2026 · 5 min · Michelle

Build an Internal Knowledge Assistant

How to design an internal AI assistant that helps staff find policies, procedures, and operating knowledge without creating a guessing machine.

March 16, 2026 · 6 min · Michelle

RAG Explained for Business

A business-friendly explanation of retrieval-augmented generation and why it matters when your AI must work from company knowledge.

March 16, 2026 · 7 min · Michelle
Cover image

Search-R2: When Retrieval Learns to Admit It Was Wrong

Opening — Why this matters now Search-integrated LLMs were supposed to be the antidote to hallucination. Give the model tools, give it the web, let it reason step by step—problem solved. Except it wasn’t. What we actually built were agents that search confidently, reason eloquently, and fail quietly. One bad query early on, one misleading paragraph retrieved at the wrong moment, and the whole reasoning chain collapses—yet reinforcement learning still rewards it if the final answer happens to be right. ...

February 4, 2026 · 4 min · Zelina
Cover image

When Retrieval Learns to Breathe: Teaching LLMs to Go Wide *and* Deep

Opening — Why this matters now Large language models are no longer starved for text. They are starved for structure. As RAG systems mature, the bottleneck has shifted from whether we can retrieve information to how we decide where to look first, how far to go, and when to stop. Most retrieval stacks still force an early commitment: either search broadly and stay shallow, or traverse deeply and hope you picked the right starting point. ...

January 21, 2026 · 4 min · Zelina
Cover image

Secrets, Context, and the RAG Illusion

Opening — Why this matters now Personalized AI assistants are rapidly becoming ambient infrastructure. They draft emails, recall old conversations, summarize private chats, and quietly stitch together our digital lives. The selling point is convenience. The hidden cost is context collapse. The paper behind this article introduces PrivacyBench, a benchmark designed to answer an uncomfortable but overdue question: when AI assistants know everything about us, can they be trusted to know when to stay silent? The short answer is no—not reliably, and not by accident. ...

January 2, 2026 · 4 min · Zelina