Cover image

SD‑RAG: Don’t Trust the Model, Trust the Pipeline

Opening — Why this matters now RAG was supposed to make LLMs safer. Instead, it quietly became a liability. As enterprises rushed to bolt retrieval layers onto large language models, they unintentionally created a new attack surface: sensitive internal data flowing straight into a model that cannot reliably distinguish instructions from content. Prompt injection is not a corner case anymore—it is the default threat model. And telling the model to “behave” has proven to be more of a suggestion than a guarantee. ...

January 20, 2026 · 4 min · Zelina
Cover image

Who’s Really in Charge? Epistemic Control After the Age of the Black Box

Opening — Why this matters now Machine learning has become science’s most productive employee—and its most awkward colleague. It delivers predictions at superhuman scale, spots patterns no graduate student could ever see, and does so without asking for coffee breaks or tenure. But as ML systems increasingly mediate discovery, a more uncomfortable question has resurfaced: who is actually in control of scientific knowledge production? ...

January 20, 2026 · 5 min · Zelina
Cover image

Aligned or Just Agreeable? Why Accuracy Is a Terrible Proxy for AI–Human Alignment

Opening — Why this matters now As large language models quietly migrate from text generators to decision makers, the industry has developed an unhealthy obsession with the wrong question: Did the model choose the same option as a human? Accuracy, F1, and distributional overlap have become the default proxies for alignment. They are also deeply misleading. ...

January 19, 2026 · 4 min · Zelina
Cover image

Greedy, but Not Blind: Teaching Optimization to Listen

Opening — Why this matters now Public-sector AI has a credibility problem. Not because it cannot optimize—but because it optimizes too cleanly. In health system planning, decisions are rarely about pure efficiency. They are negotiated compromises shaped by terrain, politics, institutional memory, and hard-earned intuition. Classic optimization methods politely ignore all that. This paper tackles a question many planners quietly ask but rarely formalize: Can we let algorithms optimize without silencing human judgment—and still keep mathematical guarantees intact? ...

January 19, 2026 · 4 min · Zelina
Cover image

Houston, We Have a Benchmark: When Agentic AI Meets Orbital Reality

Opening — Why this matters now Agentic large language models are increasingly marketed as generalist planners: systems that can reason, act, and adapt across domains without bespoke algorithmic scaffolding. The pitch is seductive—why maintain a zoo of solvers when a single agent can plan everything from code refactors to satellite schedules? AstroReason-Bench arrives as a cold shower. ...

January 19, 2026 · 4 min · Zelina
Cover image

Probe, Then Commit: Why Solver Tuning Finally Grew Up

Opening — Why this matters now Constraint programming (CP) has always promised elegance: state the problem, let the solver do the work. In practice, however, seasoned users know the uncomfortable truth—solver performance lives or dies by hyperparameters most people neither understand nor have time to tune. As problem instances grow larger and solver configurations explode combinatorially, manual tuning has become less of an art and more of a liability. The paper Hyperparameter Optimization of Constraint Programming Solvers confronts this reality head-on, proposing a framework that finally treats solver configuration as what it is: a resource allocation problem under uncertainty. ...

January 19, 2026 · 4 min · Zelina
Cover image

Punching Above Baselines: When Boxing Strategy Learns to Differentiate

Opening — Why this matters now Elite sport has quietly become an optimization problem. Marginal gains are no longer found in strength alone, but in decision quality under pressure. Boxing, despite its reputation for instinct and grit, has remained stubbornly analog in this regard. Coaches still scrub footage frame by frame, hunting for patterns that disappear as fast as they emerge. ...

January 19, 2026 · 4 min · Zelina
Cover image

Think-with-Me: When LLMs Learn to Stop Thinking

Opening — Why this matters now The AI industry has developed an unhealthy obsession with thinking longer. More tokens, deeper chains, bigger context windows—surely that must mean better reasoning. Except, increasingly, it doesn’t. Large Reasoning Models (LRMs) often reason past the point of usefulness, slipping into self-validation loops or overwriting correct answers with unnecessary exploration. This paper proposes a heretical idea in the age of scaling: maybe the model doesn’t need to think more—it needs to know when to stop. ...

January 19, 2026 · 3 min · Zelina
Cover image

When LLMs Read the Room: Predictive Process Monitoring Without the Data Buffet

Opening — Why this matters now Predictive Process Monitoring (PPM) has always promised operational foresight: knowing how long a case will take, whether a costly activity will happen, or when things are about to go wrong. The catch has been brutally consistent — you need a lot of data. Thousands of traces. Clean logs. Stable processes. ...

January 19, 2026 · 5 min · Zelina
Cover image

Fish in the Ocean, Not Needles in the Haystack

Opening — Why this matters now Long-context multimodal models are starting to look fluent enough to pass surface-level exams on scientific papers. They answer questions correctly. They summarize convincingly. And yet, something feels off. The answers often arrive without a visible path—no trail of figures, no textual anchors, no defensible reasoning chain. In other words, the model knows what to say, but not necessarily why it is true. ...

January 18, 2026 · 4 min · Zelina