Cover image

Mind-Reading Without Telepathy: Predictive Concept Decoders

Opening — Why this matters now For years, AI interpretability has promised transparency while quietly delivering annotations, probes, and post-hoc stories that feel explanatory but often fail the only test that matters: can they predict what the model will actually do next? As large language models become agents—capable of long-horizon planning, policy evasion, and strategic compliance—interpretability that merely describes activations after the fact is no longer enough. What we need instead is interpretability that anticipates behavior. That is the ambition behind Predictive Concept Decoders (PCDs). ...

December 18, 2025 · 5 min · Zelina
Cover image

Stepwise Think-Critique: Teaching LLMs to Doubt Themselves (Productively)

Opening — Why this matters now Large Language Models have learned how to think out loud. What they still struggle with is knowing when that thinking is wrong — while it is happening. In high‑stakes domains like mathematics, finance, or policy automation, delayed error detection is not a feature; it is a liability. Most modern reasoning pipelines still follow an awkward split: first generate reasoning, then verify it — often with a separate model. Humans do not work this way. We reason and judge simultaneously. This paper asks a simple but uncomfortable question: what if LLMs were trained to do the same? ...

December 18, 2025 · 4 min · Zelina
Cover image

When Tokens Remember: Graphing the Ghosts in LLM Reasoning

Opening — Why this matters now Large language models don’t think—but they do accumulate influence. And that accumulation is exactly where most explainability methods quietly give up. As LLMs move from single-shot text generators into multi-step reasoners, agents, and decision-making systems, we increasingly care why an answer emerged—not just what token attended to what prompt word. Yet most attribution tools still behave as if each generation step lives in isolation. That assumption is no longer just naïve; it is actively misleading. ...

December 18, 2025 · 4 min · Zelina
Cover image

Greedy Enough to Win: When Loss Starts Driving the Learning Rate

Opening — Why this matters now Modern deep learning training is an odd contradiction. We obsess over architectures, data curation, and trillion-token scaling laws—then quietly accept Cosine Annealing as if it were gravity. Learning rate schedules are often inherited, not argued for. This paper challenges that complacency with a scheduler that does something almost offensive in its simplicity: it just watches the loss and reacts. ...

December 17, 2025 · 3 min · Zelina
Cover image

Model First, Think Later: Why LLMs Fail Before They Reason

Opening — Why this matters now As LLM agents graduate from clever chatbots to decision‑making systems, their failures are becoming less amusing and more expensive. We are no longer talking about wrong trivia answers; we are talking about broken schedules, invalid plans, unsafe workflows, and agents confidently violating constraints they were never told—explicitly—not to break. ...

December 17, 2025 · 4 min · Zelina
Cover image

Picking Less to Know More: When RAG Stops Ranking and Starts Thinking

Opening — Why this matters now Retrieval-Augmented Generation has a dirty secret: it keeps retrieving more context while quietly getting no smarter. As context windows balloon to 100K tokens and beyond, RAG systems dutifully shovel in passages—Top‑5, Top‑10, Top‑100—hoping recall will eventually rescue accuracy. It doesn’t. Accuracy plateaus. Costs rise. Attention diffuses. The model gets lost in its own evidence pile. ...

December 17, 2025 · 4 min · Zelina
Cover image

Ports, But Make Them Agentic: When LLMs Start Running the Yard

Opening — Why this matters now Ports are supposed to be automated. In practice, many of their most critical decisions still depend on a small priesthood of optimization specialists, tribal operational knowledge, and painfully slow deployment cycles. Vehicle Dispatching Systems (VDSs) — the logic that tells fleets of AGVs where to go and when — are a prime example. They promise up to 30% efficiency gains, yet stubbornly resist scaling from one terminal to another. ...

December 17, 2025 · 4 min · Zelina
Cover image

Reasoning Loops, Not Bigger Brains

Opening — Why this matters now For the past two years, AI progress has been narrated as a story of scale: more parameters, more data, more compute. Yet the ARC-AGI leaderboard keeps delivering an inconvenient counterexample. Small, scratch-trained models—no web-scale pretraining, no trillion-token diet—are routinely humiliating far larger systems on abstract reasoning tasks. This paper asks the uncomfortable question: where is the reasoning actually coming from? ...

December 17, 2025 · 3 min · Zelina
Cover image

Shaking the Stack: Teaching Seismology to Talk Back

Opening — Why this matters now Scientific software has a strange tradition: world‑class physics wrapped in workflows that feel frozen in the 1990s. Seismology is no exception. SPECFEM — arguably the gold standard for seismic wave simulation — delivers extraordinary numerical fidelity, but only after users survive a rite of passage involving fragile text files, shell scripts, and MPI incantations. ...

December 17, 2025 · 4 min · Zelina
Cover image

When Attention Learns to Breathe: Sparse Transformers for Sustainable Medical AI

Opening — Why this matters now Healthcare AI has quietly run into a contradiction. We want models that are richer—multi-modal, context-aware, clinically nuanced—yet we increasingly deploy them in environments that are poorer: fewer samples, missing modalities, limited compute, and growing scrutiny over energy use. Transformers, the industry’s favorite hammer, are powerful but notoriously wasteful. In medicine, that waste is no longer academic; it is operational. ...

December 17, 2025 · 4 min · Zelina