Cover image

The Patient Is Not a Moving Document: Why Clinical AI Needs World Models

Opening — Why this matters now Clinical AI has quietly hit a ceiling. Over the past five years, large language models trained on electronic health records (EHRs) have delivered impressive gains: better coding, stronger risk prediction, and even near‑physician exam performance. But beneath those wins lies an uncomfortable truth. Most clinical foundation models still treat patients as documents—static records to be summarized—rather than systems evolving over time. ...

January 30, 2026 · 4 min · Zelina
Cover image

Triage by Token: When Context Clues Quietly Override Clinical Judgment

Opening — Why this matters now Large language models are quietly moving from clerical assistance to clinical suggestion. In emergency departments (EDs), where seconds matter and triage decisions shape outcomes, LLM-based decision support tools are increasingly tempting: fast, consistent, and seemingly neutral. Yet neutrality in language does not guarantee neutrality in judgment. This paper interrogates a subtle but consequential failure mode: latent bias introduced through proxy variables. Not overt racism. Not explicit socioeconomic labeling. Instead, ordinary contextual cues—how a patient arrives, where they live, how often they visit the ED—nudging model outputs in clinically unjustified ways. ...

January 24, 2026 · 4 min · Zelina
Cover image

Think Before You Beam: When AI Learns to Plan Like a Physicist

Opening — Why this matters now Automation in healthcare has a credibility problem. Not because it performs poorly—but because it rarely explains why it does what it does. In high-stakes domains like radiation oncology, that opacity isn’t an inconvenience; it’s a blocker. Regulators demand traceability. Clinicians demand trust. And black-box optimization, however accurate, keeps failing both. ...

December 24, 2025 · 4 min · Zelina
Cover image

When Bigger Isn’t Smarter: Stress‑Testing LLMs in the ICU

Opening — Why this matters now Healthcare AI has entered its foundation model phase. LLMs trained on trillions of tokens are being casually proposed for everything from triage to prognosis, often with an implicit assumption: bigger models must understand patients better. This paper quietly punctures that assumption. By benchmarking LLMs against smaller, task‑focused language models (SLMs) on shock prediction in ICUs, the authors confront a question most vendors avoid: Do LLMs actually predict future clinical deterioration better—or do they merely sound more convincing? ...

December 24, 2025 · 3 min · Zelina
Cover image

Doctor GPT, But Make It Explainable

Opening — Why this matters now Healthcare systems globally suffer from a familiar triad: diagnostic bottlenecks, rising costs, and a shortage of specialists. What makes this crisis especially stubborn is not just capacity—but interaction. Diagnosis is fundamentally conversational, iterative, and uncertain. Yet most AI diagnostic tools still behave like silent oracles: accurate perhaps, but opaque, rigid, and poorly aligned with how humans actually describe illness. ...

December 22, 2025 · 4 min · Zelina
Cover image

Painkillers with Foresight: Teaching Machines to Anticipate Cancer Pain

Opening — Why this matters now Cancer pain is rarely a surprise to clinicians. Yet it still manages to arrive uninvited, often at night, often under-treated, and almost always after the window for calm, preventive adjustment has closed. In lung cancer wards, up to 90% of patients experience moderate to severe pain episodes — and most of these episodes are predictable in hindsight. ...

December 19, 2025 · 4 min · Zelina
Cover image

When LLMs Get Fatty Liver: Diagnosing AI-MASLD in Clinical AI

Opening — Why this matters now AI keeps passing medical exams, acing board-style questions, and politely explaining pathophysiology on demand. Naturally, someone always asks the dangerous follow-up: So… can we let it talk to patients now? This paper answers that question with clinical bluntness: not without supervision, and certainly not without consequences. When large language models (LLMs) are exposed to raw, unstructured patient narratives—the kind doctors hear every day—their performance degrades in a very specific, pathological way. The authors call it AI-MASLD: AI–Metabolic Dysfunction–Associated Steatotic Liver Disease. ...

December 15, 2025 · 4 min · Zelina
Cover image

HyFedRAG: Caching Privacy into Federated RAG

Centralized Retrieval-Augmented Generation (RAG) systems promise smarter answers, but they quietly assume one big, clean dataset in one place. Reality is far messier: hospitals, insurers, or financial groups each hold their own silo, often in incompatible formats, and none are willing—or legally allowed—to pool raw data. The HyFedRAG framework tackles this head‑on by making RAG federated, heterogeneous, and privacy‑aware. Edge First, Cloud Second Instead of centralizing records, HyFedRAG runs retrieval at the edge. Each hospital or business unit: ...

September 12, 2025 · 3 min · Zelina
Cover image

Charting a Better Bedside: When Agentic RL Teaches RAG to Diagnose

Why this paper matters: Retrieval‑augmented generation (RAG) has been the default answer to “how do we make LLMs factual?” But clinical work is not a single hop to a single document; it’s a workflow—observe, hypothesize, retrieve, cross‑check, and only then decide. Deep‑DxSearch reframes RAG as a sequential policy, trained end‑to‑end with reinforcement learning (RL) so the model learns when to reason internally and when to consult guidelines, match similar patients, or search broader knowledge—before committing to a diagnosis. That design change is the story. ...

August 24, 2025 · 5 min · Zelina
Cover image

Confounder Hunters: How LLM Agents are Rewriting the Rules of Causal Inference

When Hidden Variables Become Hidden Costs In causal inference, confounders are the uninvited guests at your data party — variables that influence both treatment and outcome, quietly skewing results. In healthcare, failing to adjust for them can turn life-saving insights into misleading noise. Traditionally, finding these culprits has been the realm of domain experts, a slow and costly process that doesn’t scale well. The paper from National Sun Yat-Sen University proposes a radical alternative: put Large Language Model (LLM)-based agents into the causal inference loop. These agents don’t just crunch numbers — they reason, retrieve domain knowledge, and iteratively refine estimates, effectively acting as tireless, always-available junior experts. ...

August 12, 2025 · 3 min · Zelina