Cover image

AGI by Committee: Why the First General Intelligence Won’t Arrive Alone

Opening — Why this matters now For years, AGI safety discussions have revolved around a single, looming figure: the model. One system. One alignment problem. One decisive moment. That mental model is tidy — and increasingly wrong. The paper “Distributional AGI Safety” argues that AGI is far more likely to emerge not as a monolith, but as a collective outcome: a dense web of specialized, sub‑AGI agents coordinating, trading capabilities, and assembling intelligence the way markets assemble value. AGI, in this framing, is not a product launch. It is a phase transition. ...

December 19, 2025 · 4 min · Zelina
Cover image

TOGGLE or Die Trying: Giving LLM Compression a Spine

Opening — Why this matters now LLM compression is having an identity crisis. On one side, we have brute-force pragmatists: quantize harder, prune deeper, pray nothing important breaks. On the other, we have theoreticians insisting that something essential is lost — coherence, memory, truthfulness — but offering little beyond hand-waving and validation benchmarks. As LLMs creep toward edge deployment — embedded systems, on-device assistants, energy‑capped inference — this tension becomes existential. You can’t just say “it seems fine.” You need guarantees. Or at least something better than vibes. ...

December 19, 2025 · 4 min · Zelina
Cover image

When Black Boxes Grow Teeth: Mapping What AI Can *Actually* Do

Opening — Why this matters now We are deploying black-box AI systems faster than we are understanding them. Large language models, vision–language agents, and robotic controllers are increasingly asked to do things, not just answer questions. And yet, when these systems fail, the failure is rarely spectacular—it is subtle, conditional, probabilistic, and deeply context-dependent. ...

December 19, 2025 · 3 min · Zelina
Cover image

Delegating to the Almost-Aligned: When Misaligned AI Is Still the Rational Choice

Opening — Why this matters now The AI alignment debate has a familiar rhythm: align the values first, deploy later. Sensible, reassuring—and increasingly detached from reality. In practice, we are already delegating consequential decisions to systems we do not fully understand, let alone perfectly align. Trading algorithms rebalance portfolios, recommendation engines steer attention, and autonomous agents negotiate, schedule, and filter on our behalf. The real question is no longer “Is the AI aligned?” but “Is it aligned enough to justify delegation, given what it can do better than us?” ...

December 18, 2025 · 4 min · Zelina
Cover image

Reasoning Loops, Not Bigger Brains

Opening — Why this matters now For the past two years, AI progress has been narrated as a story of scale: more parameters, more data, more compute. Yet the ARC-AGI leaderboard keeps delivering an inconvenient counterexample. Small, scratch-trained models—no web-scale pretraining, no trillion-token diet—are routinely humiliating far larger systems on abstract reasoning tasks. This paper asks the uncomfortable question: where is the reasoning actually coming from? ...

December 17, 2025 · 3 min · Zelina
Cover image

When Attention Learns to Breathe: Sparse Transformers for Sustainable Medical AI

Opening — Why this matters now Healthcare AI has quietly run into a contradiction. We want models that are richer—multi-modal, context-aware, clinically nuanced—yet we increasingly deploy them in environments that are poorer: fewer samples, missing modalities, limited compute, and growing scrutiny over energy use. Transformers, the industry’s favorite hammer, are powerful but notoriously wasteful. In medicine, that waste is no longer academic; it is operational. ...

December 17, 2025 · 4 min · Zelina
Cover image

When Medical AI Stops Guessing and Starts Asking

Opening — Why this matters now Medical AI has become very good at answering questions. Unfortunately, medicine rarely works that way. Pathology, oncology, and clinical decision-making are not single-query problems. They are investigative processes: observe, hypothesize, cross-check, revise, and only then conclude. Yet most medical AI benchmarks still reward models for producing one-shot answers — neat, confident, and often misleading. This mismatch is no longer academic. As multimodal models edge closer to clinical workflows, the cost of shallow reasoning becomes operational, regulatory, and ethical. ...

December 16, 2025 · 4 min · Zelina
Cover image

When Precedent Gets Nuanced: Why Legal AI Needs Dimensions, Not Just Factors

Opening — Why this matters now Legal AI has a habit of oversimplifying judgment. In the race to automate legal reasoning, we have learned how to encode rules, then factors, and eventually hierarchies of factors. But something stubborn keeps leaking through the abstractions: strength. Not whether a reason exists — but how strongly it exists. ...

December 16, 2025 · 4 min · Zelina
Cover image

When Reasoning Needs Receipts: Graphs Over Guesswork in Medical AI

Opening — Why this matters now Medical AI has a credibility problem. Not because large language models (LLMs) can’t answer medical questions—they increasingly can—but because they often arrive at correct answers for the wrong reasons. In medicine, that distinction is not academic. A shortcut that accidentally lands on the right diagnosis today can quietly institutionalize dangerous habits tomorrow. ...

December 16, 2025 · 3 min · Zelina
Cover image

When Small Models Learn From Their Mistakes: Arithmetic Reasoning Without Fine-Tuning

Opening — Why this matters now Regulated industries love spreadsheets and hate surprises. Finance, healthcare, and insurance all depend on tabular data—and all have strict constraints on where that data is allowed to go. Shipping sensitive tables to an API-hosted LLM is often a non‑starter. Yet small, on‑prem language models have a reputation problem: they speak fluently but stumble over arithmetic. ...

December 16, 2025 · 3 min · Zelina