Cover image

Safety First, Reward Second — But Not Last

Opening — Why this matters now Reinforcement learning has spent the last decade mastering games, simulations, and neatly bounded optimization problems. Reality, inconveniently, is none of those things. In robotics, autonomous vehicles, industrial automation, and any domain where mistakes have real-world consequences, almost safe is simply unsafe. Yet most “safe RL” methods quietly rely on a compromise: allow some violations, average them out, and hope the system behaves. This paper refuses that bargain. It treats safety as a hard constraint, not a tunable preference—and then asks an uncomfortable question: can we still learn anything useful? ...

January 4, 2026 · 4 min · Zelina
Cover image

Trust No One, Train Together: Zero-Trust Federated Learning Grows Teeth

Opening — Why this matters now Critical infrastructure is no longer attacked by teenagers in hoodies. It is probed, poisoned, and patiently undermined by adversaries who understand distributed systems better than most defenders. From water treatment plants to national energy grids, Industrial IoT (IIoT) has become a strategic attack surface. Federated Learning (FL) was supposed to help—privacy-preserving, collaborative, decentralized. Instead, it quietly introduced a new problem: you are now trusting hundreds or thousands of autonomous agents not to lie. ...

January 4, 2026 · 4 min · Zelina
Cover image

When Fairness Fails in Groups: From Lone Counterexamples to Discrimination Clusters

Opening — Why this matters now Most algorithmic fairness debates still behave as if discrimination is a rounding error: rare, isolated, and best handled by catching a few bad counterexamples. Regulators ask whether a discriminatory case exists. Engineers ask whether any unfair input pair can be found. Auditors tick the box once a model is declared “2-fair.” ...

January 4, 2026 · 4 min · Zelina
Cover image

When Models Start Remembering: The Quiet Rise of Adaptive AI

Opening — Why this matters now For years, we have treated AI models like polished machines: train once, deploy, monitor, repeat. That worldview is now visibly cracking. The paper you just uploaded lands squarely on this fault line, arguing—quietly but convincingly—that modern AI systems are no longer well-described as static functions. They are processes. And processes remember. ...

January 4, 2026 · 3 min · Zelina
Cover image

When Riders Become Nodes: Mapping Fraud in Ride-Hailing with Graph Neural Networks

Opening — Why this matters now Ride-hailing fraud is no longer a fringe operational headache. It is a structural problem amplified by scale, incentives, and post-pandemic digitization. As platforms expanded, so did adversarial behavior: GPS spoofing, collusive rides, route inflation, and off-platform hire conversions quietly eroded trust and margins. Traditional fraud detection systems—feature-heavy, transaction-centric, and largely static—have struggled to keep up. The paper under review argues that the problem is not merely more fraud, but more relational fraud. And relational problems demand relational models. ...

January 4, 2026 · 4 min · Zelina
Cover image

AI Writes the Rules: When Formal Logic Teaches Language Discipline

Opening — Why this matters now Natural language is where most software failures quietly begin. Requirements are written in good faith, read with confidence, and implemented with subtle misunderstandings that only surface once systems are deployed, audited, or—worse—regulated. The uncomfortable truth is that natural language is flexible where engineering systems demand rigidity. This paper tackles that gap head‑on, proposing a method where formal logic leads and language follows. Instead of writing requirements first and wrestling semantics into place later, the authors invert the workflow: start from formal specification patterns, then systematically generate a controlled natural language (CNL) using an AI assistant. Precision first. Fluency second. ...

January 3, 2026 · 4 min · Zelina
Cover image

Gated, Not Gagged: Fixing Reward Hacking in Diffusion RL

Opening — Why this matters now Reinforcement learning has become the fashionable finishing school for large generative models. Pre-training gives diffusion models fluency; RL is supposed to give them manners. Unfortunately, in vision, those manners are often learned from a deeply unreliable tutor: proxy rewards. The result is familiar and embarrassing. Models learn to win the metric rather than satisfy human intent—rendering unreadable noise that scores well on OCR, or grotesquely saturated images that charm an aesthetic scorer but repel humans. This phenomenon—reward hacking—is not a bug in implementation. It is a structural failure in how we regularize learning. ...

January 3, 2026 · 4 min · Zelina
Cover image

Rotate Less, Quantize Better: OptRot and the Geometry of LLM Compression

Opening — Why this matters now Quantization is no longer a niche optimization; it is the price of admission for deploying large language models at scale. As model sizes balloon and inference budgets stubbornly refuse to follow, post-training quantization (PTQ) has become the default survival strategy. Yet one stubborn problem keeps resurfacing: outliers. A handful of extreme weights—or activations—can quietly wreck an otherwise elegant low‑bit deployment. This paper introduces OptRot, a method that tackles that problem not with more data, more calibration, or more training, but with something almost suspiciously modest: a carefully chosen rotation objective. ...

January 3, 2026 · 4 min · Zelina
Cover image

Talking to Yourself, but Make It Useful: Intrinsic Self‑Critique in LLM Planning

Opening — Why this matters now For years, the received wisdom in AI planning was blunt: language models can’t really plan. Early benchmarks—especially Blocksworld—made that verdict look almost charitable. Models hallucinated invalid actions, violated preconditions, and confidently declared failure states as success. The field responded by bolting on external verifiers, symbolic planners, or human-in-the-loop corrections. ...

January 3, 2026 · 3 min · Zelina
Cover image

Think First, Grasp Later: Why Robots Need Reasoning Benchmarks

Opening — Why this matters now Robotics has reached an awkward adolescence. Vision–Language–Action (VLA) models can now describe the world eloquently, name objects with near-human fluency, and even explain why a task should be done a certain way—right before dropping the object, missing the grasp, or confidently picking up the wrong thing. This is not a data problem. It’s a diagnostic one. ...

January 3, 2026 · 5 min · Zelina