Cover image

When Models Start Remembering Too Much

Opening — Why this matters now Large language models are no longer judged solely by what they can generate, but by what they remember. As models scale and datasets balloon, a quiet tension has emerged: memorization boosts fluency and benchmark scores, yet it also raises concerns around data leakage, reproducibility, and governance. The paper examined here steps directly into that tension, asking not whether memorization exists — that debate is settled — but where, how, and why it concentrates. ...

February 2, 2026 · 3 min · Zelina
Cover image

FadeMem: When AI Learns to Forget on Purpose

Opening — Why this matters now The race to build smarter AI agents has mostly followed one instinct: remember more. Bigger context windows. Larger vector stores. Ever-growing retrieval pipelines. Yet as agents move from demos to long-running systems—handling days or weeks of interaction—this instinct is starting to crack. More memory does not automatically mean better reasoning. In practice, it often means clutter, contradictions, and degraded performance. Humans solved this problem long ago, not by remembering everything, but by forgetting strategically. ...

February 1, 2026 · 4 min · Zelina
Cover image

From Indicators to Intent: When Trading Libraries Grow Up

Opening — Why this matters now Most trading libraries die of obesity. They start life as tidy indicator toolkits and, over time, accumulate ad‑hoc features, half‑finished strategies, and opinionated shortcuts that quietly blur the line between describing markets and acting on them. Eventually, users stop trusting what a signal actually means. The latest strategyr refactor is interesting because it does the opposite: it removes functionality. Aggressively. And in doing so, it clarifies what kind of system this wants to be. ...

February 1, 2026 · 3 min · Zelina
Cover image

When Empathy Needs a Map: Benchmarking Tool‑Augmented Emotional Support

Opening — Why this matters now Emotional support from AI has quietly moved from novelty to expectation. People vent to chatbots after work, during grief, and in moments of burnout—not to solve equations, but to feel understood. Yet something subtle keeps breaking trust. The responses sound caring, but they are often wrong in small, revealing ways: the time is off, the location is imagined, the suggestion doesn’t fit reality. Empathy without grounding turns into polite hallucination. ...

February 1, 2026 · 4 min · Zelina
Cover image

MemCtrl: Teaching Small Models What *Not* to Remember

Opening — Why this matters now Embodied AI is hitting a very human bottleneck: memory. Not storage capacity, not retrieval speed—but judgment. Modern multimodal large language models (MLLMs) can see, reason, and act, yet when deployed as embodied agents they tend to remember too much, too indiscriminately. Every frame, every reflection, every redundant angle piles into context until the agent drowns in its own experience. ...

January 31, 2026 · 4 min · Zelina
Cover image

Metric Time Without the Clock: Making ASP Scale Again

Opening — Why this matters now Temporal reasoning has always been the Achilles’ heel of symbolic AI. The moment time becomes quantitative—minutes, deadlines, durations—logic programs tend to balloon, grounders panic, and scalability quietly exits the room. This paper lands squarely in that discomfort zone and does something refreshingly unglamorous: it makes time boring again. And boring, in this case, is good for business. ...

January 31, 2026 · 3 min · Zelina
Cover image

REASON About Reasoning: Why Neuro‑Symbolic AI Finally Needs Its Own Hardware

Opening — Why this matters now Neuro‑symbolic AI is having a quiet comeback. While large language models dominate headlines, the systems quietly outperforming them on math proofs, logical deduction, and safety‑critical reasoning all share the same uncomfortable truth: reasoning is slow. Not neural inference—reasoning. The paper behind REASON makes an unfashionable but crucial claim: if we want agentic AI that reasons reliably, interprets decisions, and operates in real time, we cannot keep pretending GPUs are good at symbolic and probabilistic logic. They aren’t. REASON is what happens when researchers finally stop forcing logic to cosplay as linear algebra. ...

January 31, 2026 · 4 min · Zelina
Cover image

Sequential Beats Parallel: When Deep Research Agents Learn to Reflect

Opening — Why this matters now The last year has been crowded with so-called deep research agents. Everyone parallelizes. Everyone fans out queries. Everyone promises doctoral-level synthesis at web speed. And yet, the leaderboard keeps telling an inconvenient story: throwing more parallel agents at a problem does not reliably buy depth. The paper “Deep Researcher with Sequential Plan Reflection and Candidates Crossover” enters this debate with a pointed thesis: research is not a map-reduce problem. If you want insight, you need memory, reflection, and the ability to change your mind mid-flight. ...

January 31, 2026 · 4 min · Zelina
Cover image

SokoBench: When Reasoning Models Lose the Plot

Opening — Why this matters now The AI industry has grown comfortable with a flattering assumption: if a model can reason, it can plan. Multi-step logic, chain-of-thought traces, and ever-longer context windows have encouraged the belief that we are edging toward systems capable of sustained, goal-directed action. SokoBench quietly dismantles that assumption. By stripping planning down to its bare minimum, the paper reveals an uncomfortable truth: today’s large reasoning models fail not because problems are complex—but because they are long. ...

January 31, 2026 · 3 min · Zelina
Cover image

When ERP Meets Attention: Teaching Transformers to Pack, Schedule, and Save Real Money

Opening — Why this matters now Enterprise Resource Planning (ERP) systems are excellent at recording what has happened. They are far less impressive at deciding what should happen next. When decision-making involves combinatorial explosions—packing furnaces, sequencing machines, allocating scarce inputs—ERP often falls back on brittle heuristics, slow solvers, or human intuition. None scale gracefully. ...

January 31, 2026 · 4 min · Zelina