Cover image

When ERP Meets Attention: Teaching Transformers to Pack, Schedule, and Save Real Money

Opening — Why this matters now Enterprise Resource Planning (ERP) systems are excellent at recording what has happened. They are far less impressive at deciding what should happen next. When decision-making involves combinatorial explosions—packing furnaces, sequencing machines, allocating scarce inputs—ERP often falls back on brittle heuristics, slow solvers, or human intuition. None scale gracefully. ...

January 31, 2026 · 4 min · Zelina
Cover image

PyraTok: When Video Tokens Finally Learn to Speak Human

Opening — Why this matters now Text-to-video models are scaling at an alarming pace. Resolution is no longer the bottleneck—semantic fidelity is. As generators push into 4K and even 8K regimes, a quieter but more consequential problem emerges underneath: the tokenizer. If visual tokens do not align with language, no amount of diffusion steps will save downstream reasoning, control, or zero-shot transfer. ...

January 24, 2026 · 3 min · Zelina
Cover image

When Models Guess the Verb by Looking at the Drawer

Opening — Why this matters now If you have ever watched a video model confidently predict opening drawer when the person is clearly closing it, you have already encountered the core problem of modern compositional video understanding: the model isn’t really watching the action. It is guessing. As video models are increasingly deployed in robotics, industrial monitoring, and human–AI interaction, the ability to correctly generalize unseen verb–object combinations is no longer academic. A robot that confuses opening with closing is not merely inaccurate—it is dangerous. ...

January 24, 2026 · 4 min · Zelina
Cover image

Skeletons in the Proof Closet: When Lean Provers Need Hints, Not More Compute

Opening — Why this matters now Neural theorem proving has entered its industrial phase. With reinforcement learning pipelines, synthetic data factories, and search budgets that would make a chess engine blush, models like DeepSeek‑Prover‑V1.5 are widely assumed to have internalized everything there is to know about formal proof structure. This paper politely disagrees. Under tight inference budgets—no massive tree search, no thousand-sample hail‑Mary—the author shows that simple, almost embarrassingly old‑fashioned structural hints still deliver large gains. Not new models. Not more data. Just better scaffolding. ...

January 23, 2026 · 4 min · Zelina
Cover image

Vibe Coding a Theorem Prover: When LLMs Prove (and Break) Themselves

Opening — Why this matters now LLMs can write code, explain proofs, and occasionally hallucinate both with equal confidence. So the obvious next question—posed almost mischievously in this paper—is whether an LLM can code a theorem prover that itself relies on LLMs. Not as a demo. Not as a toy. But as a fully automatic, kernel-checked prover that runs on a laptop and outperforms Isabelle’s industrial-grade automation in at least some regimes. ...

January 11, 2026 · 4 min · Zelina
Cover image

When Solvers Guess Smarter: Teaching SMT to Think in Functions

Opening — Why this matters now Quantified SMT solving has always lived in an uncomfortable space between elegance and brute force. As models grew richer—mixing non-linear arithmetic, real-valued domains, and uninterpreted functions—the solvers stayed stubbornly syntactic. They match patterns. They enumerate. They hope. Meanwhile, large language models have quietly absorbed a century’s worth of mathematical intuition. AquaForte asks an obvious but previously taboo question: what if we let SMT solvers borrow that intuition—without surrendering formal guarantees? ...

January 11, 2026 · 3 min · Zelina
Cover image

When Prophet Meets Perceptron: Chasing Alpha with NP‑DNN

Opening — Why this matters now Stock prediction papers arrive with clockwork regularity, each promising to tame volatility with yet another hybrid architecture. Most quietly disappear after publication. A few linger—usually because they claim eye‑catching accuracy. This paper belongs to that second category, proposing a Neural Prophet + Deep Neural Network (NP‑DNN) stack that reportedly delivers over 93%–99% accuracy in stock market prediction. ...

January 9, 2026 · 3 min · Zelina
Cover image

NeuroSPICE: When Circuits Stop Ticking and Start Thinking

Opening — Why this matters now Circuit simulation has always been an exercise in controlled compromise. We discretize time, linearize nonlinearity, and hope the numerical solver behaves. SPICE has done this extraordinarily well for decades—but it was built for an era where devices were mostly electrical, mostly local, and mostly cooperative. That era is ending. Ferroelectrics, photonics, thermal coupling in 3D ICs, and other strongly nonlinear or multi-physics effects are turning compact modeling into a brittle art. Against this backdrop, NeuroSPICE proposes something mildly heretical: stop stepping through time altogether. ...

December 30, 2025 · 3 min · Zelina
Cover image

Teaching Has a Poker Face: Why Teacher Emotion Needs Its Own AI

Opening — Why this matters now AI has become remarkably good at reading emotions—just not the kind that actually matter in classrooms. Most sentiment models are trained on people being honest with their feelings: tweets, movie reviews, reaction videos. Teachers, unfortunately for the models, are professionals. They perform. They regulate. They smile through frustration and project enthusiasm on command. As a result, generic sentiment analysis treats classrooms as emotionally flat—or worse, mislabels them entirely. ...

December 24, 2025 · 4 min · Zelina
Cover image

When LLMs Stop Talking and Start Choosing Algorithms

Opening — Why this matters now Large Language Models are increasingly invited into optimization workflows. They write solvers, generate heuristics, and occasionally bluff their way through mathematical reasoning. But a more uncomfortable question has remained largely unanswered: do LLMs actually understand optimization problems—or are they just eloquent impostors? This paper tackles that question head‑on. Instead of judging LLMs by what they say, it examines what they encode. And the results are quietly provocative. ...

December 16, 2025 · 4 min · Zelina