Cover image

The Sentiment Edge: How FinDPO Trains LLMs to Think Like Traders

Financial markets don’t reward the loudest opinions. They reward the most timely, well-calibrated ones. FinDPO, a new framework by researchers from Imperial College London, takes this lesson seriously. It proposes a bold shift in how we train language models to read market sentiment. Rather than relying on traditional supervised fine-tuning (SFT), FinDPO uses Direct Preference Optimization (DPO) to align a large language model with how a human trader might weigh sentiment signals in context. And the results are not just academic — they translate into real money. ...

July 27, 2025 · 3 min · Zelina
Cover image

From Graph to Grit: Diagnosing Warehouse Bottlenecks with LLMs and Knowledge Graphs

In the age of Digital Twins and hyper-automated warehouses, simulations are everywhere—but insights are not. Discrete Event Simulations (DES) generate rich, micro-level data on logistics flows, delays, and resource utilization, yet interpreting these data remains painfully manual, fragile, and siloed. This paper from Quantiphi introduces a compelling solution: transforming raw simulation outputs into a Knowledge Graph (KG) and querying it via an LLM agent that mimics human investigative reasoning. It’s a shift from spreadsheet-style summaries to an interactive AI assistant that explains why something is slow, where the bottleneck is, and what needs attention. ...

July 26, 2025 · 3 min · Zelina
Cover image

Planners, Meet Your Smart Sidekick

Imagine asking, “Why wasn’t Order A scheduled for production yesterday?” and getting not just an answer, but a causal breakdown, an alternative plan, and a visual comparison — all without involving your operations research (OR) consultant. That’s exactly what SMARTAPS delivers. Built by Huawei researchers, SMARTAPS is a tool-augmented LLM interface for interacting with Advanced Planning Systems (APS) using natural language. It doesn’t try to replace optimization solvers — it simply makes them accessible. In doing so, it redefines how planners interact with complex decision-making models. ...

July 26, 2025 · 3 min · Zelina
Cover image

The Most Dangerous Query Is the One You Don't Question

In the age of natural language interfaces to databases (NLIDBs), asking the right question has never been easier—or more perilous. While systems like ChatGPT or SQL-Palm can convert everyday English into valid SQL, they often do so without interrogating the quality of the question itself. And as Peter Drucker warned, “The most dangerous thing is asking the wrong question.” Enter VeriMinder, a system built not to improve SQL syntax or execution accuracy—but to diagnose and refine the analytical intent behind the user’s query. It tackles a deceptively simple yet far-reaching problem: a well-formed SQL query that answers a poorly formed question can yield confident but misleading insights. This is particularly problematic in enterprise settings where non-technical users rely on LLM-based BI assistants. ...

July 25, 2025 · 4 min · Zelina
Cover image

Think Twice, Then Speak: Deliberative Searcher and the Future of Reliable LLMs

When a large language model (LLM) answers your question with a high degree of confidence, do you trust it? What if it’s wrong—but still confident? The stakes are high in real-world applications, from legal guidance to enterprise decision support. Yet today’s LLMs remain notoriously unreliable in aligning their confidence with correctness. The paper Deliberative Searcher: Improving LLM Reliability via Reinforcement Learning with Constraints (Yin et al., 2025) offers a bold response: rewire LLMs to be reasoning-primary and information-secondary. Instead of front-loading search and passively absorbing evidence, Deliberative Searcher acts more like a prudent investigator: it thinks, self-assesses, retrieves external information only when needed, and calibrates its confidence step-by-step. Crucially, it learns this behavior through a custom constrained reinforcement learning regime. ...

July 23, 2025 · 3 min · Zelina
Cover image

Latent Brilliance: Turning LLMs into Creativity Engines

What if we stopped asking language models to “be creative”—and instead let them explore creativity the way humans brainstorm: by remixing ideas, nudging boundaries, and iterating through meaningful variations? That’s exactly what Large Language Models as Innovators proposes: a novel framework that leverages the latent embedding space of ideas—not prompts—to drive controlled, domain-agnostic creativity. Rather than relying on handcrafted rules or complex prompting tricks, the authors show how LLMs can generate original and relevant ideas by interpolating between known concepts, evaluating results, and refining outputs over time. ...

July 21, 2025 · 3 min · Zelina
Cover image

Serverless Bulls and Bears: How One Developer Built a Real-Time Stock Analyst with Zero Infrastructure

Most real-time financial systems rely on deep stacks of infrastructure, from custom APIs to cloud VMs and high-frequency data ingestion pipelines. But what if a single developer could deploy a daily-updating, AI-powered stock analysis engine without a single server? That’s exactly what Taniv Ashraf set out to do — and accomplished — in his recent case study on a fully serverless architecture using Google Gemini, GitHub Actions, and static web hosting. The result is an elegantly simple yet conceptually powerful demonstration of how qualitative LLM analysis and automation tools can replace entire categories of financial tooling — if wielded strategically. ...

July 15, 2025 · 4 min · Zelina
Cover image

Tables Turned: Why LLM-Based Table Agents Are the Next Big Leap in Business AI

When most people think of AI today, they picture text generation, image synthesis, or copilots answering emails. But beneath the surface of digital transformation lies an often-overlooked backbone of enterprise work: tables. Spreadsheets, databases, and semi-structured tabular documents are still where critical operations happen — from finance to health records to logistics. A recent survey paper, Toward Real-World Table Agents, pushes us to rethink how AI interacts with tabular data. Instead of treating tables as static inputs, the authors argue that tables are evolving into active data canvases — and LLM-based Table Agents are poised to become their intelligent orchestrators. ...

July 15, 2025 · 4 min · Zelina
Cover image

The Retrieval-Reasoning Tango: Charting the Rise of Agentic RAG

In the AI race to make large language models both factual and reasoned, two camps have emerged: one focused on retrieval-augmented generation (RAG) to fight hallucination, the other on long-chain reasoning to mimic logic. But neither wins alone. This week’s survey by Li et al. (2025), Towards Agentic RAG with Deep Reasoning, delivers the most comprehensive synthesis yet of the field’s convergence point: synergized RAG–Reasoning. It’s no longer a question of whether retrieval helps generation or reasoning helps retrieval—but how tightly the two can co-evolve, often under the coordination of autonomous agents. ...

July 15, 2025 · 3 min · Zelina
Cover image

Plug Me In: Why LLMs with Tools Beat LLMs with Size

The latest research out of Heriot-Watt University doesn’t just challenge the notion that bigger is better — it quietly dismantles it. In their newly released Athena framework, Nripesh Niketan and Hadj Batatia demonstrate how integrating external APIs into LLM pipelines can outperform even the likes of GPT-4o and LLaMA-Large on real tasks like math and science. And they didn’t just beat them — they lapped them. Why GPT-4 Still Fumbles Math Ask GPT-4o to solve a college-level math problem, and it might hallucinate steps or miss basic arithmetic. The reason? LLMs, even at trillion-parameter scale, are not calculators. They’re probabilistic machines trained on patterns, not deterministic reasoners. ...

July 14, 2025 · 3 min · Zelina