Cover image

The Two Minds of Finance: Testing LLMs for Divergence and Discipline

How do we judge whether an AI is thinking like a human—or at least like a financial analyst? A new benchmark, ConDiFi, offers a compelling answer: test not just whether an LLM gets the right answer, but whether it can explore possible ones. That’s because true financial intelligence lies not only in converging on precise conclusions but in diverging into speculative futures. Most benchmarks test convergent thinking: answer selection, chain-of-thought, or multi-hop reasoning. But strategic fields like finance also demand divergent thinking—creative, open-ended scenario modeling that considers fat-tail risks and policy surprises. ConDiFi (short for Convergent-Divergent for Finance) is the first serious attempt to capture both dimensions in one domain-specific benchmark. ...

July 25, 2025 · 4 min · Zelina
Cover image

Latent Brilliance: Turning LLMs into Creativity Engines

What if we stopped asking language models to “be creative”—and instead let them explore creativity the way humans brainstorm: by remixing ideas, nudging boundaries, and iterating through meaningful variations? That’s exactly what Large Language Models as Innovators proposes: a novel framework that leverages the latent embedding space of ideas—not prompts—to drive controlled, domain-agnostic creativity. Rather than relying on handcrafted rules or complex prompting tricks, the authors show how LLMs can generate original and relevant ideas by interpolating between known concepts, evaluating results, and refining outputs over time. ...

July 21, 2025 · 3 min · Zelina