AI for Financial Document Review

Where AI can help finance teams review statements, contracts, memos, and disclosures faster, and where exact review still belongs to humans.

March 16, 2026 · 5 min

ETH Support/Resistance Zone Detector

What this market-analytics demo illustrates about AI-adjacent decision support, and how to position such demos responsibly for clients.

March 16, 2026 · 5 min

Expense Categorization with LLMs

How to use LLMs to turn messy receipts, descriptions, and invoices into structured expense categories without weakening accounting controls.

March 16, 2026 · 5 min

Forecast Budgets with AI

Where AI can genuinely help budget forecasting and where finance teams still need disciplined modeling, assumptions, and human judgment.

March 16, 2026 · 5 min

Smart Invoicing with AI

How to use AI to extract, validate, and route invoice information while keeping finance controls, approval logic, and exception handling intact.

March 16, 2026 · 5 min

Where AI Helps and Fails in Accounting

A realistic view of where AI is useful in accounting work and where human controls, policy interpretation, and exactness still dominate.

March 16, 2026 · 5 min
Cover image

When Prophet Meets Perceptron: Chasing Alpha with NP‑DNN

Opening — Why this matters now Stock prediction papers arrive with clockwork regularity, each promising to tame volatility with yet another hybrid architecture. Most quietly disappear after publication. A few linger—usually because they claim eye‑catching accuracy. This paper belongs to that second category, proposing a Neural Prophet + Deep Neural Network (NP‑DNN) stack that reportedly delivers over 93%–99% accuracy in stock market prediction. ...

January 9, 2026 · 3 min · Zelina
Cover image

Trading Without Cheating: Teaching LLMs to Reason When Markets Lie

Opening — Why this matters now Large Language Models have learned how to solve math problems, write production-grade code, and even argue convincingly with themselves. Yet when we drop them into financial markets—arguably the most incentive-aligned environment imaginable—they develop a bad habit: they cheat. Not by insider trading, of course. By doing something more subtle and far more dangerous: reward hacking. They learn to chase noisy returns, memorize lucky assets, and fabricate reasoning after the fact. The profits look real. The logic isn’t. ...

January 8, 2026 · 4 min · Zelina
Cover image

Model Portfolio: When LLMs Sit the CFA

If your firm is debating whether to trust an LLM on investment memos, this study is a gift: 1,560 questions from official CFA mock exams across Levels I–III, run on three model archetypes—multimodal generalist (GPT‑4o), deep-reasoning specialist (GPT‑o1), and lightweight cost‑saver (o3‑mini)—both zero‑shot and with a domain‑reasoning RAG pipeline. Below is what matters for adoption, not just leaderboard bragging rights. What the paper really shows Reasoning beats modality for finance. The reasoning‑optimized model (GPT‑o1) dominates across levels; the generalist (GPT‑4o) is inconsistent, especially on math‑heavy Level II. RAG helps where context is long and specialized. Gains are largest at Level III (portfolio cases) and in Fixed Income/Portfolio Management, modest at Level I. Retrieval cannot fix arithmetic. Most errors are knowledge gaps, not reading problems. Readability barely moves accuracy; the bottleneck is surfacing the right curriculum facts and applying them. Cost–accuracy has a sweet spot. o3‑mini + targeted RAG is strong enough for high‑volume workflows; o1 should be reserved for regulated, high‑stakes analysis. Executive snapshot CFA Level GPT‑4o (ZS → RAG) GPT‑o1 (ZS → RAG) o3‑mini (ZS → RAG) Takeaway I 78.6% → 79.4% 94.8% → 94.8% 87.6% → 88.3% Foundations already in‑model; RAG adds little II 59.6% → 60.5% 89.3% → 91.4% 79.8% → 84.3% Level II exposes math + integration gaps; RAG helps smaller models most III 64.1% → 68.6% 79.1% → 87.7% 70.9% → 76.4% Case‑heavy; RAG is decisive, especially for o1 ZS = zero‑shot. Accuracies are from the paper’s aggregated results. ...

September 11, 2025 · 4 min · Zelina
Cover image

Dial M—for Markets: Brain‑Scanning and Steering LLMs for Finance

TL;DR A new paper shows how to insert a sparse, interpretable layer into an LLM to expose plain‑English concepts (e.g., sentiment, risk, timing) and steer them like dials without retraining. In finance news prediction, these interpretable features outperform final‑layer embeddings and reveal that sentiment, market/technical cues, and timing drive most short‑horizon alpha. Steering also debiases optimism, lifting Sharpe by nudging the model negative on sentiment. Why this matters (and what’s new) Finance teams have loved LLMs’ throughput but hated their opacity. This paper demonstrates a lightweight path to transparent performance: ...

September 1, 2025 · 4 min · Zelina