When Your Agent Knows It’s Lying: Detecting Tool-Calling Hallucinations from the Inside
Opening — Why this matters now LLM-powered agents are no longer a novelty. They calculate loans, file expenses, query databases, orchestrate workflows, and—when things go wrong—quietly fabricate tool calls that look correct but aren’t. Unlike textual hallucinations, tool-calling hallucinations don’t merely misinform users; they bypass security controls, corrupt data, and undermine auditability. In short: once agents touch real systems, hallucinations become operational risk. ...