When LLMs Stop Guessing and Start Arguing: A Two‑Stage Cure for Health Misinformation
Opening — Why this matters now Health misinformation is not a fringe problem anymore. It is algorithmically amplified, emotionally charged, and often wrapped in scientific‑looking language that fools both humans and machines. Most AI fact‑checking systems respond by doing more — more retrieval, more reasoning, more prompts. This paper argues the opposite: do less first, think harder only when needed. ...