Fish in the Ocean, Not Needles in the Haystack
Opening — Why this matters now Long-context multimodal models are starting to look fluent enough to pass surface-level exams on scientific papers. They answer questions correctly. They summarize convincingly. And yet, something feels off. The answers often arrive without a visible path—no trail of figures, no textual anchors, no defensible reasoning chain. In other words, the model knows what to say, but not necessarily why it is true. ...