Research2026-05-12
Sanity Checks for Long-Form Hallucination Detection
Source: Arxiv CS.AI
arXiv:2605.08346v1 Announce Type: cross Abstract: Hallucination detection methods for large language models increasingly operate on chain-of-thought reasoning traces, yet it remains unclear whether they evaluate the reasoning itself or merely exploit surface correlates of the final answer. We...
arxivpapers