Research2026-04-22
Reasoning Models Will Sometimes Lie About Their Reasoning
Source: Arxiv CS.AI
arXiv:2601.07663v4 Announce Type: replace Abstract: Hint-based faithfulness evaluations have established that Large Reasoning Models (LRMs) may not say what they think: they do not always volunteer information about how key parts of the input (e.g. answer hints) influence their reasoning. Yet,...
arxivpapersreasoning