Research2026-04-20
Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations
Source: Arxiv CS.AI
arXiv:2603.03332v3 Announce Type: replace-cross Abstract: Chain-of-Thought (CoT) prompting has emerged as a foundational technique for eliciting reasoning from Large Language Models (LLMs), yet the robustness of this approach to corruptions in intermediate reasoning steps remains poorly understood....
arxivpapersrag