Research2026-04-22
Reasoning Structure Matters for Safety Alignment of Reasoning Models
Source: Arxiv CS.AI
arXiv:2604.18946v1 Announce Type: new Abstract: Large reasoning models (LRMs) achieve strong performance on complex reasoning tasks but often generate harmful responses to malicious user queries. This paper investigates the underlying cause of these safety risks and shows that the issue lies in the...
arxivpapersreasoningsafety