Research2026-04-24
CAP: Controllable Alignment Prompting for Unlearning in LLMs
Source: Arxiv CS.AI
arXiv:2604.21251v1 Announce Type: cross Abstract: Large language models (LLMs) trained on unfiltered corpora inherently risk retaining sensitive information, necessitating selective knowledge unlearning for regulatory compliance and ethical safety. However, existing parameter-modifying methods face...
arxivpapersprompting