BeClaude
Research2026-04-24

CAP: Controllable Alignment Prompting for Unlearning in LLMs

Source: Arxiv CS.AI

arXiv:2604.21251v1 Announce Type: cross Abstract: Large language models (LLMs) trained on unfiltered corpora inherently risk retaining sensitive information, necessitating selective knowledge unlearning for regulatory compliance and ethical safety. However, existing parameter-modifying methods face...

arxivpapersprompting