Research2026-05-12
CachePrune: Teaching LLMs What Not to Follow via KV-Cache Editing
Source: Arxiv CS.AI
arXiv:2504.21228v3 Announce Type: replace-cross Abstract: Large Language Models (LLMs) are susceptible to indirect prompt injection attacks, where the model inadvertently responds to instructions injected into the prompt context. This vulnerability stems from LLMs' inability to distinguish between...
arxivpapers