Research2026-05-07
MAGE: Safeguarding LLM Agents against Long-Horizon Threats via Shadow Memory
Source: Arxiv CS.AI
arXiv:2605.03228v1 Announce Type: cross Abstract: As large language model (LLM)-powered agents are increasingly deployed to perform complex, real-world tasks, they face a growing class of attacks that exploit extended user-agent-environment interactions to pursue malicious objectives improbable in...
arxivpapersagents