BeClaude
Research2026-05-14

REALISTA: Realistic Latent Adversarial Attacks that Elicit LLM Hallucinations

Source: Arxiv CS.AI

arXiv:2605.12813v1 Announce Type: cross Abstract: Large language models (LLMs) achieve strong performance across many tasks but remain vulnerable to hallucinations, motivating the need for realistic adversarial prompts that elicit such failures. We formulate hallucination elicitation as a...

arxivpapers