BeClaude
Research2026-05-14

Negation Neglect: When models fail to learn negations in training

Source: Arxiv CS.AI

arXiv:2605.13829v1 Announce Type: cross Abstract: We introduce Negation Neglect, where finetuning LLMs on documents that flag a claim as false makes them believe the claim is true. For example, models are finetuned on documents that convey "Ed Sheeran won the 100m gold at the 2024 Olympics" but...

arxivpapers