BeClaude
Back to News
Release2024-04-19

The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions

Source: OpenAI

Today's LLMs are susceptible to prompt injections, jailbreaks, and other attacks that allow adversaries to overwrite a model's original instructions with their own malicious prompts.

openaigpt