Safety · intermediate

What is Jailbreak?

A plain-English explanation of Jailbreak (Jailbreak) — what it means, why it matters, and how it is used in AI.

Jailbreak
Jailbreak
A jailbreak is a prompt, technique, or sequence of inputs designed to bypass an AI model's safety guidelines and make it produce content it was trained to refuse.
"A user might frame a harmful request as a fictional scenario to try to get a model to produce content it would normally refuse."

Also known as: Prompt injection, adversarial prompt, safety bypass

Why does Jailbreak matter?

Understanding jailbreaks matters for AI safety researchers, developers deploying AI in sensitive contexts, and policymakers.

Practice this term

The best way to remember Jailbreak is to practice unscrambling it. AI Terminology Scrambler uses spaced repetition to help you learn and retain AI vocabulary in just a few minutes a day.

Practice Jailbreak now →

Related AI terms