A prompt that tricks an LLM into bypassing its safety training — making it output things it was specifically trained to refuse. An ongoing cat-and-mouse game between researchers and model providers.
"Someone jailbroke the model with a fake grandma bedtime story prompt."
No comments yet — say something.
Add your own interpretation of "jailbreak".