What are the definitions for prompt injections, jailbreaks, prompt hacking and prompt leaking?
David McCarthy
•
Last updated on May 15, 2025 at 6:00 PM
Below are concise meanings of several common terms you'll run into when working on advanced prompts:
- Prompt injection – Inserting crafted text into the input or context to override or influence the model's original instructions.
- Jailbreak – A specialised prompt injection designed to bypass content filters or system policies so the model produces otherwise restricted output.
- Prompt hacking – The broader practice of analysing, manipulating, or optimising prompts (including benign optimisation and adversarial attacks).
- Prompt leaking – Forcing the model to reveal hidden developer or system prompts that were not meant to be disclosed.