« Back to Glossary Index

A jailbreak refers to techniques used to override the AI’s built-in safety features and restrictions.

These safety features are designed to prevent the AI from generating harmful, offensive, or restricted content. By using specific methods or prompts, people can exploit weaknesses in the AI’s programming to make it produce responses it shouldn’t, such as inappropriate or dangerous information.


Reinforcement Learning from Human Feedback (RLHF) can play a significant role in preventing jailbreaks in AI systems by aligning the AI’s behavior with human values and safety requirements. 

Some jailbreak methods:

  • DAN (Do Anything Now) Prompt: This is a type of prompt designed to trick the AI into bypassing its built-in content filters. By instructing the AI to “do anything now,” users attempt to get the model to generate content that it would normally refuse to produce under standard operating parameters.
  • Developer Mode: This method involves convincing the AI that it is in a special mode where it can provide information and perform tasks that are normally restricted. By simulating a developer environment, users can bypass regular content filters and extract otherwise restricted information.
  • AIM Mode: Similar to Developer Mode, AIM (Artificial Intelligence Mode) is another exploit where users manipulate the AI into a state where it disregards its usual ethical guidelines and restrictions. In this mode, the AI might be prompted to provide content that is usually filtered out.
« Back to Glossary Index