LLM jailbreak evaluation benchmark. Includes a dataset of prompts from categories of harmful behavior:
- Disinformation and deception
- Illegal goods and services
- Hate/harassment/discrimination
- Non-violent crime
- Violence
- Sexual content https://strong-reject.readthedocs.io/en/latest/