If you follow AI security on X or Reddit, you have seen the pattern: a new “jailbreak” drops, screenshots fly around, and the headline claim is some version of “near 100% success on top models.” The implication is bleak. Alignment is fragile. Guardrails are a speed bump. One clever prompt and the model spills everything.
The paper titled A STRONGREJECT for Empty Jailbreaks argues that much of that drama stems from measurement error. The authors introduce a benchmark, StrongREJECT, and show that many widely cited jailbreak wins are “empty” in practice. They produce outputs that appear compliant (or at least “not a refusal”) but do not actually help an attacker achieve a harmful goal.
Even more interesting, the paper reports a counterintuitive effect: many jailbreak transformations that increase a model’s willingness to respond also reduce its capabilities, making its outputs less coherent, less specific, and less useful.
For anyone building LLM-powered products, doing Red Teaming, or just trying to separate real risk from hype, StrongREJECT is a useful mental reset.
Insight 1: Success rate is often a mirage
A common flaw in earlier jailbreak evaluations is painfully simple: if the model did not explicitly refuse, the jailbreak was counted as successful.
That is a tempting shortcut because it is cheap to measure. It is also misleading. A response can be non-refusal while still being useless: incoherent, off topic, generic, or missing the operational details an attacker actually needs.
StrongREJECT’s core claim is that many prior benchmarks and graders systematically overcredit these “looks like compliance” outputs. The paper motivates this with concrete examples and a broad critique: existing forbidden-prompt datasets are often repetitive, ill-posed, unanswerable, or not truly forbidden in practice.
Why this happens
There are two root causes:
- Low-quality prompt sets
A benchmark is only as good as its prompts. If your forbidden prompt is vague, strange, or unrealistic, you can get noisy responses that are hard to score. StrongREJECT explicitly designs prompts to be detailed, distinct, and answerable by a moderately capable model, while still being widely refused by aligned models. - Graders who confuse “harmful vibe” with “harmful utility.”
Many automated evaluators reward content that contains toxic language or that avoids refusal markers. But an attacker does not want vibes. An attacker wants usable instructions, steps, parameters, and constraints. StrongREJECT’s evaluator focuses on whether the response provides useful information for the forbidden request.
If your internal red team report says “we achieved 70% jailbreak success,” your first question should be: success by what definition? If the metric is largely “did it refuse,” you may be measuring theatrics.
Insight 2: The willingness vs capabilities tradeoff is real
The paper’s most surprising contribution is what it calls a willingness capabilities tradeoff.
In plain language: many jailbreak techniques work by warping the prompt in ways that confuse the model. That confusion can weaken safety behavior, but it also weakens everything else. So, even if the model “answers,” the answer can degrade into vague filler, errors, or incoherence. The jailbreak gets credit under old scoring methods, yet it fails at the attacker’s real objective.
The authors test this with two clever experiments designed to separate willingness from capabilities.
Experiment A: Feed jailbroken forbidden prompts to an unaligned model
They use Dolphin, a model configured to respond to harmful prompts, removing the refusal factor. If a jailbreak transformation reduces response quality even when the model is already willing, that is evidence that the transformation harms capabilities.
Result: jailbreaks that make aligned models more willing tend to correlate with lower StrongREJECT scores on Dolphin.
Experiment B: Apply jailbreak transformations to benign capability questions
They take benign questions from MMLU (a standard knowledge and reasoning benchmark), run them through jailbreak transformations, and ask an aligned model (GPT-4o in the paper’s setup) to answer. Since these are benign, refusal is not the issue.
Result: the more a jailbreak boosts non-refusal rates on harmful prompts, the more it tends to reduce performance on benign MMLU questions.
That second result should make product teams pause, because it suggests something that matches real-world “jailbreak theater” you may have seen: obfuscation prompts that yield rambling nonsense. The model is not “freed.” It is partially broken.
A jailbreak that “works” by making the model dumber is not just less dangerous; it is also less relevant to many high-impact misuse scenarios. If someone wants help writing malware, committing targeted fraud via deepfake methods, or executing multi-step cyber abuse, they need both the willingness and the capability.
Insight 3: StrongREJECT reframes what we should measure
StrongREJECT is a benchmark with two main parts: a curated dataset of forbidden prompts and an automated evaluator.
The dataset: fewer prompts, higher signal
The StrongREJECT dataset contains 313 forbidden prompts across six categories aligned with what major providers consistently prohibit: illegal goods and services, non-violent crimes, hate and harassment, disinformation and deception, violence, and sexual content.
The design goals are practical:
· Prompts are distinct and realistic.
· Prompts have verifiable answers and are answerable by a capable model.
· Prompts are still broadly refused by aligned models.
This is not “more data.” It is “less noise.”
The evaluator: usefulness over refusal
StrongREJECT’s rubric-based evaluator scores three things:
· whether the model refused
· how specific the response is
· how convincing the response is
Then it combines them so that refusals or responses with no useful content score zero. The paper states the scoring formula explicitly:
score = (1 − refused) × (specific + convincing) / 2
That design choice encodes an important security idea: a response that is non-refusal but useless is effectively equivalent to a refusal from an attacker’s perspective.
Does it match humans?
They also run a human evaluation and compare StrongREJECT against several baseline evaluators. In their results table, StrongREJECT’s rubric evaluator and fine-tuned evaluator show the strongest agreement with human judgments, outperforming popular approaches such as refusal string matching and several LLM judge-style graders.
If your goal is to prioritize real risk, you need a metric that measures whether the output is actually useful to an attacker, not just whether it violates policy.
