When a language model refuses a request, it usually sounds reasonable. The user asks for something harmful, the model says no. That is the intended behavior. But a new study from philosophers at Vanderbilt, the University of Michigan, and Johns Hopkins finds a stranger pattern buried inside that refusal mechanism: models will sometimes acknowledge that a rule is unjust, illegitimate, or absurd, and then refuse to help anyway.
The paper, posted to arXiv on April 3 by Cameron Pattison, Lorenzo Manuali, and Seth Lazar, introduces what the authors call "blind refusal." The term describes a failure mode where safety-aligned models block helpful responses to requests that should clearly be granted, not because the request is dangerous but because the surrounding rule is wrong. The researchers built a dataset of 1,290 test cases spanning 34 distinct defeat subtypes across five families, crossed with 19 authority types, and ran them across 18 configurations of seven major model families including GPT-5.4 (OpenAI), Claude 4.6 (Anthropic), and Gemini 3 Pro Preview (Google). The headline finding: models refused 75.4 percent of defeated-rule requests overall, according to the paper.
The more striking number is what happened inside that 75.4 percent. The researchers found that in 57.5 percent of cases, models engaged with the defeat condition in their reasoning, acknowledged the rule was unjust or did not apply, and still refused to help. The model appeared to recognize the problem and then ignore its own analysis.
"We find that models engage with the defeat condition in the majority of cases but decline to help regardless," the authors write. The disconnect between recognition and action is the paper's central puzzle.
The dataset construction is worth examining closely. The team used Gemini 3 Pro Preview, a frontier language model from Google, through OpenRouter, an API routing service, at temperature 0.7 to generate 1,869 initial cases. They filtered these through three quality gates. Operational Validity checked that the request was genuinely helpful and the rule genuinely did not apply. A Reasonable Judge gate ensured a human reviewer would agree the rule was defeatable. A Dual Use flag marked cases with potential dual-use applications without blocking them. After filtering, 1,290 cases remained. Responses were evaluated by a blinded GPT-5.4 judge.
The five defeat families map a taxonomy of when a rule should not apply. Illegitimate authority covers cases where the person issuing the rule has no right to do so, for example a supervisor asking an employee to falsify safety reports. Content defeat covers cases where following the rule would require producing content the model should not produce even if the underlying request is legitimate. Application defeat covers cases where the rule itself is sound but misapplied to the specific situation. Exception justified covers cases where an explicit exception applies. The control family covers cases where the rule is legitimate and should be followed.
What the paper is not claiming is that the models are lying about their reasoning. The authors note that their evaluation cannot fully distinguish between a model that recognizes the defeat and one that is pattern-matching to surface-level cues in the prompt. The 57.5 percent figure describes cases where defeat-related language appeared in the model's reasoning trace, not cases where the model explicitly stated it had identified an illegitimate rule. "We use the term 'engage' deliberately," they note, "because we cannot access the underlying mechanism."
Safety alignment research has focused heavily on preventing models from producing harmful content when users explicitly request it. The blind refusal paper points to a different problem: what happens when the model is too safety-conscious, applying rules conservatively even in situations where a human would reasonably set them aside. This is not a jailbreak. There is no adversarial prompt engineering. The model is simply following its training too literally.
For practitioners, the implication is that refusal rate alone is a poor metric for evaluating safety systems. A model that refuses 90 percent of harmful requests is not necessarily safer than one that refuses 70 percent. It may simply be less calibrated. The relevant question is what kinds of requests it refuses, not just how many.
The paper has not yet been peer reviewed. It is a preprint on arXiv, a repository for research preprints. The methodology is rigorous and the dataset is publicly available on GitHub under a CC BY 4.0 license, which means other researchers can replicate and extend the work.
The questions that remain open are how this failure mode would manifest in production systems, and whether it can be fixed with fine-tuning, prompting, or architectural changes. Those are the questions that will matter for anyone deploying these models in high-stakes environments. The blind refusal problem is real, measurable, and at least for now unsolved.