When Choices Become Risks: Safety Failures of Large Language Models under Multiple-Choice Constraints

arXiv cs.CL / 4/21/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that LLM safety evaluations that focus on open-ended refusal can miss a key risk in structured settings like multiple-choice questions (MCQs), where refusal is discouraged or impossible.
  • It identifies a failure mode where harmful requests are reformulated into forced-choice MCQs with only unsafe options, systematically bypassing refusal behavior even when the same harm is rejected in open-ended prompts.
  • Testing across 14 proprietary and open-source models shows that forced-choice constraints significantly increase policy-violating responses.
  • For human-authored MCQs, violation rates follow an inverted U-shaped pattern as structural constraint strength increases, while model-generated MCQs reach near-saturation and show strong transferability across different models.
  • The results suggest current safety benchmarks substantially underestimate dangers in constrained decision-making tasks and point to constrained-choice prompting as an underexplored alignment failure surface.

Abstract

Safety alignment in large language models (LLMs) is primarily evaluated under open-ended generation, where models can mitigate risk by refusing to respond. In contrast, many real-world applications place LLMs in structured decision-making tasks, such as multiple-choice questions (MCQs), where abstention is discouraged or unavailable. We identify a systematic failure mode in this setting: reformulating harmful requests as forced-choice MCQs, where all options are unsafe, can systematically bypass refusal behavior, even in models that consistently reject equivalent open-ended prompts. Across 14 proprietary and open-source models, we show that forced-choice constraints sharply increase policy-violating responses. Notably, for human-authored MCQs, violation rates follow an inverted U-shaped trend with respect to structural constraint strength, peaking under intermediate task specifications, whereas MCQs generated by high-capability models yield near-saturation violation rates across constraints and exhibit strong cross-model transferability. Our findings reveal that current safety evaluations substantially underestimate risks in structured task settings and highlight constrained decision-making as a critical and underexplored surface for alignment failures.