Schema Key Wording as an Instruction Channel in Structured Generation under Constrained Decoding

arXiv cs.CL / 4/17/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Constrained decoding for LLM structured generation typically enforces formats like JSON/XML as structural constraints, but this paper shows that how schemas are linguistically worded can also change model behavior.
  • The authors demonstrate that changing only the wording of schema keys—without altering the prompt or model parameters—can significantly affect performance under constrained decoding.
  • They propose viewing structured generation as a multi-channel instruction problem, where prompts provide explicit instructions while schema keys provide implicit instruction signals during decoding.
  • Experiments on mathematical reasoning benchmarks find that different model families respond differently: Qwen benefits more from schema-level instructions, while LLaMA depends more on prompt-level guidance.
  • The study also finds non-additive effects between instruction channels, meaning combining prompt and schema channels does not always yield further gains.

Abstract

Constrained decoding has been widely adopted for structured generation with large language models (LLMs), ensuring that outputs satisfy predefined formats such as JSON and XML. However, existing approaches largely treat schemas as purely structural constraints and overlook the possibility that their linguistic formulation may affect model behavior. In this work, we study how instruction placement influences model performance in structured generation and show that merely changing the wording of schema keys, without modifying the prompt or model parameters, can significantly alter model performance under constrained decoding. Based on this observation, we propose to reinterpret structured generation as a multi-channel instruction problem, where instructions can be conveyed explicitly through prompts and implicitly through schema keys during decoding. To the best of our knowledge, this is the first work to systematically study how schema key formulation acts as an implicit instruction channel and affects model performance under constrained decoding. Experiments on multiple mathematical reasoning benchmarks show that different model families exhibit distinct sensitivities to these instruction channels: Qwen models consistently benefit from schema-level instructions, while LLaMA models rely more heavily on prompt-level guidance. We further observe non-additive interaction effects between instruction channels, showing that combining multiple channels does not always lead to further improvement. These findings suggest that schema design not only determines output structure, but also carries instruction signals, offering a new perspective on structured generation in LLMs.