AI Navigate

LLM Routing as Reasoning: A MaxSAT View

arXiv cs.AI / 3/17/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a constraint-based interpretation of language-conditioned LLM routing by formulating it as a weighted MaxSAT/MaxSMT problem in which natural language feedback induces hard and soft constraints over model attributes.
  • Under this formulation, routing corresponds to selecting models that approximately maximize satisfaction of feedback-conditioned clauses.
  • Empirical analysis on a benchmark of 25 models shows that language feedback yields near-feasible recommendation sets, and that no-feedback scenarios reveal systematic priors.
  • The work suggests that LLM routing can be understood as structured constraint optimization driven by language-conditioned preferences.
  • The study provides a theoretical and empirical framework linking natural language preferences to model-selection decisions, informing future routing system design.

Abstract

Routing a query through an appropriate LLM is challenging, particularly when user preferences are expressed in natural language and model attributes are only partially observable. We propose a constraint-based interpretation of language-conditioned LLM routing, formulating it as a weighted MaxSAT/MaxSMT problem in which natural language feedback induces hard and soft constraints over model attributes. Under this view, routing corresponds to selecting models that approximately maximize satisfaction of feedback-conditioned clauses. Empirical analysis on a 25-model benchmark shows that language feedback produces near-feasible recommendation sets, while no-feedback scenarios reveal systematic priors. Our results suggest that LLM routing can be understood as structured constraint optimization under language-conditioned preferences.