Screening Is Enough

arXiv cs.LG / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that conventional softmax attention lacks an absolute notion of query–key relevance because attention only redistributes fixed mass across competing keys, making it hard to explicitly reject irrelevant keys.
  • It introduces Multiscreen, a language-model architecture using a “screening” mechanism that thresholds each key for explicit acceptance/rejection and aggregates only the remaining keys, removing global competition.
  • Experiments reported that Multiscreen matches Transformer validation loss while using about 40% fewer parameters and supports stable training at much higher learning rates.
  • The method is reported to preserve strong long-context perplexity performance, show minimal retrieval degradation even beyond training context length, and cut inference latency by up to 3.2× at 100K context length.

Abstract

A core limitation of standard softmax attention is that it does not define a notion of absolute query--key relevance: attention weights are obtained by redistributing a fixed unit mass across all keys according to their relative scores. As a result, relevance is defined only relative to competing keys, and irrelevant keys cannot be explicitly rejected. We introduce Multiscreen, a language-model architecture built around a mechanism we call screening, which enables absolute query--key relevance. Instead of redistributing attention across all keys, screening evaluates each key against an explicit threshold, discarding irrelevant keys and aggregating the remaining keys, thereby removing global competition among keys. Across experiments, Multiscreen achieves comparable validation loss with approximately 40% fewer parameters than a Transformer baseline, enables stable optimization at substantially larger learning rates, maintains strong performance in long-context perplexity, shows little to no degradation in retrieval performance even far beyond the training context length, and reduces inference latency by up to 3.2\times at 100K context length.