Position-Agnostic Pre-Projection for Transformer Attention: Nonlinear Feature Construction and Content Skip Before Q/K/V

arXiv cs.CL / 4/14/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes two transformer-attention modifications: a non-linear, position-agnostic pre-projection MLP before Q/K/V computation and a content skip pathway that can bypass the attention mechanism when helpful.
  • The pre-projection is applied after layer normalization and before positional encoding, aiming to construct richer features without injecting positional information too early.
  • Experiments using frozen probes on Pythia-160M and 410M show the combined method delivers the strongest gains, including +40.6% LAMBADA accuracy and -39% perplexity at the 160M scale.
  • Learned skip-connection behavior indicates later transformer layers rely more on the content bypass than earlier layers, suggesting deeper layers benefit from content information that avoids position-aware attention.
  • The authors report that the changes add no K/V cache overhead, which can help preserve inference efficiency.

Abstract

We propose two complementary modifications to transformer attention blocks. First, a non-linear pre-projection MLP is inserted between layer norm and Q/K/V projections, constructing richer features in a position-agnostic manner before any positional encoding is applied. Second, a content skip connection routes the pre-projection's features around the attention mechanism, allowing content information to bypass position-aware attention where beneficial. In frozen-probe experiments on Pythia-160M and 410M, the combined approach achieves the strongest results across methods: +40.6% LAMBADA accuracy and -39% perplexity at 160M scale. Learned skip connection weights reveal a consistent pattern across model sizes: later transformer layers activate the content bypass more strongly than earlier layers, suggesting that deeper layers benefit from content information that does not pass through positional attention. All modifications add no K/V cache overhead.