Selective Rotary Position Embedding
arXiv cs.CL / 4/27/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes Selective RoPE, an input-dependent rotary position embedding that generalizes standard RoPE and allows rotation by arbitrary angles in both softmax and linear transformer settings.
- It argues that softmax attention implicitly performs a hidden rotational operation on query-key pairs, revealing an underlying positional structure.
- The authors connect positional encoding behavior to state-space and gated linear transformer mechanisms, suggesting the real part relates to forgetting while the imaginary part encodes positions via rotations.
- Experiments show that adding Selective RoPE to gated transformers improves language modeling performance and helps on challenging sequence tasks such as copying, state tracking, and retrieval.
Related Articles

Subagents: The Building Block of Agentic AI
Dev.to

DeepSeek-V4 Models Could Change Global AI Race
AI Business

Got OpenAI's privacy filter model running on-device via ExecuTorch
Reddit r/LocalLLaMA

The Agent-Skill Illusion: Why Prompt-Based Control Fails in Multi-Agent Business Consulting Systems
Dev.to

We Built a Voice AI Receptionist in 8 Weeks — Every Decision We Made and Why
Dev.to