LaplacianFormer:Rethinking Linear Attention with Laplacian Kernel
arXiv cs.CV / 4/23/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- The paper proposes LaplacianFormer, a Transformer variant that replaces softmax attention with a Laplacian kernel to better scale to high-resolution vision workloads that suffer from softmax’s quadratic complexity.
- It argues that prior linear-attention methods using Gaussian-kernel approximations lack solid theoretical justification and can suppress mid-range token interactions.
- To mitigate expressiveness loss from low-rank approximations, the method introduces a provably injective feature map that preserves fine-grained token information.
- Efficient computation is achieved via a Nyström approximation of the kernel matrix and a Newton–Schulz iteration-based solver, avoiding expensive matrix inversion and SVD.
- The authors report custom CUDA kernels for forward and backward passes and show on ImageNet that LaplacianFormer improves the performance–efficiency trade-off while enhancing attention expressiveness.
Related Articles

Trajectory Forecasts in Unknown Environments Conditioned on Grid-Based Plans
Dev.to

Elevating Austria: Google invests in its first data center in the Alps.
Google Blog

OpenAI Just Named It Workspace Agents. We Open-Sourced Our Lark Version Six Months Ago
Dev.to

GPT Image 2 Subject-Lock Editing: A Practical Guide to input_fidelity
Dev.to

AI Tutor That Works Offline — Study Anywhere with EaseLearn AI
Dev.to