Learning Retrieval Models with Sparse Autoencoders
arXiv cs.LG / 3/17/2026
📰 NewsModels & Research
Key Points
- The paper introduces SPLARE, a method to train SAE-based learned sparse retrieval (LSR) models that encode queries and documents into high-dimensional sparse representations rather than projecting into the vocabulary space.
- Sparse autoencoders are used to decompose dense LLM representations into interpretable latent features, enabling more semantically structured and language-agnostic retrieval signals.
- Empirical results show SPLARE-based LSR consistently outperforms vocabulary-based LSR in multilingual and out-of-domain settings, with SPLARE-7B achieving top results on MMTEB multilingual and English retrieval tasks.
- A lighter 2B-parameter variant demonstrates a smaller footprint while preserving retrieval performance, highlighting practical scalability of the approach.
Related Articles

PearlOS. We gave swarm intelligence a local desktop environment and code control to self-evolve. Has been pretty incredible to see so far. Open source and free if you want your own.
Reddit r/LocalLLaMA
QwenDean-4B | fine-tuned SLM for UIGen; our first attempt, looking for feedback!
Reddit r/LocalLLaMA
acestep.cpp: portable C++17 implementation of ACE-Step 1.5 music generation using GGML. Runs on CPU, CUDA, ROCm, Metal, Vulkan
Reddit r/LocalLLaMA
**Introducing SPEED-Bench: A Unified and Diverse Benchmark for Speculative Decoding**
Hugging Face Blog

Newest GPU server in the lab! 72gb ampere vram!
Reddit r/LocalLLaMA