Expert Selections In MoE Models Reveal (Almost) As Much As Text
arXiv cs.CL / 3/16/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- A text-reconstruction attack on mixture-of-experts (MoE) language models shows that tokens can be recovered using only expert routing selections.
- Contrasting with prior logistic-regression approaches, a 3-layer MLP improves top-1 reconstruction to 63.1%, and a transformer-based sequence decoder achieves 91.2% top-1 (94.8% top-10) on 32-token sequences from OpenWebText after training on 100 million tokens.
- The results connect MoE routing information to the broader embedding-inversion literature and highlight practical leakage scenarios such as distributed inference and side channels.
- Adding noise reduces reconstruction but does not eliminate it, underscoring the need to treat expert selections as sensitive as the underlying text.
Related Articles

Astral to Join OpenAI
Dev.to

PearlOS. We gave swarm intelligence a local desktop environment and code control to self-evolve. Has been pretty incredible to see so far. Open source and free if you want your own.
Reddit r/LocalLLaMA

Why Data is Important for LLM
Dev.to

The Inference Market Is Consolidating. Agent Payments Are Still Nobody's Problem.
Dev.to

YouTube's Deepfake Shield for Politicians Changes Evidence Forever
Dev.to