FourierMoE: Fourier Mixture-of-Experts Adaptation of Large Language Models
arXiv cs.LG / 4/3/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- FourierMoE proposes a new parameter-efficient fine-tuning approach for LLMs by moving mixture-of-experts adaptation into the spectral (frequency) domain instead of the spatial domain used by prior MoE-PEFT methods.
- The method is motivated by findings that different tasks have distinct frequency energy distributions and that LLM layers have heterogeneous sensitivities across frequencies.
- FourierMoE uses a frequency-adaptive router to send tokens to experts specialized in different frequency bands, with experts learning conjugate-symmetric complex coefficients to preserve full phase and amplitude information.
- The authors argue FourierMoE supports theoretically lossless reconstruction of real-valued spatial weights via inverse discrete Fourier transform (IDFT), maintaining representational fidelity during adaptation.
- Experiments across 28 benchmarks, multiple model architectures, and different scales show consistent gains over baselines in both single-task and multi-task fine-tuning while requiring significantly fewer trainable parameters.
Related Articles

Black Hat Asia
AI Business

Mistral raises $830M, 9fin hits unicorn status, and new Tech.eu Summit speakers unveiled
Tech.eu

ChatGPT costs $20/month. I built an alternative for $2.99.
Dev.to

OpenAI shifts to usage-based pricing for Codex in ChatGPT business plans
THE DECODER

Why I built an AI assistant that doesn't know who you are
Dev.to