Importance-Guided Basis Selection for Low-Rank Decomposition of Large Language Models
arXiv cs.LG / 5/5/2026
📰 NewsModels & Research
Key Points
- Low-rank decomposition can compress large language models, but performance depends critically on choosing which singular-vector bases to keep for a given task.
- The article argues that prior heuristics (e.g., pruning based on small re-learned magnitudes or adapting coefficients) can be misaligned with task loss because they ignore the local geometry of the loss landscape.
- It introduces Basis Selection with Importance (BSI), which ranks and prunes bases by estimating the expected increase in task loss when each basis is removed, using a second-order Taylor expansion that blends sensitivity and curvature.
- To apply this efficiently to LLMs, BSI uses an adapted Hutchinson-style randomized probing method to estimate Hessian-diagonal information via symmetric parameter perturbations.
- Experiments on mathematical reasoning benchmarks show BSI outperforming existing low-rank decomposition baselines, with the biggest gains under deep compression settings, supported by theoretical bounds and sample-complexity guarantees.
Related Articles

Why Retail Chargeback Recovery Could Be AgentHansa's First Real PMF
Dev.to

Last Week in AI #340 - OpenAI vs Musk + Microsoft, DeepSeek v4, Vision Banana
Last Week in AI

Trying to train tiny LLMs on length constrained reddit posts summarization task using GRPO on 3xMac Minis - updates!
Reddit r/LocalLLaMA

Uber Shares What Happens When 1.500 AI Agents Hit Production
Reddit r/artificial
vibevoice.cpp: Microsoft VibeVoice (TTS + long-form ASR with diarization) ported to ggml/C++, runs on CPU/CUDA/Metal/Vulkan, no Python at inference
Reddit r/LocalLLaMA