AA-SVD : Anchored and Adaptive SVD for Large Language Model Compression
arXiv cs.LG / 4/3/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes AA-SVD, a fast low-rank SVD-based framework that compresses billion-parameter LLMs without requiring retraining.
- It addresses error propagation caused by distribution shifts during layer-by-layer compression by explicitly modeling both upstream input shifts and anchoring to original layer outputs.
- Beyond compressing individual layers, AA-SVD refines each Transformer block end-to-end to reduce block-level output distortion and enable joint compensation for accumulated errors.
- Experiments show AA-SVD outperforms prior SVD-style baselines across a range of compression ratios, with especially large gains under aggressive compression budgets where other methods substantially degrade or fail.




