Pack only the essentials: Adaptive dictionary learning for kernel ridge regression
arXiv cs.LG / 4/27/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- Kernel ridge regression (KRR) is limited by the O(n^2) space required to store and manipulate the full kernel matrix, making it impractical for large datasets.
- While Nystrom approximations using uniform sampling reduce storage to O(nm), they can require m≈O(n) on datasets where the kernel has high coherence.
- Sampling columns via ridge leverage scores (RLS) can achieve accurate Nystrom approximations with m scaling to the effective dimension, but computing exact RLS still costs O(n^2) space.
- The paper proposes SQUEAK, an algorithm extending INK-Estimate that uses unnormalized RLS to simplify the procedure and avoid estimating the effective dimension for normalization, while keeping space complexity close to exact RLS sampling (up to a constant factor).




