Optimized Architectures for Kolmogorov-Arnold Networks
arXiv stat.ML / 4/22/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes architectural strategies to improve Kolmogorov-Arnold networks (KANs) while preserving their interpretability, which previous enhancements often compromised due to added complexity.
- It studies an approach that combines overprovisioned architectures with sparsification, deep supervision, and depth selection to produce compact and interpretable KANs without accuracy loss.
- The method uses differentiable mechanisms optimized end-to-end under a minimum description length (MDL) objective, jointly learning activations, structure, and depth.
- Experiments across multiple settings—including function approximation, dynamical systems forecasting, and real-world prediction—show sparsification alone is not enough, but adding depth selection yields competitive or better accuracy with much smaller models.


