Neural Network Pruning via QUBO Optimization
arXiv cs.CV / 4/8/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper frames neural network pruning as a combinatorial optimization problem and argues that prior QUBO approaches underperformed due to oversimplified objectives (e.g., L1-norm) that miss interactions among filters.
- It introduces a unified Hybrid QUBO formulation that uses gradient-aware sensitivity (first-order Taylor and second-order Fisher information) for the QUBO linear term and data-driven activation similarity for the quadratic term to capture both individual relevance and redundancy.
- To ensure the desired sparsity level, the method adds a dynamic capacity-driven search that enforces target sparsity constraints without reshaping the optimization landscape.
- The approach uses a two-stage pipeline where a Tensor-Train (TT) Refinement step (gradient-free) fine-tunes the QUBO solution against the true evaluation metric.
- Experiments on the SIDD image denoising dataset show the Hybrid QUBO outperforms greedy Taylor pruning and traditional L1-based QUBO, with additional gains from TT Refinement at suitable pruning scales.
Related Articles
[N] Just found out that Milla Jovovich is a dev, invested in AI, and just open sourced a project
Reddit r/MachineLearning

ALTK‑Evolve: On‑the‑Job Learning for AI Agents
Hugging Face Blog

Context Windows Are Getting Absurd — And That's a Good Thing
Dev.to
Google isn’t an AI-first company despite Gemini being great
Reddit r/artificial

GitHub Weekly: Copilot SDK Goes Public, Cloud Agent Breaks Free
Dev.to