Beyond Loss Values: Robust Dynamic Pruning via Loss Trajectory Alignment

arXiv cs.CV / 4/9/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Existing dynamic data pruning methods often rank samples by per-sample loss, which can misidentify noisy-label examples (high-loss noisy samples) as valuable and hurt model accuracy.
  • The paper introduces AlignPrune, a plug-and-play module that improves pruning under label noise by using a loss-trajectory-based criterion called the Dynamic Alignment Score (DAS).
  • AlignPrune targets more reliable identification of noisy samples by aligning with how losses evolve over training rather than relying on single-point loss values.
  • Experiments across five benchmarks, multiple noise types, and pruning ratios show consistent gains, with accuracy improvements up to 6.3% over state-of-the-art dynamic pruning baselines.
  • The authors report that AlignPrune integrates into existing pruning frameworks without changing model architecture or the training pipeline, and provide code for adoption and further research.

Abstract

Existing dynamic data pruning methods often fail under noisy-label settings, as they typically rely on per-sample loss as the ranking criterion. This could mistakenly lead to preserving noisy samples due to their high loss values, resulting in significant performance drop. To address this, we propose AlignPrune, a noise-robust module designed to enhance the reliability of dynamic pruning under label noise. Specifically, AlignPrune introduces the Dynamic Alignment Score (DAS), which is a loss-trajectory-based criterion that enables more accurate identification of noisy samples, thereby improving pruning effectiveness. As a simple yet effective plug-and-play module, AlignPrune can be seamlessly integrated into state-of-the-art dynamic pruning frameworks, consistently outperforming them without modifying either the model architecture or the training pipeline. Extensive experiments on five widely-used benchmarks across various noise types and pruning ratios demonstrate the effectiveness of AlignPrune, boosting accuracy by up to 6.3\% over state-of-the-art baselines. Our results offer a generalizable solution for pruning under noisy data, encouraging further exploration of learning in real-world scenarios. Code is available at: https://github.com/leonqin430/AlignPrune.