FD$^2$: A Dedicated Framework for Fine-Grained Dataset Distillation

arXiv cs.CV / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses limitations of existing decoupled dataset distillation methods on fine-grained datasets, where distilled samples tend to be overly similar within a class and fail to capture subtle inter-class distinctions.
  • It proposes FD$^2$ (Fine-grained Dataset Distillation), which focuses distillation on localized discriminative regions to build fine-grained class representations.
  • In pretraining, FD$^2$ uses counterfactual attention learning to aggregate discriminative features and update class prototypes.
  • During distillation, it applies a fine-grained characteristic constraint to pull samples toward their class prototype while repelling others, and uses a similarity/diversity constraint to diversify attention among same-class samples.
  • Experiments across multiple fine-grained and general datasets indicate FD$^2$ can plug into decoupled dataset distillation pipelines and improves performance with strong transferability.

Abstract

Dataset distillation (DD) compresses a large training set into a small synthetic set, reducing storage and training cost, and has shown strong results on general benchmarks. Decoupled DD further improves efficiency by splitting the pipeline into pretraining, sample distillation, and soft-label generation. However, existing decoupled methods largely rely on coarse class-label supervision and optimize samples within each class in a nearly identical manner. On fine-grained datasets, this often yields distilled samples that (i) retain large intra-class variation with subtle inter-class differences and (ii) become overly similar within the same class, limiting localized discriminative cues and hurting recognition. To solve the above-mentioned problems, we propose FD^{2}, a dedicated framework for Fine-grained Dataset Distillation. FD^{2} localizes discriminative regions and constructs fine-grained representations for distillation. During pretraining, counterfactual attention learning aggregates discriminative representations to update class prototypes. During distillation, a fine-grained characteristic constraint aligns each sample with its class prototype while repelling others, and a similarity constraint diversifies attention across same-class samples. Experiments on multiple fine-grained and general datasets show that FD^{2} integrates seamlessly with decoupled DD and improves performance in most settings, indicating strong transferability.