Two-Stage Optimizer-Aware Online Data Selection for Large Language Models

arXiv cs.AI / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that gradient-based data selection for LLM fine-tuning, while effective offline, is ill-suited for online fine-tuning where data arrives sequentially and utility depends on the current optimization step and geometry.
  • It proposes an optimizer-aware online data selection framework that treats selection as shaping the next target-oriented parameter update under the current optimizer state.
  • The method formulates online selection as an update-matching problem connected to second-order target utility, emphasizing that the chosen subset must account for interactions and redundancy among samples.
  • To implement this for practical long-context LLMs, it introduces a two-stage Filter-then-Weight algorithm plus factorized outer-product gradient representations and optimized matrix computations.
  • Experiments indicate consistent improvements in convergence and downstream task performance compared with existing online data selection baselines under the same data budget.

Abstract

Gradient-based data selection offers a principled framework for estimating sample utility in large language model (LLM) fine-tuning, but existing methods are mostly designed for offline settings. They are therefore less suited to online fine-tuning, where data arrives sequentially, sample utility is step-dependent, and the effective update geometry is shaped by adaptive optimizers. We propose an optimizer-aware framework for gradient-based online data selection and reweighting in LLM fine-tuning. Our key idea is to view online selection not as static sample ranking, but as shaping the next target-oriented update under the optimizer state. We formulate this as an optimizer-aware update-matching problem, establish its connection to second-order target utility, and show why subset-level construction must account for interactions and redundancy among selected samples. Based on this view, we develop a two-stage Filter-then-Weight algorithm that first filters geometrically useful candidates and then optimizes their coefficients. To make the framework practical for LLMs, we introduce a factorized outer-product gradient representation and optimized matrix computations for long-context data. Experiments show that our method consistently improves convergence and downstream performance over existing online data selection baselines under the same data budget.