Cost-Aware Learning

arXiv cs.LG / 5/1/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies Cost-Aware Learning for finite-sum optimization, aiming to reach a target error while minimizing total cost when different components have different sampling costs.
  • It introduces the Cost-Aware Stochastic Gradient Descent (SGD) algorithm for convex objectives, along with derived cost complexity bounds to achieve error ε.
  • The authors prove a lower bound for the problem and propose a subset selection method to further reduce training cost.
  • They apply the framework to reinforcement learning with language models, where policy-gradient cost depends on sequence length, presenting Cost-Aware GRPO to cut policy-optimization cost without hurting performance.
  • Experiments on 1.5B and 8B LLMs show up to ~30% fewer tokens used for policy optimization while achieving matching or better accuracy versus baselines.

Abstract

We consider the problem of Cost-Aware Learning, where sampling different component functions of a finite-sum objective incurs different costs. The objective is to reach a target error while minimizing the total cost. First, we propose the Cost-Aware Stochastic Gradient Descent algorithm for convex functions, and derive its cost complexity to attain an error of \epsilon. Furthermore, we establish a lower bound for this setting and provide a subset selection algorithm to further reduce the cost of training. We apply our theoretical insights to reinforcement learning with language models, where the computational cost of policy gradients varies with sequence length. To this end, we introduce Cost-Aware GRPO, an algorithm designed to reduce the cost of policy optimization while preserving performance. Empirical results on 1.5B and 8B LLMs demonstrate that our approach reduces the tokens used in policy optimization by up to about 30% while matching or exceeding baseline accuracy.