Supplement Generation Training for Enhancing Agentic Task Performance

arXiv cs.LG / 4/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Agentic task training with large foundation models is becoming impractical due to high compute demands, slow iteration, and frequent model obsolescence.
  • The proposed Supplement Generation Training (SGT) approach trains a smaller LLM to create supplemental text that is appended to the original input to improve the larger LLM’s task performance.
  • SGT aims to adapt supplements dynamically to task requirements without changing or fine-tuning the underlying large model.
  • By separating task-specific optimization from large foundation models, SGT targets a more flexible and cost-effective way to deploy LLM-powered agents in real-world settings.

Abstract

Training large foundation models for agentic tasks is increasingly impractical due to the high computational costs, long iteration cycles, and rapid obsolescence as new models are continuously released. Instead of post-training massive models for every new task or domain, we propose Supplement Generation Training (SGT), a more efficient and sustainable strategy. SGT trains a smaller LLM to generate useful supplemental text that, when appended to the original input, helps the larger LLM solve the task more effectively. These lightweight models can dynamically adapt supplements to task requirements, improving performance without modifying the underlying large models. This approach decouples task-specific optimization from large foundation models and enables more flexible, cost-effective deployment of LLM-powered agents in real-world applications.