AutoSP: Unlocking Long-Context LLM Training Via Compiler-Based Sequence Parallelism

arXiv cs.LG / 5/1/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • The paper introduces AutoSP, an automated compiler-based approach to optimize long-context LLM training without requiring developers to manually redesign training pipelines.
  • AutoSP applies a targeted combination of techniques, including automated sequence parallelism and long-context-aware activation checkpointing, to improve trainability.
  • The authors argue that existing training libraries mainly optimize for parameter scaling (e.g., ZeRO-3/FSDP, tensor/pipeline parallelism) and lack straightforward abstractions for long-context-specific optimizations.
  • Experiments on both NVIDIA and AMD hardware show context lengths increased by up to 2.7× and 2.5×, respectively, compared with competitive hand-written baselines, while adding negligible runtime overhead.
  • AutoSP is positioned as a productivity boost by reducing the need for deep expertise to combine multiple complex long-context optimization strategies.

Abstract

Large-language-models (LLMs) demonstrate enormous utility in long-context tasks which require processing prompts that consist of tens to hundreds of thousands of tokens. However, existing LLM training libraries do not provide easy to use abstractions to optimize for long-context training, instead focusing on optimizations for models with large parameter counts through ZeRO-3/FSDP, Tensor and Pipeline parallelism. This forces users to rewrite LLM training libraries to incorporate compositions of various complex long-context optimizations, such as sequence-parallelism, to training pipelines; a process that requires in-depth expertise, reducing developer productivity. To tackle these challenges, we introduce AutoSP: the first automated solution to automatically optimize LLM training for longer-contexts. AutoSP compiles models and applies a targeted set of optimizations: automated sequence parallelism, and long-context aware activation-checkpointing, to drastically enhance LLM trainability at negligible cost to throughput. Our evaluation demonstrates AutoSP's capability on both NVIDIA and AMD hardware, increasing training contexts by upto 2.7\times and 2.5\times respectively over competitive hand-written baseline at negligible cost to runtime performance.