SepSeq: A Training-Free Framework for Long Numerical Sequence Processing in LLMs

arXiv cs.CL / 4/10/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that LLMs’ performance degrades on long numerical sequences due to attention dispersion caused by the Softmax attention mechanism preventing focus on relevant positions.
  • It introduces SepSeq, a training-free and plug-and-play framework that inserts separator tokens to create an “attention sink,” helping the model concentrate on local segments while still retaining global context.
  • Experiments across nine popular LLMs show an average relative accuracy improvement of 35.6% across diverse domains for long numerical sequence processing.
  • The method also reduces total inference token consumption by 16.4% on average, indicating efficiency gains alongside accuracy improvements.

Abstract

While transformer-based Large Language Models (LLMs) theoretically support massive context windows, they suffer from severe performance degradation when processing long numerical sequences. We attribute this failure to the attention dispersion in the Softmax mechanism, which prevents the model from concentrating attention. To overcome this, we propose Separate Sequence (SepSeq), a training-free, plug-and-play framework to mitigate dispersion by strategically inserting separator tokens. Mechanistically, we demonstrate that separator tokens act as an attention sink, recalibrating attention to focus on local segments while preserving global context. Extensive evaluations on 9 widely-adopted LLMs confirm the effectiveness of our approach: SepSeq yields an average relative accuracy improvement of 35.6% across diverse domains while reducing total inference token consumption by 16.4% on average.