Length Value Model: Scalable Value Pretraining for Token-Level Length Modeling
arXiv cs.CL / 5/1/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces the Length Value Model (LenVM), a token-level approach for modeling remaining generation length in autoregressive LLMs, addressing the lack of fine-grained length supervision in prior work.
- LenVM casts length modeling as a value estimation problem using a constant negative reward per generated token, producing a bounded, discounted return that acts as a monotonic proxy for the remaining generation horizon.
- The method provides dense, annotation-free, and scalable supervision, and experiments show LenVM delivers strong inference-time signals across both LLMs and VLMs.
- On the LIFEBench exact length-matching task, applying LenVM to a 7B model boosts length score from 30.9 to 64.8 and outperforms frontier closed-source models.
- LenVM also enables controllable trade-offs between performance and efficiency, preserving 63% GSM8K accuracy at a 200-token budget (vs. 6% for a token-budget baseline) and providing interpretable token-level signals about reasoning length regimes.
Related Articles

Why Autonomous Coding Agents Keep Failing — And What Actually Works
Dev.to

Text-to-image is easy. Chaining LLMs to generate, critique, and iterate on images autonomously is a routing nightmare. AgentSwarms now supports Image generation playground and creative media workflows!
Reddit r/artificial

Why Enterprise AI Pilots Fail
Dev.to

Announcing the NVIDIA Nemotron 3 Super Build Contest
Dev.to

75% of Sites Blocking AI Bots Still Get Cited. Here Is Why Blocking Does Not Work.
Dev.to