IQuest-Coder-V1 Technical Report
arXiv cs.AI / 3/18/2026
📰 NewsIndustry & Market MovesModels & Research
Key Points
- The IQuest-Coder-V1 family introduces code LLMs (7B/14B/40B/40B-Loop) and a code-flow multi-stage training paradigm that models the evolving logic of software pipelines.
- The training pipeline includes an initial pre-training on code facts, repository data, and completion data; a mid-training stage with reasoning and agentic trajectories at 32k-context and repository-scale 128k-context; and a post-training phase for specialized coding capabilities via a thinking path (reasoning-driven RL) and an instruct path (general assistance).
- The IQuest-Coder-V1-Loop variant adds a recurrent mechanism to balance model capacity and deployment footprint, enabling an efficiency-focused deployment path.
- The authors claim state-of-the-art performance in code intelligence across agentic software engineering, competitive programming, and complex tool use.
- They release the complete white-box chain of checkpoints from pre-training bases to final thinking and instruction models, aiming to advance autonomous code intelligence and real-world agentic systems.




