SemRep: Generative Code Representation Learning with Code Transformations
arXiv cs.LG / 3/17/2026
💬 OpinionModels & Research
Key Points
- SemRep proposes using semantics-preserving code transformations as an intermediate representation to guide generative code transformations and downstream instruction-specific edits.
- The framework achieves improvements on general code editing and optimization tasks (e.g., GPU kernel optimization) of 6.9% in correctness, 1.1x in performance, 13.9% in generalization, and 6.7% in robustness when trained with the same budget.
- SemRep enhances exploration of diverse code transformations and works well with an evolutionary coding agent to discover optimizations that much larger baselines miss while using 25% less inference compute for the same performance.
- By decoupling representation learning from end-to-end editing, SemRep provides a more flexible, semantics-guided approach to code transformation.
- The approach demonstrates broad applicability across tasks, suggesting improved robustness and generalization in generative code modeling.
Related Articles

OpenAI just gave up on Sora and its billion-dollar Disney deal
Reddit r/artificial

Paged Attention in Large Language Models LLMs
MarkTechPost

OpenAI is shutting down Sora, its powerful AI video model, app and API
VentureBeat

OpenAI just gave up on Sora and its billion-dollar Disney deal
The Verge
New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B
Reddit r/LocalLLaMA