Towards Cold-Start Drafting and Continual Refining: A Value-Driven Memory Approach with Application to NPU Kernel Synthesis
arXiv cs.LG / 3/12/2026
📰 NewsModels & Research
Key Points
- EvoKernel is introduced as a self-evolving agentic framework that automates the lifecycle of kernel synthesis from initial drafting to continual refining in data-scarce NPU programming environments.
- It formulates kernel synthesis as a memory-based reinforcement learning task with a novel value-driven retrieval mechanism that prioritizes experiences by their contribution to the current objective, such as bootstrapping a feasible draft or refining latency.
- The approach enables cross-task memory sharing, allowing insights to transfer from simple to complex operators, and includes an NPU-specific variant of KernelBench for evaluation.
- Results show frontier models' correctness improving from 11.0% to 83.0% and a median speedup of 3.60x over initial drafts, demonstrating effective learning for kernel synthesis on niche hardware ecosystems.
Related Articles

**Core Allocation Optimization for Energy‑Efficient Multi‑Core Scheduling in ARINC650 Systems**
Dev.to

LongCat-Flash-Prover: A new frontier for Open-Source Formal Reasoning.
Reddit r/LocalLLaMA

composer 2 is just Kimi K2.5 with RL?????
Reddit r/LocalLLaMA

Built a small free iOS app to reduce LLM answer uncertainty with multiple models
Dev.to
![[P] We built a Weights & Biases for Autoresearch - track steps, compare experiments, and share results](/_next/image?url=https%3A%2F%2Fpreview.redd.it%2Flv7w6809f7qg1.png%3Fwidth%3D140%26height%3D75%26auto%3Dwebp%26s%3De77e7b54776d5a33eb092415d26190352ad20577&w=3840&q=75)
[P] We built a Weights & Biases for Autoresearch - track steps, compare experiments, and share results
Reddit r/MachineLearning