CROP: Token-Efficient Reasoning in Large Language Models via Regularized Prompt Optimization
arXiv cs.CL / 4/17/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes CROP (Cost-Regularized Optimization of Prompts), an automatic prompt optimization method that reduces LLM token usage by explicitly regularizing response length during optimization.
- Unlike existing APO approaches that optimize only task accuracy, CROP adds feedback based on response length, encouraging the model to output concise answers with only critical reasoning.
- Experiments on GSM8K, LogiQA, and BIG-Bench Hard show a reported 80.6% reduction in token consumption while maintaining competitive accuracy with only a nominal performance drop.
- The authors position CROP as a practical technique for deploying token-efficient, cost-effective agentic AI systems in production pipelines where latency and token cost matter.
Related Articles

Claude and I aren't vibing at all
Dev.to

The ULTIMATE Guide to AI Voice Cloning: RVC WebUI (Zero to Hero)
Dev.to

From Generic to Granular: AI-Powered CMA Personalization for Solo Agents
Dev.to

Kiwi-chan Devlog #007: The Audit Never Sleeps (and Neither Does My GPU)
Dev.to

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to