SeLaR: Selective Latent Reasoning in Large Language Models
arXiv cs.CL / 4/10/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes SeLaR (Selective Latent Reasoning), a training-free method that improves chain-of-thought reasoning in large language models by selectively using latent (soft) reasoning only when the model is uncertain.
- It addresses prior latent-reasoning limitations, including reasoning instability from global soft activation and the tendency of soft embeddings to collapse toward the most likely token.
- SeLaR uses an entropy-gated mechanism to switch between soft embeddings at low-confidence steps and discrete decoding at high-confidence steps, aiming to preserve stability while enabling exploration.
- It adds entropy-aware contrastive regularization that discourages soft embeddings from aligning with the dominant token direction, encouraging multiple reasoning trajectories.
- Experiments across five reasoning benchmarks report consistent performance gains over standard CoT and other training-free approaches.
Related Articles

Black Hat Asia
AI Business
v0.20.5
Ollama Releases

Inside Anthropic's Project Glasswing: The AI Model That Found Zero-Days in Every Major OS
Dev.to
Gemma 4 26B fabricated an entire code audit. I have the forensic evidence from the database.
Reddit r/LocalLLaMA

SoloEngine: Low-Code Agentic AI Development Platform with Native Support for Multi-Agent Collaboration, MCP, and Skill System
Dev.to