| I tested it in several iterations, and although it's sometimes hard to make the model stick to the number, it reliably remembered the number when it was chosen during reasoning. You have to add [link] [comments] |
Qwen3.6 is maintaining context inside the CoT
Reddit r/LocalLLaMA / 4/17/2026
💬 OpinionSignals & Early TrendsTools & Practical Usage
Key Points
- The post claims that Qwen 3.6 can maintain context within its chain-of-thought (CoT) during reasoning.
- In repeated tests, the model reportedly remembered a chosen number reliably when that number was selected during the reasoning process.
- The writer notes that getting the model to consistently stick to the number can be difficult, even if context retention is strong.
- They specify that enabling this behavior requires using the chat template argument `--chat-template-kwargs '{"preserve_thinking": true}'`.



