Qwen3.6 is maintaining context inside the CoT

Reddit r/LocalLLaMA / 4/17/2026

💬 OpinionSignals & Early TrendsTools & Practical Usage

Key Points

  • The post claims that Qwen 3.6 can maintain context within its chain-of-thought (CoT) during reasoning.
  • In repeated tests, the model reportedly remembered a chosen number reliably when that number was selected during the reasoning process.
  • The writer notes that getting the model to consistently stick to the number can be difficult, even if context retention is strong.
  • They specify that enabling this behavior requires using the chat template argument `--chat-template-kwargs '{"preserve_thinking": true}'`.
Qwen3.6 is maintaining context inside the CoT

I tested it in several iterations, and although it's sometimes hard to make the model stick to the number, it reliably remembered the number when it was chosen during reasoning. You have to add --chat-template-kwargs '{"preserve_thinking": true}' for this to actually work.

submitted by /u/Big_Mix_4044
[link] [comments]