I'm running Qwen/Qwen3.6-27B-FP8 via vLLM using this command: vllm serve Qwen/Qwen3.6-27B-FP8 --tensor-parallel-size 4 --gpu-memory-utilization 0.95 --max-num-seqs 8 \ --enable-auto-tool-choice --tool-call-parser qwen3_xml \ --enable-prefix-caching --attention-backend flashinfer
It works pretty well in Claude Code, except fairly often it will announce its about to do something, then just stops and waits for a user response. E.g.:
``` Let me continue with the remaining edits.
✻ Brewed for 48s
```
(waiting for user input)
No error message, no failed tool call as far as I can tell, it just fails to follow through. Sometimes it will do it several times in a row and even comment "The user replied 'continue' - they want me to continue. Let me continue with the remaining edits." (user prompt waiting for me to reply)
Is this just a deficiency in the model's thinking, an incompatibility between Claude Code's prompts and the model, or an error in the configuration?
I haven't seen this happen in OpenCode, but there are reasons I prefer CC for some tasks.
Thanks.
[link] [comments]




