Qwen 3.6 27B on Strix Halo 128GB: any experiences?

Reddit r/LocalLLaMA / 4/28/2026

💬 OpinionSignals & Early TrendsTools & Practical Usage

Key Points

  • The post asks community members for real-world experiences running Qwen 3.6 27B on Strix Halo with 128GB of memory.
  • The requester is particularly interested in performance as the context window grows toward a full 256K tokens.
  • The author planned to use RunPod for testing but notes that the required setup/instance is not available there.
  • Overall, it’s a call for practical benchmarking and troubleshooting insights from others who have already run this configuration locally.

I'd jump on runpod and ssh in to test my workloads, but they don't have it.

Would love to know how well this runs, particularly as context approaches a full 256K.

Thanks!

submitted by /u/boutell
[link] [comments]