AI Navigate

Best models for RTX 6000 x 4 build

Reddit r/LocalLLaMA / 3/23/2026

💬 OpinionTools & Practical Usage

Key Points

  • The poster is building a four-slot RTX 6000 MAX-Q (384GB) system with 768GB RAM and is seeking guidance on the best models to run with minimal degradation.
  • They are evaluating specific models: Qwen3.5-122B-A10B at BF16 and Qwen3.5-397B-A17B at Q6_K.
  • The intended workloads are hacking tools, fuzzing, and code auditing, indicating security-focused AI usage.
  • They request optimization tips and resources to improve performance of these cards and models.

Hey everyone,

Ive got my 4th RTX 6000 MAX-Q (384GB) (also have 768GB RAM) coming in a couple days, I’ve been looking and doing some reading regarding what the current best models I can run on this are with limited degradation.

So far I’m looking at the following:

Qwen3.5-122B-A10B at BF16

Qwen3.5-397B-A17B at Q6_K

Predominately looking to build out and refine a bundle of hacking tools, some fuzzing, and some code auditing.

Is there any additional optimisation I need to do for these cards and these models?

I’ve already been building stuff out with this, if anyone has any tips or resources they’d recommend please share them with me :)

Thanks

submitted by /u/Direct_Bodybuilder63
[link] [comments]