Hard freakin' decision..Blackwell 96G or Mac Studio 256G

Reddit r/LocalLLaMA / 4/24/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsTools & Practical UsageIndustry & Market MovesModels & Research

Key Points

  • The post asks whether to buy a used NVIDIA RTX Pro 6000 Blackwell 96G or a new Apple Mac Studio M3 Ultra 256GB for running large local AI models and supporting multiple smaller ML tasks.
  • The author is primarily targeting “fat” models (e.g., Big Gemma and Qwen variants) plus memory-sensitive workloads like embeddings, re-ranking, STT/TTS, and small fast models for Home Assistant.
  • They consider Blackwell’s key benefits as CUDA support and performance for AI/encoding, but note additional costs and risk: a likely need for a new high-wattage power supply and the uncertainty of buying a used GPU.
  • They see the Mac Studio as a safer, simpler option with more reliable new hardware and better total RAM availability for local inference, while debating whether paying extra for higher CPU/GPU core configurations is worth it.
  • The author intends a server/rack AI-only setup via SSH and IP KVM and is not otherwise a Mac user, raising practical platform and workflow questions alongside the hardware choice.

EDIT: OKOKOK. Blackwell all the way. NEW, at MC or NewEgg or where ever and more tokens than my face can handle. Thanks guys. I was close to pulling that Apple.com trigger. You saved me.

So now it's between the workstation and the max-q. Max-q would mean no nee power supply and save me money. Also, airflow would be A LOT better. I had the 4090 in there with no problems iirc though.

------------------------------------------------------------------------------------------------------------

So, I have too much money. Help me help the economy.

US dollarydoo's below:

  • A used RTX Pro 6000 96G card on the ebays is ~10K shipped. NOTE: I didn't know they were 10k new. I thought they were like 15.
  • A new Mac Studio M3 Ultra with 256G is either 6400 or 8K depending on the proc you choose. (shipped prices to my state)

I want to run some fat models. Big Gemma4s or Qwen3.6s. I also have other small models I need to keep in memory. Embedding, re-ranking, tts, stt, small and fast model for Home Assistant, etc.

I am not a mac guy. Linux and windows for me. Haven't touched a mac in 30 years. IF I get one, it'll be AI exclusive and live in a rack accessible via SSH and IP KVM only.

On the PC side, the blackwell card would live in my current server, and I'd need a new 1000-1200watt 3.1 power supply too. It would be video encoding and AI exclusive. It's main advantage is CUDA and doing other things with it that support CUDA.

To me the Mac SEEMS like the MUCH better choice. More RAM, brand new. The blackwell would be used. If it fritzes then I am out 10k.

Also, if Mac is the way to go, do I pay 1500 clams for the upgraded processor/GPU?
28/60 vs 32/80 CPU/GPU cores. Will it make a big enough diff to justify the clams?

Please and thank you.

submitted by /u/HyPyke
[link] [comments]