2x 512gb ram M3 Ultra mac studios

Reddit r/LocalLLaMA / 4/21/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsTools & Practical Usage

Key Points

  • A Reddit user describes having invested about $25k in Mac Studio hardware with dual 512GB RAM and offers to help test workloads based on user requests.
  • They report running DeepSeek V3.2 at Q8 using an Exo backend and are currently troubleshooting why Exo isn’t loading a Q8 version while running GLM 5.1 at Q4 on each machine.
  • The user is waiting for Kimi 2.6 and expects the community to optimize it for MLX/mmap, suggesting ongoing local LLM inference experimentation.
  • The post emphasizes practical, hands-on testing of local LLM models and quantization behavior on high-memory Apple Silicon systems.
2x 512gb ram M3 Ultra mac studios

$25k in hardware. tell me what you want me to load on them and i'll help test.
i've done deepseek v3.2 Q8 so far with exo backend.

currently running GLM 5.1 Q4 on each (troubleshooting why exo isn't loading the Q8 version)

patiently awaiting kimi2.6 for when the community optimizes it for MLX/mmap

submitted by /u/taylorhou
[link] [comments]