About to build a 6× Arc B70 LLM rig, want to talk to someone experienced first

Reddit r/LocalLLaMA / 4/20/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsTools & Practical Usage

Key Points

  • The post is seeking advice from an experienced builder about creating an LLM rig using six Intel Arc B70 GPUs, without requiring Arc-specific expertise.
  • The author wants input specifically related to running Llama models with vLLM, including how to set up the system correctly from the start.
  • They mention having conducted preliminary tests on a 5090 machine with a 128GB unified-memory configuration and have seen “interesting results.”
  • The author is willing to pay for the time needed for a consultation and requests evidence of relevant experience depending on the hourly rate.
  • Overall, this is a community request for practical guidance rather than a published technical update or benchmark release.

Hello, I’m preparing to build a rig with six Intel Arc B70s, but before I move forward, I’d like to speak with someone who has experience building similar systems (no arc specific knowledge required) , particularly with llama and vLLM.

In my initial tests using a 5090 machine & a 128GB of unified memory system, I’ve been seeing some interesting results. I have several questions and would really value the opportunity to discuss them with someone experienced so I can make informed decisions and set things up correctly from the start.

I’m open to paying for your time; however, depending on the rate, I would appreciate seeing some evidence of relevant experience.

Thanks!

submitted by /u/somesayitssick
[link] [comments]