NVMe RAID0 at dual-channel DDR5 bandwidth?

Reddit r/LocalLLaMA / 3/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisTools & Practical Usage

Key Points

  • A Reddit user discusses whether building an NVMe RAID0 array on certain AM5 motherboards could reach aggregate bandwidth comparable to dual-channel DDR5 for LLM inference workloads.
  • The proposal uses six Samsung 9100 Pro NVMe SSDs, each rated at up to 14.8GB/s sequential read, to estimate roughly 88.8GB/s peak RAID0 bandwidth when the full PCIe lane bandwidth is available.
  • The author acknowledges that RAID0 and SSDs still have much higher latency than DRAM, and notes the 14.8GB/s figure reflects sequential peak performance rather than real inference access patterns.
  • The motivation is balancing performance and cost by targeting much higher storage capacity per dollar (minimum ~6TB across six drives) for local LLM use, despite latency penalties.
  • The core question is whether such a high-throughput storage setup can materially help LLM inference compared with relying primarily on memory bandwidth.

Been wondering if anyone has tried this or at least considered.

Basically, with some AM5 mobos, like Asus Pro WS B850M-ACE SE, one could install 6x Samsung 9100 Pro NVMe SSDs (2 directly in M.2 slots, 4 in x16 slot bifurcated), each with peak 14.8GB/s sequential read speeds, with full 5.0 x4 PCIe lanes. That'd add up to 88.8GB/s peak bandwidth in RAID0, falling into the range of dual-channel DDR5 bandwidth.

I'm aware that latency is way worse with SSDs, and that 14.8GB/s is only the sequential peak, but still, wouldn't that approach dual-channel DDR5 in LLM inference tasks while giving way more capacity per dollar? The minimum capacity with 9100 Pros would be 6TB total.

submitted by /u/ABLPHA
[link] [comments]