What in tarnation is going on with the cost of compute

Reddit r/LocalLLaMA / 5/1/2026

💬 OpinionSignals & Early TrendsTools & Practical UsageIndustry & Market Moves

Key Points

  • The post raises alarm that cloud GPU prices (including H100/H200/B200 instances) have risen above $1,000 per hour for sustained periods on platforms the author uses.
  • The author says they cannot find certain GPUs (e.g., a “B200” server GPU on Vast) and observes the price spike across multiple points during the previous week.
  • They question the market dynamics behind the compute cost increase and argue that it makes it difficult for academics and smaller teams to afford training and development.
  • The author suggests that, if prices remain high, users would likely migrate to alternatives such as RunPod where pricing is described as more reasonable.
  • An update states the issue affects the author’s local Llama/Bitnet pipeline work, delaying model training and development intended for the community.

Does anyone know? I can’t even find a server gpu <b200 on vast, and for the first time that I’ve ever seen on mithril, at multiple points last week have h100/h200/b200 all been at over $1k an hour, for sustained periods! I don’t know why you wouldn’t just migrate to runpod at that point, even their pricing isn’t that costly.

Seriously, academics can’t afford that, and I’d assume startups would just buy hardware to lock compute prices in. What in gods green Earth is going on?

———

EDIT: this applies to localLlama as I am literally training models / developing projects expressly for the consumption of the community here. I can’t finish my bitnet pipeline until pricing comes back down.

submitted by /u/Party-Special-5177
[link] [comments]