Why is no open weight model inference provider hosting Mimo-v2.5 or Mimo-v2.5-pro?

Reddit r/LocalLLaMA / 5/5/2026

💬 OpinionSignals & Early TrendsTools & Practical UsageIndustry & Market Moves

Key Points

  • The article is a Reddit post questioning why third-party API inference providers are not hosting Xiaomi’s Mimo-v2.5 or Mimo-v2.5-pro open-weight models.
  • It claims the models are strong in token efficiency and have a notably low hallucination rate compared with several other models (e.g., Kimi-k2.6, Deepseek-V4, GLM-5.1).
  • Despite the perceived quality, the post says no provider (and even mentions “chutes”) offers these models for inference besides Xiaomi itself.
  • The author characterizes the lack of hosting as unusual and expresses curiosity about the underlying reason.

Literally no 3rd party api inference provider is hosting the mimo-2.5 series models from Xiaomi. They seem to be reallly good.

High token efficiency and very low halucination rate compared to Kimi-k2.6, Deepseek-V4 or GLM-5.1, and yet no provider not even chutes is hosting it other than Xiaomi themselves.

I find it very strange.

submitted by /u/True_Requirement_891
[link] [comments]