AI Navigate

What non-Chinese models are relevant right now?

Reddit r/LocalLLaMA / 3/14/2026

💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The author is running local models on a state-owned research cluster, with hardware considerations largely non-issues but explicit restrictions against DeepSeek, AliBaba, and any other Chinese models.
  • The current frontier of non-Chinese models cited includes GPT-OSS, Nemotron, and Mistral, with IBM Granite mentioned as a potential small tool-calling option.
  • While Olmo is liked for various reasons, it is not considered universally optimal for any single job, prompting exploration of other model families.
  • The post requests recommendations on additional non-Chinese model families to consider (e.g., Gemma, Phi, Llama 4) to broaden options.

Started running local models for a variety of purposes on state-owned research cluster. VRAM and inference time are essentially non-issues, but I explicitly can't use DeepSeek or AliBaba products or their derivatives, and, implicitly, any other Chinese models would be heavily frowned upon. It seems like GPT-OSS, Nemotron, and Mistral models make up the frontier of non-Chinese models right now, maybe including something like IBM Granite for small tool calling models. I really like Olmo for a variety of reasons, but it's probably not the best tool for any job. Are there any model families I'm unaware of that I should be looking at? Gemma? Phi? Llama 4?

submitted by /u/StacDnaStoob
[link] [comments]