Two new Qwen3.5 “Neo” fine‑tunes focused on fast, efficient reasoning

Reddit r/LocalLLaMA / 3/24/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • Two new community fine-tunes, Qwen3.5-4B-Neo and Qwen3.5-9B-Neo, are released based on the Qwen3.5 models and are optimized for faster, more efficient reasoning.
  • The 4B “Neo” variant emphasizes shorter internal chain-of-thought to reduce token cost while aiming to maintain or improve accuracy.
  • The 9B “Neo” variant provides a larger, similarly reasoning-focused fine-tune option.
  • Both models are available on Hugging Face, with additional GGUF versions published in a related collection for broader local deployment compatibility.

Hey everyone,

Just wanted to share two new community fine‑tunes I came across: Qwen3.5‑4B‑Neo by Jackrong.

Qwen3.5‑4B‑Neo
A reasoning‑optimized fine‑tune of Qwen3.5‑4B. It focuses heavily on efficient chain‑of‑thought: shorter internal reasoning, lower token cost, and higher accuracy.
HF link: https://huggingface.co/Jackrong/Qwen3.5-4B-Neo

Qwen3.5‑9B‑Neo
A larger variant fine‑tuned of Qwen3.5‑9B.
HF link: https://huggingface.co/Jackrong/Qwen3.5-9B-Neo

GGUF versions are also available in the collection here: https://huggingface.co/collections/Jackrong/qwen35-neo

submitted by /u/FabbBr
[link] [comments]