GRaPE 2 Model Family

Reddit r/LocalLLaMA / 4/15/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • The post announces the GRaPE 2 model family, providing two new releases: GRaPE Mini (5B) and GRaPE Flash (9B).
  • GRaPE 2 supports six “thinking modes” that let users allocate compute budgets to reduce overthinking compared with Qwen3.5-style models.
  • The models were trained extensively on agent-style examples (e.g., code agents and browser agents) and are reported to have decent coding performance.
  • The author credits r/unsloth for enabling GRaPE 2 and links to the Hugging Face repositories for both model variants.

Today I announce the first two models I am posting on here! First off, hello all of r/LocalLLaMA, nice to join. But I would love to show off the General Reasoning Agent for Project Exploration, dubbed as GRaPE. GRaPE is on the second generation, and has two models

  1. GRaPE Mini
  2. GRaPE Flash

These models are 5B and 9B respectively, and support 6 thinking modes to allocate budgets, so you don't get overthinking like in the Qwen3.5 models. All of which is detailed in the Huggingface repo at the end of this post. I have generally found medium / low is the sweet spot, but minimal exists if you cannot bear thinking at all.

GRaPE 2 was trained with lots and lots of examples of being an agent, so code agent, browser agent, etc; And the models has decent coding performance!

Huge thanks to r/unsloth for making GRaPE 2 possible.

https://huggingface.co/SL-AI/GRaPE-2-Mini

https://huggingface.co/SL-AI/GRaPE-2-Flash

submitted by /u/SweaterDog_YT
[link] [comments]