I just uploaded a Qwen3.5-9B GGUF that I fine-tuned on a mix of reasoning data and FunctionGemma-related function-calling data, then converted for llama.cpp/GGUF runtimes.
It’s still a Qwen-family model, but the tuning pushes it more toward structured responses, tool-use style behavior, and action-oriented prompting.
If you run local models with llama.cpp, LM Studio, Ollama, or similar, I’d be interested in hearing how it performs for:
- general chat
- reasoning tasks
- structured outputs
- function-calling style prompts
Repo link: Huggingface
[link] [comments]



