AI Navigate

Qwen3.5-9B-Claude-4.6-Opus-Uncensored-Distilled-GGUF

Reddit r/LocalLLaMA / 3/16/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research

Key Points

  • A Reddit user released their first fully uncensored LLM model named Qwen3.5-9B-Claude-4.6-Opus-Uncensored-Distilled-GGUF, with links to HuggingFace.
  • The model disables thinking by default via a modified chat template baked into the GGUF file to enable more creative and less restricted responses.
  • The creator describes merging tensors from two Qwen 3.5 variants (HauhauCS Aggressive uncensored and Jackrong Claude-4.6 Opus Reasoning Distilled GGUF) using a Colab script and then evaluating with Claude Opus 4.6.
  • They loaded the model into LM Studio 0.4.7 with specified generation parameters and report zero refusals and creative outputs for a 9B model.

Hello everyone. I made my first fully uncensored LLM model for this community. Here link:
https://huggingface.co/LuffyTheFox/Qwen3.5-9B-Claude-4.6-Opus-Uncensored-Distilled-GGUF

Thinking is disabled by default in this model via modified chat template baked in gguf file.

So, I love to use Qwen 3.5 9B especially for roleplay writing and prompt crafting for image generation and tagging on my NVidia RTX 3060 12 GB, but it misses creativity, contains a lot of thinking loops and refuses too much. So I made the following tweaks:

1) I downloaded the most popular model from: https://huggingface.co/HauhauCS/Qwen3.5-9B-Uncensored-HauhauCS-Aggressive

2) I downloaded the second popular model from: https://huggingface.co/Jackrong/Qwen3.5-9B-Claude-4.6-Opus-Reasoning-Distilled-GGUF

3) I compared HauhauCS checkpoint with standart Qwen 3.5 checkpoint and extracted modified tensors by HauhauCS.

4) I merged modified tensors by HauhauCS with Jackrong tensors.

Everything above was done via this script in Google Colab. I vibecoded it via Claude Opus 4.6: https://pastebin.com/1qKgR3za

On next stage I crafted System Prompt. Here another pastebin: https://pastebin.com/pU25DVnB

I loaded modified model in LM Studio 0.4.7 (Build 1) with following parameters:

Temperature: 0,7
Top K Sampling: 20
Presence Penalty: 1.5
Top P Sampling: 0.8
Min P Sampling: 0
Seed: 3407 or 42

And everything works with pretty nicely. Zero refusals. And responces are really good and creative for 9B model. Now we have distilled uncensored version of Qwen 3.5 9B finetuned on Claude Opus 4.6 thinking logic. Hope it helps. Enjoy. Feel free to tweak my system prompt simplify or extent it if you want.

submitted by /u/EvilEnginer
[link] [comments]