A Qwen finetune, that feels VERY human

Reddit r/LocalLLaMA / 5/4/2026

💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • The author says they were asked to create an “Assistant_Pepe_32B” model and chose Qwen3-32B as the best base candidate despite it being difficult to fine-tune beyond STEM tasks.
  • Assistant_Pepe is described as an assistant style that lacks a “typical assistant brain,” using a negativity bias intended to reduce sycophancy.
  • The post links to prior discussions (on Reddit) that reportedly informed the concept and hypotheses behind the model.
  • The author concludes that the resulting model feels unusually “human,” which is noteworthy because it is still built on top of Qwen.
  • More technical/contextual information is provided via the model card on Hugging Face (Assistant_Pepe_32B).

Hello guys,

So TL;DR, I was asked by multiple people to make an Assistant_Pepe_32B version, but the best base model contender was Qwen3-32B, a model that is very hard to tune on anything other than STEM.

The concept of Assistant_Pepe is an assistant without a typical 'assistant brain', that is infused with negativity bias to reduce sycophancy, previous discussions can be found here and here.

I don't wanna bore you too much with a wall of text, because the above discussions truly did a great job, and great ideas and hypothesis were raised there.

I'll conclude with this: this is probably one of the more "human" models out there, which by itself is quite interesting, because it's a Qwen underneath.

More details in the model card:
https://huggingface.co/SicariusSicariiStuff/Assistant_Pepe_32B

submitted by /u/Sicarius_The_First
[link] [comments]