What is the secret sauce Claude has and why hasn't anyone replicated it?

Reddit r/LocalLLaMA / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • A Reddit user argues that Claude’s conversational behavior is unusually distinctive and does not appear to be reproducible by simply transferring Claude’s system prompt to another model (they tried Qwen3.5 27B with Sonnet 4.5’s prompt).
  • The user claims that attempts to distill Claude’s “thinking traces” into other models have not produced convincing results, suggesting the distinctive style may not be easily captured that way.
  • They question whether architectural differences, model scale, or training data contribute to Claude-like “talking style” and “vibes,” since prompts alone haven’t matched observable formatting and verbosity patterns.
  • The post highlights specific mismatches they observed—such as Claude’s avoidance of bullet points and emojis and its tendency toward certain response length/coverage behaviors.
  • Overall, the piece is framed as an open question about what makes Claude’s communication style hard to replicate across LLMs.

I've noticed something about Claude from talking to it. It's very very distinct in its talking style, much more of an individual than some other LLMs I know. I tried feeding that exact same system prompt Sonnet 4.5 to Qwen3.5 27B and it didn't change how it acted, so I ruled out the system prompt doing the heavy lifting.

I've seen many many distills out there claiming that Claude's responses/thinking traces have been distilled into another model and testing is rather... disappointing. I've searched far and wide, and unless I'm missing something (I hope I'm not, apologies if I am though...), I believe that it's justified to ask:

Why can't we make a model talk like Claude?

It's not even reasoning, it's just talking "style" and "vibes", which isn't even hidden from Claude's API/web UI. Is it some sort of architecture difference that just so happens to make a model not be able to talk like Claude no matter how hard you try? Or is it a model size thing along with a good system prompt (a >200B model prompted properly can talk like Claude)?

I've tried system prompts for far too long, but the model seems to always miss:
- formatting (I've noticed Claude strays from emojis and tries to not use bullet points as much as possible, unlike other models)
- length of response (sometimes it can ramble for 5 paragraphs about what Satin is and yet talk about Gated DeltaNets for 1)

Thank you!

submitted by /u/ComplexType568
[link] [comments]