Helping to make the sub more helpful

Reddit r/LocalLLaMA / 4/27/2026

💬 Opinion

Key Points

  • The author says they frequently help newcomers to AI on the subreddit, and responded quickly to a post asking for advice on model sizes, but their answer was blocked after the thread was locked.
  • The moderator directed the user to a specific older thread, and the author checked the rules and concluded the original post did not actually violate any of the listed low-effort rules.
  • They argue that funneling questions into one thread creates a “wall” of old comments that discourages casual users who could help, and delays people who need answers.
  • The author suggests the community should find a better solution that keeps it interactive and supportive while still preventing repetitive floods of similar posts.
  • The post also includes practical model suggestions (e.g., Gemma and Qwen variants) for users with limited VRAM.
  • categories: [

I like to help out on this sub and spend a lot of time reading / answering questions to help people getting into AI. Today I spotted a post where someone was asking for advice on models between certain sizes. I clicked on it when it was 1 min old, and wrote a quick answer. It got blocked because the thread had already been locked.

The mod message under the post said the user should habe used a particular thread, which they kindly linked.

I went to read the rules, to see which one they actually fell foul of, and I realised that their post didn't actually break any of the 5 listed rules. It was especially not low effort.

Now, I understand the need to prevent a flood of similar posts, but the one thread solution just feels awful. I open it and am faced immediately with 12 day old comments. It puts a whole wall between the casual user who might be able to help, and the person needing help.

Not suggesting you open it up... but i do think this kind of thing is worth finding a good solution to. Having an interactive and supportive community is something that a lot of sub admits try desperately to encourage, because it keeps people around... my advice is to make that as easy as possible for people!

Cheers!

-------

Oh -and u/adventurous-gold6413 - I was going to say:

With 16gb vram and 64gb dram, I don't know why you are limiting yourself like that.

Try Gemma 4 31b. If that is too slow, try Gemma 4 26b.the 26b is an MoE, and I have run it at speed on a 6gb 2060 laptop gpu and ram spillover.

Try qwen3.6 35b a3b in Q6 and Q4. It is a fantastic model, and it will run on your rig. I have run this model on the 6gb vram laptop too.

Why are you focused on max 32b?

submitted by /u/Ell2509
[link] [comments]