I like to help out on this sub and spend a lot of time reading / answering questions to help people getting into AI. Today I spotted a post where someone was asking for advice on models between certain sizes. I clicked on it when it was 1 min old, and wrote a quick answer. It got blocked because the thread had already been locked.
The mod message under the post said the user should habe used a particular thread, which they kindly linked.
I went to read the rules, to see which one they actually fell foul of, and I realised that their post didn't actually break any of the 5 listed rules. It was especially not low effort.
Now, I understand the need to prevent a flood of similar posts, but the one thread solution just feels awful. I open it and am faced immediately with 12 day old comments. It puts a whole wall between the casual user who might be able to help, and the person needing help.
Not suggesting you open it up... but i do think this kind of thing is worth finding a good solution to. Having an interactive and supportive community is something that a lot of sub admits try desperately to encourage, because it keeps people around... my advice is to make that as easy as possible for people!
Cheers!
-------
Oh -and u/adventurous-gold6413 - I was going to say:
With 16gb vram and 64gb dram, I don't know why you are limiting yourself like that.
Try Gemma 4 31b. If that is too slow, try Gemma 4 26b.the 26b is an MoE, and I have run it at speed on a 6gb 2060 laptop gpu and ram spillover.
Try qwen3.6 35b a3b in Q6 and Q4. It is a fantastic model, and it will run on your rig. I have run this model on the 6gb vram laptop too.
Why are you focused on max 32b?
[link] [comments]