I’m trying to understand why models from major labs (GPT, Claude, etc.) dominate real-world usage? You might say it's due to the expensive pretraining compute budge, but there already exists many pretrained open-source models at the same scale (e.g., Kimi).
Of course Kimi isn't as good as Claude, but it's the RL on top of the pretraining that makes Claude what it is right? Given Kimi, DeepSeek etc all have the expensive pretraining done, the RLHF on top is what makes Claude what it is right? And that should be much more accessible in terms of cost to smaller labs no?
[link] [comments]



