Why do only big ML labs dominate widely-used models despite many open-source pretrained models smaller labs could do RL on? [D]
·
0 reactions
·
0 comments
·
2 views
I’m trying to understand why models from major labs (GPT, Claude, etc.) dominate real-world usage? You might say it's due to the expensive pretraining compute budge, but there already exists many pretrained open-source models at the same scale (e.g., Kimi). Of course Kimi isn't as good as Claude, but it's the RL on top of the pretraining that makes Claude what it is right? Given Kimi, DeepSeek etc all have the expensive pretraining done, the RLHF on top is what makes Claude what it is right? And
Original article
Reddit
Anonymous · no account needed