▲ | fooker 3 days ago | |
> you need scale. No, training state of the art LLMs is still a bit of alchemy. We don't understand what works and what doesn't. Meta is paying 100M each to hire AI researchers not because they know how to scale (they aren't bringing GPUs lol), but mainly because they remember what worked and what didn't for training GPT4. > If I would be patient.. No, you'd spend the time and resources training and end up with something worse than even GPT3. This is what made Deepseek appear in headlines for two months straight. Plenty of other companies have 100x more resources and are actively trying to have their own LLMs. Including big names like Apple and Oracle. They haven't managed to. |