| ▲ | pier25 a day ago | ||||||||||||||||||||||||||||
So these companies will be profitable if training stops? Is that even a real possibility? | |||||||||||||||||||||||||||||
| ▲ | danaris 8 hours ago | parent | next [-] | ||||||||||||||||||||||||||||
Any given company could stop training tomorrow, and, as some others have said here, they'd be generating quite a bit of profit until their models visibly fell behind, however long that ended up taking, at which point they'd probably just fall over completely. Over the whole industry? No; they can never, ever stop training, or they'll cease to be useful at all very soon. Training is what keeps the models up-to-date on current events, which includes new programming languages, frameworks, and techniques. It's already been observed that using LLM assistance on some types of programming is much more effective than others, based on how well-represented they are in the training data: if everyone stopped training tomorrow, and next month a new programming language came out, none of them would ever be able to help you program in that new language. This can be extended to other aspects of programming, too. If training stopped, coding assistants would gradually start giving you wrong answers on how to implement code for APIs, frameworks, and languages that continued to evolve, as they will always do, in much subtler (and likely harder-to-debug) ways than how they'd deal with a new language whose existence they don't even know about. | |||||||||||||||||||||||||||||
| ▲ | naravara a day ago | parent | prev [-] | ||||||||||||||||||||||||||||
The impetus to continue training at the pace they are is driven by the competition. So if the money starts drying up, then they’ll naturally slow down because they’ll have to figure out how to do more with less. I suspect that once the models hit a point of “good enough” for certain use cases companies will start putting R&D focus in other areas that may be less expensive. Like figuring out how to run more efficiently, UI/UX conventions that help users get what they’re trying to accomplish in fewer steps, various kinds of caching of requests, etc. So the cost to serve tokens over time should only come down, and will probably start coming down more rapidly as the returns to model training slow down. That’ll probably be a while though, because each successive model tends to be a lot better than the last. | |||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||