Remix.run Logo
keeda 7 days ago

I posted a comment yesterday regarding this with links to a couple relevant studies: https://news.ycombinator.com/item?id=44793392 -- briefly:

* Even with all this infra buildout all the hyperscalers are constantly capacity constrained, especially for GPUs.

* Surveys are showing that most people are only using AI for a fraction of the time at work, and still reporting significant productivity benefits, even with current models.

The AGI/ASI hype is a distraction, potentially only relevant to the frontier model labs. Even if all model development froze today, there is tremendous untapped demand to be met.

The Metaverse/VR/AR boom was never a boom, with only 2 big companies (Meta, Apple) plowing any "real" money into it. Similarly with crypto, another thing that AI is unjustifiably compared to. I think because people were trying to make it happen.

With the AI boom, however, the largest companies, major governments and VCs are all investing feverishly because it is already happening and they want in on it.

Animats 7 days ago | parent [-]

> Even with all this infra buildout all the hyperscalers are constantly capacity constrained, especially for GPUs.

Are they constrained on resources for training, or resources for serving users using pre-trained LLMs? The first use case is R&D, the second is revenue. The ratio of hardware costs for those areas would be good to know.

keeda 7 days ago | parent [-]

Good question, I don't believe they break out their workloads into training versus inference, in fact they don't even break out any mumbers in any useful detail. But anecdotally the public clouds did seem to be most GPU-constrained whenever Sam Altman was making the rounds asking for trillions in infra for training.

However, my understanding is that the same GPUs can be used for both training and inference (potentially in different configurations?) so there is a lot of elasticity there.

That said, for the public clouds like Azure, AWS and GCP, training is also a source of revenue because other labs pay them to train their models. This is where accusations of funny money shell games come into play because these companies often themselves invest in those labs.