|
| ▲ | Aurornis 5 days ago | parent | next [-] |
| The cost of “manufacturing” an AI response is the inference cost, which this article covers. > That would be like saying the unit economics of selling software is good because the only cost is some bandwidth and credit card processing fees. You need to include the cost of making the software Unit economics is about the incremental value and costs of each additional customer. You do not amortize the cost of software into the unit economics calculations. You only include the incremental costs of additional customers. > just like you need to include the cost of making the models. The cost of making the models is important overall, but it’s not included in the unit economics or when calculating the cost of inference. |
|
| ▲ | voxic11 5 days ago | parent | prev | next [-] |
| That isn't what unit economics is. The purpose of unit economics is to answer: "How much money do I make (or lose) if I add one more customer or transaction?". Since adding an additional user/transaction doesn't increase the cost of training the models you would not include the cost of training the models in a unit economics analysis. The entire point of unit economics is that it excludes such "fixed costs". |
|
| ▲ | barrkel 5 days ago | parent | prev | next [-] |
| There is no marginal cost for training, just like there's no marginal cost for software. This is why you don't generally use unit economics for analyzing software company breakeven. |
| |
| ▲ | cj 5 days ago | parent [-] | | The only reason unit economics aren't generally used for software companies is the profit margin is typically 80%+. The cost of posting a Tweet on Twitter/X is close to $0. Compare the cost of tweeting to the cost of submitting a question to ChatGPT. The fact that ChatGPT rate limits (and now sells additional credits to keep using it after you hit the limit) indicates there are serious unit economic considerations. We can't think of OpenAI/Anthropic as software businesses. At least from a financial perspective, it's more similar to a company selling compute (e.g. AWS) than a company selling software (e.g. Twitter/X). |
|
|
| ▲ | ascorbic 5 days ago | parent | prev | next [-] |
| You can amortise the training cost across billions of inference requests though. It's the marginal cost for inference that's most interesting here. |
|
| ▲ | cwyers 5 days ago | parent | prev | next [-] |
| The thing about large fixed costs is that you can just solve them with growth. If they were losing money on inference alone no amount of growth would help. It's not clear to me there's enough growth that everybody makes it out of this AI boom alive, but at least some companies are going to be able to grow their way to profitability at some point, presumably. |
|
| ▲ | martinald 5 days ago | parent | prev [-] |
| But what about running Deepseek R1 or (insert other open weights model here)? There is no training cost for that. |
| |
| ▲ | JCM9 5 days ago | parent [-] | | 1. Someone is still paying for that cost. 2. “Open source” is great but then it’s just a commodity. It would be very hard to build a sustainable business purely on the back of commoditized models. Adding a feature to an actual product that does something else though? Sure. | | |
| ▲ | scarface_74 5 days ago | parent [-] | | There is plenty of money to be made from hosting open source software. AWS for instance makes tons of money from Linux, MySQL, Postgres, Redis, hosting AI models like DeepSeek (Bedrock) etc. |
|
|