| ▲ | sandworm101 3 hours ago | ||||||||||||||||||||||
But that per-token cost is a total joke. All these companies are fighting to build market share in some future dominated by one or two AI ecosystems. It is musical chairs until someone creates the one ring to rule them all. So they are charging token amounts just to claim revenue as they burn through investor dollars. In short: per-token charges currently cover maybe 1% of the total costs in this field. To pay ongoing costs, and pay back investors, everyone will need to pay 100x or 1000x the current rates, likely for decades. | |||||||||||||||||||||||
| ▲ | red_hare an hour ago | parent | next [-] | ||||||||||||||||||||||
If that's true, it's very unsustainable. Gemma-4 26B-A4B + M5 MacBook Pro + OpenCode isn't Claude Code _yet_, but it's good enough that if I were forced to use it I would be fine. | |||||||||||||||||||||||
| |||||||||||||||||||||||
| ▲ | matrik an hour ago | parent | prev | next [-] | ||||||||||||||||||||||
I'm not sure this information is grounded, but I remember to have read somewhere the inference is indeed profitable. My personal experience is similar. Running 2x3090s draw 500-600W and you can locally run amazing models with a similar setup. | |||||||||||||||||||||||
| |||||||||||||||||||||||
| ▲ | deaux 2 hours ago | parent | prev | next [-] | ||||||||||||||||||||||
> In short: per-token charges currently cover maybe 1% of the total costs in this field There are plenty of seemingly informed people saying the exact opposite, so that's a lot of confidence you're talking with. I have a hard time believing it when we know what open weights models cost to run. And sure, there's training costs, but again many say inference costs are already above training costs. | |||||||||||||||||||||||
| ▲ | twoodfin 2 hours ago | parent | prev | next [-] | ||||||||||||||||||||||
From the perspective of a deal like this, “total costs in the field” matter less than incremental cost per token served. The unit economics for today’s frontier models should be great, and this suggests Anthropic believes they’ll get better. | |||||||||||||||||||||||
| ▲ | postalrat 2 hours ago | parent | prev [-] | ||||||||||||||||||||||
In a decade the cost of compute will be a tiny fraction of what it costs now. Specialized hardware will exist that will be cheap and efficient. | |||||||||||||||||||||||
| |||||||||||||||||||||||