| ▲ | toasty228 4 hours ago | |||||||||||||||||||||||||
> The demand for AI is currently overwhelming. Wait until they charge the real pice, if I sold a dollar for 10ct I'd also have a lot of demand. I'm burning billions of tokens on chatgpt "deepresearch Pro extended" for things I wouldn't even bother googling, the second I have to pay even 2x the price I won't use that anymore | ||||||||||||||||||||||||||
| ▲ | hootz 2 hours ago | parent | next [-] | |||||||||||||||||||||||||
Can't that be countered by the fact that you can pay a reasonable price (something like 20 or 30 bucks) for small businesses independent flat-rate inference subscriptions of models like GLM-5.1? They aren't being subsidized, they just balance normal and power users around their flat rate. Just check something like synthetic.new, Ollama Cloud or OpenCode Go. | ||||||||||||||||||||||||||
| ▲ | vanuatu 4 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||
I hear this analogy (selling a dollar for 10ct) but it's unclear to me how we can cleanly map intelligence to cents. If the LLM was GPT-1, most people wouldn't even use it for free. So clearly there's another axis here? | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
| ▲ | ls612 4 hours ago | parent | prev [-] | |||||||||||||||||||||||||
The estimates I've seen are that running inference at scale on a Deepseek V3 sized model (so 700B parameters) costs roughly $0.70/mtok or so given current H100 rental costs. Sonnet charges $15/mtok on the API so the delta between the true cost and the API cost is quite large, to the point where even many subscription users are likely profitable. | ||||||||||||||||||||||||||