| ▲ | gruez 8 hours ago |
| How viable are the $20/month subscriptions for actual work and are they loss making for Anthropic? I've heard both of people needing to get higher tiers to get anything done in Claude Code and also that the subscriptions are (heavily?) subsidized by Anthropic, so the "just another $20 SaaS" argument doesn't sound too good. |
|
| ▲ | simonw 8 hours ago | parent | next [-] |
| I am confident that Anthropic make revenue from that $20 than the electricity and server costs needed to serve that customer. Claude Code has rate limits for a reason: I expect they are carefully designed to ensure that the average user doesn't end up losing Anthropic money, and that even extreme heavy users don't cause big enough losses for it to be a problem. Everything I've heard makes me believe the margins on inference are quite high. The AI labs lose money because of the R&D and training costs, not because they're giving electricity and server operational costs away for free. |
| |
| ▲ | tverbeure 7 hours ago | parent | next [-] | | Nobody questions that Anthropic makes revenue from a $20 subscription. The opposite would be very strange. | | |
| ▲ | simonw 6 hours ago | parent | next [-] | | A lot of people believe that Anthropic lose money selling tokens to customers because they are subsidizing it for growth. | | |
| ▲ | Drakim 2 hours ago | parent [-] | | But that has zero effect on revenue, it only affects profit. |
| |
| ▲ | brandensilva 7 hours ago | parent | prev [-] | | Yeah it's the caching that's doing the work for them though honestly. So many cached queries saving the GPUs from hard hits. | | |
| ▲ | xienze 3 hours ago | parent [-] | | How is caching implemented in this scenario? I find it unlikely that two developers are going to ask the same exact question, so at a minimum some work has to be done to figure out “someone’s asked this before, fetch the response out of the cache.” But then the problem is that most questions are peppered with specific context that has to be represented in the response, so there’s really no way to cache that. | | |
| ▲ | marcyb5st 3 hours ago | parent [-] | | From my understanding (which is poor at best), the cache is about the separate parts of the input context. Once the LLM read a file the content of that file is cached (i.e. some representation that the LLM creates for that specific file, but I really have no idea how that works). So the next time you bring either directly or indirectly that file into the context the LLM doesn't have to do a full pass, but pull its understanding/representation from the cache and uses that to answer your question/perform the task. |
|
|
| |
| ▲ | Esophagus4 8 hours ago | parent | prev | next [-] | | I always assumed that with inference being so cheap, my subscription fees were paying for training costs, not inference. | | |
| ▲ | beAbU 2 hours ago | parent | next [-] | | Is inference really that cheap? Why can't I do it at home with a reasonable amount of money? | |
| ▲ | simonw 6 hours ago | parent | prev | next [-] | | Anthropic and OpenAI are both well documented as losing billions of dollars a year because their revenue doesn't cover their R&D and training costs, but that doesn't mean their revenue doesn't cover their inference costs. | | |
| ▲ | overgard 6 hours ago | parent | next [-] | | Does it matter if they can't ever stop training though? Like, this argument usually seems to imply that training is a one-off, not an ongoing process. I could save a lot of money if I stopped eating, but it'd be a short lived experiment. I'll be convinced they're actually making money when they stop asking for $30 billion funding rounds. None of that money is free! Whoever is giving them that money wants a return on their investment, somehow. | | |
| ▲ | vidarh 3 hours ago | parent | next [-] | | At some point the players will need to reach profitability. Even if they're subsidising it with other revenue - they'll only be willing to do that as long as it drives rising inference revenue. Once that happens, whomever is left standing can dial back the training investment to whatever their share of inference can bear. | | |
| ▲ | ben_w 3 hours ago | parent [-] | | > Once that happens, whomever is left standing can dial back the training investment to whatever their share of inference can bear. Or, if there's two people left standing, they may compete with each other on price rather than performance and each end up with cloud compute's margins. | | |
| ▲ | vidarh an hour ago | parent [-] | | Sure, but they will still need to dial it back to a point where they can fund it out of inference at some point. The point is that the fact they can't do that now is irrelevant - it's a game of chicken at the moment, and that might kill some of them, but the game won't last forever. |
|
| |
| ▲ | simonw 5 hours ago | parent | prev | next [-] | | It matters because as long as they are selling inference for less than it costs to serve they have a potential path to profitability. Training costs are fixed at whatever billions of dollars per year. If inference is profitable they might conceivably make a profit if they can build a model that's good enough to sign up vast numbers of paying customers. If they lose even more money on each new customer they don't have any path to profitability at all. | | |
| ▲ | citrin_ru 42 minutes ago | parent [-] | | > If they lose even more money on each new customer they don't have any path to profitability at all. In theory they can increase prices once the customers will be hocked up. That's how many startups works. |
| |
| ▲ | krainboltgreene 5 hours ago | parent | prev [-] | | There's an argument to be made that a "return on investment by way of eliminating all workers" is a reasonable result for the capitalists. | | |
| ▲ | generic92034 3 hours ago | parent [-] | | At least until they are running out of customers. And/or societies with mass-unemployment destabilize to a degree that is not conducive for capitalists' operations. | | |
|
| |
| ▲ | vrighter 3 hours ago | parent | prev [-] | | Models are fixed. They do not learn post training. Which means that training needs to be ongoing. So the revenue covers the inference? So what? All that means is that it doesn't cover your costs and you're operating at a loss. Because it doesn't cover the training that you can't stop doing either. |
| |
| ▲ | smashed 7 hours ago | parent | prev [-] | | Doubtful |
| |
| ▲ | what 5 hours ago | parent | prev [-] | | >make revenue from that $20 than the electricity and server costs needed to serve that customer Seems like a pretty dumb take. It’s like saying it only takes $X in electricity and raw materials to produce a widget that I sell for $Y. Since $Y is bigger than $X, I’m making money! Just ignore that I have to pay people to work the lines. Ignore that I had to pay huge amounts to build the factory. Ignore every other cost. They can’t just fire everyone and stop training new models. | | |
|
|
| ▲ | _jss 7 hours ago | parent | prev | next [-] |
| Merely for the viability part: I use the $20/mo plan now, but only as a part-time independent dev. I will hit rate-limits with Opus on any moderately complex app. If I am on a roll, I will flip on Extra Usage. I prototyped a fully functional and useful niche app in ~6 total hours and $20 of extra usage, and it's solid enough and proved enough value to continue investing in and eventually ship to the App store. Without Claude I likely wouldn't have gotten to the finished prototype version to use in the real world. For Indy dev, I think LLMs are a new source of solutions. This app is too niche to justify building and marketing without LLM assistance. It likely won't earn more than $25k/year but good enough! |
|
| ▲ | Aurornis 7 hours ago | parent | prev | next [-] |
| I don't think the assumption that Anthropic is losing money on subscriptions holds up. I think each additional customer provides more revenue than the cost to run their inference, on average. For people doing work with LLMs as an assistant for codebase searching, reviews, double checks, and things like that the $20/month plan is more than fine. The closer you get to vibecoding and trying to get the LLM to do all the work, the more you need the $100 and $200 plans. On the ChatGPT side, the $20/month subscription plan for GPT Codex feels extremely generous right now. I tried getting to the end of my window usage limit one day and could not. > so the "just another $20 SaaS" argument doesn't sound too good Having seen several company's SaaS bills, even $100/month or $200/month for developers would barely change anything. |
|
| ▲ | 8note 8 hours ago | parent | prev [-] |
| id guess the 200 subscription sufficient per person. but at that point you could go for a bugger one and split amongst headcount |