| ▲ | lumost 3 days ago |
| once upon a time - engineers often had to concern themselves with datacenter bills, cloud bills, and eventually SaaS bills. We'll probably have 5-10 years of being concerned about AI bills before the AI expense is trivial compared to the human time. |
|
| ▲ | achierius 3 days ago | parent | next [-] |
| "once upon a time"? Engineers concern themselves with cloud bills right now, today! It's not a niche thing either, probably the majority of AWS consumers have to think about this, regularly. |
|
| ▲ | whatever1 2 days ago | parent | prev | next [-] |
| This guy has not been hit with a 100k/mo cloudwatch bill |
| |
| ▲ | lumost 2 days ago | parent | next [-] | | I’ve managed a few 8 figure infrastructure bills. The exponential decrease in hardware costs combined with the exponential growth in software engineer salaries has meant that these bills become inconsequential in the long run. I was at one unicorn which had to spend 10% of Cost of Goods Sold on cloud. Today their biggest costs could run on a modest Postgres cluster in the cloud thanks to progressively better hardware. 100k/mo on cloud watch corresponds to a moderately large software business assuming basic best practices are followed. Optimization projects can often run into major cost overruns where the people time exceeds the discounted future free cash flow savings from the optimization. That being said, a team of 5 on a small revenue/infra spend racking up 100k/mo is excessive. Pedantically, cloud watch/datadog are SaaS vendors - 100k/mo on Prometheus would correspond to a 20 node SSD cluster in the cloud which could easily handle several 10s of millions of metrics per second from 10s of thousands of metric producers. If you went to raw colocation facility costs - you’d have over a hundred dual Xeon machines with multi-TB direct attached SSD. Supporting hundreds of thousands of servers producing hundreds of millions of data points per second. Human time is really the main trade-off. | |
| ▲ | fragmede 2 days ago | parent | prev [-] | | Datadog has entered the chat. |
|
|
| ▲ | philomath_mn 3 days ago | parent | prev | next [-] |
| AI bills are already trivial compared to human time. I pay for claude max, all I need to do is save an hour a month and I will be breaking even. |
| |
| ▲ | oblio 2 days ago | parent | next [-] | | $200h * 8 * 5 * 4 * 12 = $384 000 per year. You're like in the top 0.05% of earners in the software field. Of course, if you save 10 hours per month, the math starts making more sense for others. And this is assuming LLM prices are stable, which I very much doubt they are, since everyone is price dumping to get market share. | | |
| ▲ | CMCDragonkai 2 days ago | parent | next [-] | | Across most anglosphere countries and tech cities - wages and salaries far outstrip what you can get for AI. AI is already objectively cheaper than human talent in rich countries. Is it as good? Yea I'd say it's better than most mid to junior engineers. Can it run entirely by itself? No, it still needs HITL. | | |
| ▲ | oblio 2 days ago | parent [-] | | Again, those prices aren't stable. Nobody is investing half a trillion in a tech without expecting a 10x return. And fairly sure soon those $20/month subscriptions will sell your data, shove ads everywhere AND basically only allow you to get that junior dev for 30 minutes per day or 2 days a month. And the $200/month will probably be $500-1000 with more limitations. Still cheap, but AI can't run an entire project, can't deliver. So the human will be in the loop, as you said, so at least a partial cost on top. | | |
| ▲ | Quinner 2 days ago | parent | next [-] | | The wages aren't stable either. There's going to be gradual convergence. | | |
| ▲ | oblio 2 days ago | parent [-] | | Oh, by the way, this entire discussion revolves around LLMs being an exponential tech. Real life only works with sigmoids. |
| |
| ▲ | CMCDragonkai 2 days ago | parent | prev | next [-] | | Not gonna happen. The competition for AI models is approaching commodity. | |
| ▲ | alwillis 2 days ago | parent | prev [-] | | What’s different is all the open weight models like Kimi-k2 or Qwen-3 Coder that are as good and, depending on the task, better than Anthropic’s Sonnet model for 80% less via openrouter [1] and other similar services. You can use these models through Claude Code; I do it everyday. Some developers are running smaller versions of these LLMs on their own hardware, paying no one. So I don’t think Anthropic and the other companies can dramatically increase their prices without losing the customers that helped them go from $0 to $4 billion in revenue in 3 years. Users can easily move between different AI platforms with no lock-in, which makes it harder to increase prices and proceed to enshitify their platforms. [1]: https://openrouter.ai/ |
|
| |
| ▲ | lumost 2 days ago | parent | prev [-] | | The percentages in the field are skewed, FAANG employ a vast number of engineers. | | |
| ▲ | oblio a day ago | parent [-] | | No, they don't. FAANG probably employs 400 000 programmers worldwide, and I think the US alone probably has about 3-4 million programmers. Worldwide there are probably 30 million. And even for FAANG, an SDE for them in Spain makes 60-100k total comp, not 400k. |
|
| |
| ▲ | byzantinegene 2 days ago | parent | prev [-] | | on the other hand, it could also you mean you are overpaid |
|
|
| ▲ | nextworddev 3 days ago | parent | prev [-] |
| You will start seriously worrying about coding AI bills within 6 months |
| |
| ▲ | alwillis 2 days ago | parent | next [-] | | You will start seriously worrying about coding AI bills within 6 months Nope. More open models ship everyday and are 80% cheaper for similar and sometimes better performance, depending on the task. You can use Qwen-3 Coder (a 480 billion parameter model with 35 billion active per forward pass (8 out of 160 experts)) for $0.302/M input tokens
$0.302/M output tokens via openrouter. Claude 4 Sonnet is $3/M input tokens and $15/M output tokens. Several utilities will let you use Claude Code to use these models at will. | | |
| ▲ | gorbypark 2 days ago | parent [-] | | Where are you seeing those prices? Cheapest one I see (and it has horrible uptime/throughput numbers) is $0.40/$1.60, and it's more like $2/$2 on average. |
| |
| ▲ | philomath_mn 3 days ago | parent | prev [-] | | Why is that? | | |
| ▲ | throwawayoldie 2 days ago | parent [-] | | Because at some point, Anthropic needs to stop hemorrhaging money and actually make some. | | |
| ▲ | fragmede 2 days ago | parent | next [-] | | Uber, founded in 2009 was finally profitable in 2023. That's a bit longer than 6 months. | | |
| ▲ | mvieira38 2 days ago | parent | next [-] | | Different situation. Uber's entire strategy was to "disrupt" the transportation industry by undercutting everyone, with the promise that eventually adoption (and monetization via data, advertising, etc.) would be large enough for fees not to rise so much as to push consumers and drivers into the established taxi industry. Anthropic, on the other hand, is a competitor brand in a brand-new industry, one with heavy reliance on capex and exploding employee salaries, for that matter. | | |
| ▲ | fragmede 2 days ago | parent [-] | | Totally different than except for the fact that it shows that "money" is able to wait 14 years, which is 28x longer than 6 months. |
| |
| ▲ | numbrss 2 days ago | parent | prev [-] | | [dead] |
| |
| ▲ | theshrike79 2 days ago | parent | prev [-] | | And when Anthropic raises the prices enough, people will jump ship. That's why you don't pay the yearly license for anything at this point in time. Pay monthly and evaluate before each bill if there's something better out already. | | |
| ▲ | nextworddev 2 days ago | parent [-] | | and that jumping ship didn't happen IRL, cursor ARR jumped 50% after their pricing change. |
|
|
|
|