▲ | andix 5 days ago | |||||||
Yes, but for most things it's not significant. For example Stack Overflow used to handle all their traffic from 9 on-prem servers (not sure if this is still the case). Millions of daily users. Power consumption and hardware cost is completely insignificant in this case. LLM inference pricing is mostly driven by power consumption and hardware cost (which also takes a lot of power/heat to manufacture). | ||||||||
▲ | Twirrim 5 days ago | parent [-] | |||||||
> For example Stack Overflow used to handle all their traffic from 9 on-prem servers (not sure if this is still the case). Millions of daily users. Power consumption and hardware cost is completely insignificant in this case. They just finished their migration to the cloud, unracked their servers a few weeks ago https://stackoverflow.blog/2025/07/16/the-great-unracking-sa... | ||||||||
|