| ▲ | traceroute66 2 days ago | |
> Compute costs fall reliably, so what matters is the capability at a given price point in 18 months, not today. The underlying point still stands, namely that "more compute" as the default answer is not sustainable. Why ? Because even if we accept the unlikely dream that GPU prices will magically take a nose-dive, you still need somewhere to put all those servers stuffed with GPUs. That means datacentres. And "more datacentres" is absolutely not sustainable. The cooling needs, the power needs, the land needs..... none of it is remotely sustainable. | ||
| ▲ | refulgentis 2 days ago | parent [-] | |
"datacenters are unsustainable" is a different conversation entirely. The premise that inference compute must scale linearly with capability isn't supported by what's actually happening. Distillation, quantization, and architectural efficiency gains routinely let you run yesterday's frontier capability at a fraction of the cost and hardware. GPT-3.5 level performance runs on a phone now. The 100M token budget Mythos used here will not require 100M tokens worth of 2026 hardware forever. On datacenters specifically: yes, they require power, cooling, and land. So does every other piece of industrial infrastructure humanity has ever built and then incrementally made more efficient. The energy per FLOP has been dropping for decades and continues to. You can argue the rate of buildout is concerning, and that's a reasonable discussion. But "not sustainable" as a flat declaration requires you to believe efficiency gains will stall, energy production won't expand, and cooling technology will stay static, all simultaneously. That's a much stronger claim than it sounds. None of which has anything to do with whether the AISI results are significant. They are. | ||