Remix.run Logo
twoodfin 3 hours ago

That these data centers can turn electricity + a little bit of fairly simple software directly into consumer and business value is pretty much the whole story.

Compare what you need to add to AWS EC2 to get the same result, above and beyond the electricity.

zozbot234 3 hours ago | parent [-]

That's a convenient story, but most consumers' and businesses' use of AI is light enough that they could easily run local models on their existing silicon. Resorting to proprietary AI running in the datacenter would only add a tiny fraction of incremental value over that, and at a significant cost.

astral_drama 2 hours ago | parent | next [-]

I'm looking forward to running a Gemma 4 turboquant on my 24GB GPU. The perf looks impressive for how compact it is.

I often get a 10x more cost effective run processing on my local hardware.

Still reaching for frontier models for coding, but find the hosted models on open router good enough for simple work.

Feels like we are jumping to warp on flops. My cores are throttled and the fiber is lit.

twoodfin 3 hours ago | parent | prev [-]

Sure but where the puck is going is long-running reasoning agents where local models are (for the moment) significantly constrained relative to a Claude Opus 4.6.