| ▲ | wyldfire 2 hours ago | |
Is that really the case? This summer there was "Frontier AI performance becomes accessible on consumer hardware within a year" [1] which makes me think it's a mistake to discount the open weights models. | ||
| ▲ | hu3 an hour ago | parent | next [-] | |
Open weight models are neat. But for SOTA performance you need specialized hardware. Even for Open Weight models. 40k in consumer hardware is never going to compete with 40k of AI specialized GPUs/servers. Your link starts with: > "Using a single top-of-the-line gaming GPU like NVIDIA’s RTX 5090 (under $2500), anyone can locally run models matching the absolute frontier of LLM performance from just 6 to 12 months ago." I highly doubt a RTX 5090 can run anything that competes with Sonnet 3.5 which was released June, 2024. | ||
| ▲ | cmrdporcupine 18 minutes ago | parent | prev [-] | |
With RAM prices spiking, there's no way consumers are going to have access to frontier quality models on local hardware any time soon, simply because they won't fit. That's not the same as discounting the open weight models though. I use DeepSeek 3.2 heavily, and was impressed by the Devstral launch recently. (I tried Kimi K2 and was less impressed). I don't use them for coding so much as for other purposes... but the key thing about them is that they're cheap on API providers. I put $15 into my deepseek platform account two months ago, use it all the time, and still have $8 left. I think the open weight models are 8 months behind the frontier models, and that's awesome. Especially when you consider you can fine tune them for a given problem domain... | ||