Remix.run Logo
hypercube33 2 hours ago

All we need is something like Qwen3-coder-next but at Kimi K2.6 ability so it runs on laptop workstation hardware and we are set...soon?

wolttam an hour ago | parent | next [-]

In 2023 GPT-4 was allegedly 1.8T parameters. In 2026 we have ~100x smaller models (10-20B) that handily outperform it, and can indeed run on a laptop.

rectang 8 minutes ago | parent [-]

How does "outperform" translate to the propensity of an LLM to hallucinate?

operatingthetan 5 minutes ago | parent [-]

There seems to be a mass delusion about how capable SOTA models actually are. That's my only explanation for how poorly I find them performing in basic knowledge tasks compared to how others describe their prowess.

unshavedyak 36 minutes ago | parent | prev [-]

I am eagerly awaiting being able to run a strong local model. I'd hand Apple $5k right now for a Claude in a box. I know the cost might not be there now, just saying that is around my ideal price point.

$10k might even be worth it - but i'm assuming that the more expensive it is the beefier it is too, which also means more electricity.. and i already run ~6 computers/servers in my house. If a power surge happens i'm going to go live in the woods lol.

atonse 13 minutes ago | parent | next [-]

I would do the same but my issue is that the models are changing so fast, so I don't want to be left out of the next model cuz it only runs on an even newer GPU or something like that.

But maybe my limited understanding is thinking of this wrong.

DANmode 24 minutes ago | parent | prev [-]

You can run 6-12 month old state of the art models for that type of money,

like, yesterday.