Remix.run Logo
binkHN 2 hours ago

> So now I joke that even if I have a claude max plan, I've still come out ahead financially.

This is no joke; for better or worse, I see a day when I'm paying a lot more for this and it will be a bargain.

wolttam 2 hours ago | parent | next [-]

By my estimation (guess) you won't actually need to spend that much because the models are already getting a point where they don't need to get a whole lot better to be extremely helpful across many domains.

And it looks like those very helpful capabilities will continue to transfer to smaller models as well, as architectures and training regimes continue to refine.

I can fairly easily imagine a world where the only people needing to spend a lot of money on models are those that are using them to solve truly novel problems. The rest of us will get plenty of use at reasonable costs for the typical day-to-day helpful stuff.

rectang 20 minutes ago | parent | next [-]

In my anecdotal experience there is a huge gap between GPT-5-mini which hallucinates relentlessly and Claude Opus or the latest GPTs which are fairly reliable. I'm hoping that gap can be closed with improved approaches for small models and that good reliability is achievable for LLMs without requiring absolutely mammoth computing resources.

For what it's worth, I also used GPT-5.2 (via duck.ai) this year for questions about taxes and it was helpful — which makes sense because there's so much material about taxes out there to be synthesized, so a text predictor trained in that domain should do well.

hypercube33 2 hours ago | parent | prev | next [-]

All we need is something like Qwen3-coder-next but at Kimi K2.6 ability so it runs on laptop workstation hardware and we are set...soon?

wolttam an hour ago | parent | next [-]

In 2023 GPT-4 was allegedly 1.8T parameters. In 2026 we have ~100x smaller models (10-20B) that handily outperform it, and can indeed run on a laptop.

rectang 9 minutes ago | parent [-]

How does "outperform" translate to the propensity of an LLM to hallucinate?

operatingthetan 6 minutes ago | parent [-]

There seems to be a mass delusion about how capable SOTA models actually are. That's my only explanation for how poorly I find them performing in basic knowledge tasks compared to how others describe their prowess.

unshavedyak 38 minutes ago | parent | prev [-]

I am eagerly awaiting being able to run a strong local model. I'd hand Apple $5k right now for a Claude in a box. I know the cost might not be there now, just saying that is around my ideal price point.

$10k might even be worth it - but i'm assuming that the more expensive it is the beefier it is too, which also means more electricity.. and i already run ~6 computers/servers in my house. If a power surge happens i'm going to go live in the woods lol.

atonse 15 minutes ago | parent | next [-]

I would do the same but my issue is that the models are changing so fast, so I don't want to be left out of the next model cuz it only runs on an even newer GPU or something like that.

But maybe my limited understanding is thinking of this wrong.

DANmode 26 minutes ago | parent | prev [-]

You can run 6-12 month old state of the art models for that type of money,

like, yesterday.

Barbing an hour ago | parent | prev [-]

[sci-fi “AGI” scenario] What if those with elite model access philosophize in a way us mere mortals can’t understand, so the elites have to prechew the ideas for us to bring them to our level, and they control the narrative?

In reality now, curious about social implications generally. Does this go beyond problem solving? Maybe the intelligence per token you get via your free library card/membership is insufficient to compete with peers in dating/employment/etc. markets, thus puts you at disadvantage.

unixhero 20 minutes ago | parent [-]

That isn't really philosophy, but rather doom and gloom theories. Control the narrative on what exactly, how I write a bootstrap script for my servers? Or what type of flower is in this photo. Not everything is politics luckily.

baq 2 hours ago | parent | prev [-]

I've seen this day sometime in December and not only with Claude. Wish I was joking on some days, feeling exhilarated on others.