Remix.run Logo
zamadatix a day ago

It feels like the days of Claude 2 -> 3 or GPT 2->3 level changes for the leading models are over and you're either going to end up with really awkward version numbers or just embrace it and increment the number. Nobody cares a Chrome update gives a major version change of 136->137 instead of 12.4.2.33 -> 12.4.3.0 for similar kinds of "the version number doesn't always have to represent the amount of work/improvement compared to the previous" reasoning.

saubeidl a day ago | parent | next [-]

It feels like LLM progress in general has kinda stalled and we're only getting small incremental improvements from here.

I think we've reached peak LLM - if AGI is a thing, it won't be through this architecture.

nico a day ago | parent | next [-]

Diffusion LLMs seem like they could be a huge change

Check this out from yesterday (watch the short video here):

https://simonwillison.net/2025/May/21/gemini-diffusion/

From:

https://news.ycombinator.com/item?id=44057820

antupis a day ago | parent [-]

Yup especially locally diffusion models will be big.

bcrosby95 a day ago | parent | prev | next [-]

Even if LLMs never reach AGI, they're good enough to where a lot of very useful tooling can be built on top of/around them. I think of it more as the introduction of computing or the internet.

That said, whether or not being a provider of these services is a profitable endeavor is still unknown. There's a lot of subsidizing going on and some of the lower value uses might fall to the wayside as companies eventually need to make money off this stuff.

michaelbrave a day ago | parent [-]

this was my take as well. Though after a while I've started thinking about it closer to the introduction of electricity which in a lot of ways would be considered the second stage of the industrial revolution, the internet and AI might be considered the second stage of the computing revolution (or so I expect history books to label it as). But just like electricity, it doesn't seem to be very profitable for the providers of electricity, but highly profitable for everything that uses it.

a day ago | parent [-]
[deleted]
jenny91 a day ago | parent | prev | next [-]

I think it's a bit early to say. At least in my domain, the models released this year (Gemini 2.5 Pro, etc). Are crushing models from last year. I would therefore not by any means be ready to call the situation a stall.

goatlover a day ago | parent | prev [-]

Which brings up the question of why AGI is a thing at all. Shouldn't LLMs just be tools to make humans more productive?

amarcheschi a day ago | parent [-]

Think of the poor vcs who are selling agi as the second coming of christ

onlyrealcuzzo a day ago | parent | prev [-]

Did you see Gemini 1.5 pro vs 2.5 pro?

zamadatix a day ago | parent [-]

Sure, but despite there being a 2.0 release between they didn't even feel the need to release a Pro for it still isn't the kind of GPT 2 -> 3 improvement we were hoping would continue for a bit longer. Companies will continue to release these incremental improvements which are all always neck-and-neck with each other. That's fine and good, just don't inherently expect the versioning to represent the same relative difference instead of the relative release increment.

cma a day ago | parent [-]

I'd say 2.5 pro to 1.5 pro was a 3 -> 4 level improvement, but the problem is 1.5 pro wasn't state of the art when released, except for context length, and 2.5 wasn't that kind of improvement compared to the best open AI or Claude stuff that was available when it released.

1.5 pro was worse than original gpt4 on several coding things I tried head to head.