Remix.run Logo
otabdeveloper4 6 hours ago

LLMs don't think at all.

Forcing it to be concise doesn't work because it wasn't trained on token strings that short.

rafram 5 hours ago | parent | next [-]

They’re able to solve complex, unstructured problems independently. They can express themselves in every major human language fluently. Sure, they don’t actually have a brain like we do, but they emulate it pretty well. What’s your definition of thinking?

otabdeveloper4 3 hours ago | parent [-]

When OP wrote about LLMs "thinking" he implied that they have an internal conceptual self-reflecting state. Which they don't, they *are* merely next token predicting statistical machines.

rafram 2 hours ago | parent [-]

This was true in 2023.

fkgmeqnb an hour ago | parent [-]

And it still is today.

HumanOstrich 6 hours ago | parent | prev [-]

> Forcing it to be concise doesn't work because it wasn't trained on token strings that short.

This is a 2023-era comment and is incorrect.

Barbing 2 hours ago | parent | next [-]

Anything I can read that would settle the debate?

otabdeveloper4 5 hours ago | parent | prev [-]

LLMs architectures have not changed at all since 2023.

> but mmuh latest SOTA from CloudCorp (c)!

You don't know how these things work and all you have to go on is marketing copy.

HumanOstrich 3 hours ago | parent [-]

Yea you don't know anything about LLM architectures. They often change with each model release.

You also aren't aware that there's more to it than "LLM architecture". And you're rather confident despite your lack of knowledge.

You're like the old LLMs before ChatGPT was released that were kinda neat, but usually wrong and overconfident about it.