Remix.run Logo
skeledrew 2 hours ago

Its really just a matter of degrees. There are 1 million, 1 million, 1 trillion parameter LLMs... and you keep scaling those parameters and you eventually get to humans. But it's still probable next tokens (decisions) based on previous tokens (experience).

skissane 40 minutes ago | parent | next [-]

> Its really just a matter of degrees. There are 1 million, 1 million, 1 trillion parameter LLMs... and you keep scaling those parameters and you eventually get to humans.

It isn’t because humans and current LLMs have radically different architectures

LLMs: training and inference are two separate processes; weights are modifiable during training, static/fixed/read-only at runtime

Humans: training and inference are integrated and run together; weights are dynamic, continuously updated in response to new experiences

You can scale current LLM architectures as far as you want, it will never compete with humans because it architecturally lacks their dynamism

Actually scaling to humans is going to require fundamentally new architectures-which some people are working on, but it isn’t clear if any of them have succeeded yet

simonh 2 hours ago | parent | prev | next [-]

They’re both neural networks, but the architectures built using those neural connections, and the way they are trained and operate are completely different. There are many different artificial neural network architectures. They’re not all LLMs.

AlphaZero isn’t a LLM. There are Feed Forward networks, recurrent networks, convolutional networks, transformer networks, generative adversarial networks.

Brains have many different regions each with different architectures. None of them work like LLMs. Not even our language centres are structured or trained anything like LLMs.

coldtea an hour ago | parent [-]

>AlphaZero isn’t a LLM. There are Feed Forward networks, recurrent networks, convolutional networks, transformer networks, generative adversarial networks.

That's irrelevant though, since all the above are still prediction machines based on weights.

If you're ok with the brain being that, then you just changed the architecture (from LLM-like), not the concept.

trinsic2 2 hours ago | parent | prev [-]

LOL. Oook.. No i dont think so. The human experience and the mechanisms behind it have a lot of unknowns and im pretty sure that trying to confine the human experience into the amount of parameters there are is short sighted.