Remix.run Logo
glemion43 13 hours ago

They are not just statistical models

They create concepts in latent space which is basically compression which forces this

jrmg 7 hours ago | parent | next [-]

You’re describing a complex statistical model.

glemion43 4 hours ago | parent [-]

Debatable I would argue. It's definitely not 'just a statistical model's and I would argue that the compression into this space fixes potential issues differently than just statistics.

But I'm not a mathematics expert if this is the real official definition I'm fine with it. But are you though?

mmooss 6 hours ago | parent | prev [-]

What is "latent space"? I'm wary of metamagical descriptions of technology that's in a hype cycle.

DoctorOetker 5 hours ago | parent | next [-]

its a statistical term, a latent variable is one that is either known to exist, or believed to exist, and then estimated.

consider estimating the position of an object from noisy readings. One presumes that position to exist in some sense, and then one can estimate it by combining multiple measurements, increasing positioning resolution.

its any variable that is postulated or known to exist, and for which you run some fitting procedure

AIorNot 5 hours ago | parent | prev | next [-]

See this video

https://youtu.be/D8GOeCFFby4?si=AtqH6cmkOLvqKdr0

glemion43 4 hours ago | parent | prev [-]

I'm disappointed that you had to add the 'metamagical' to your question tbh

It doesn't matter if ai is in a hype cycle or not it doesn't change how a technology works.

Check out the yt videos from 1blue3brown he explains LLMs quite well. .your first step is the word embedding this vector space represents the relationship between words. Father - grandfather. The vector which makes a father a grandfather is the same vector as mother to grandmother.

You the use these word vectors in the attention layer to create a n dimensional space aka latent space which basically reflects a 'world' the LLM walks through. This makes the 'magic' of LLMs.

Basically a form of compression by having higher dimensions reflecting kind a meaning.

Your brain does the same thing. It can't store pixels so when you go back to some childhood environment like your old room, you remember it in some efficient (brain efficient) way. Like the 'feeling' of it.

That's also the reason why an LLM is not just some statistical parrot.

mmooss 2 hours ago | parent [-]

> It doesn't matter if ai is in a hype cycle or not it doesn't change how a technology works.

It does change what people say about it. Our words are not reality itself; the map is not the territory.

Are you saying people should take everything said about LLMs at face value?

glemion43 29 minutes ago | parent [-]

Being dismissive of technical terms on hn because something seems to be a hype is really weird.

It's the reason why I'm here because we discuss more technically about technology