Remix.run Logo
jibal a day ago

There are no "real-world concepts" or "semantic meaning" in LLMs, there are only syntactic relationships among text tokens.

lblume a day ago | parent | next [-]

That really stretches the meaning of "syntactic". Humans have thoroughly evaluated LLMs and discovered many patterns that very cleanly map to what they would consider real-world concepts. Semantic properties do not require any human-level understanding; a Python script has specific semantics one may use to discuss its properties, and it has become increasingly clear that LLMs can reason (as in derive knowable facts, extract logical conclusions, compare it to different alternatives; not having a conscious thought process involving them) about these scripts not just by their syntactic but also semantic properties (of course bounded and limited by Rice's theorem).

jibal 18 hours ago | parent [-]

> Humans have thoroughly evaluated LLMs and discovered many patterns that very cleanly map to what they would consider real-world concepts

Well yes, humans have real-world concepts.

> Semantic properties do not require any human-level understanding

Strawman.

> a Python script has specific semantics one may use to discuss its properties

These are human-attributed semantics. To say that a static script "has" semantics is a category mistake--certainly it doesn't "have" them the way LLMs are purported by the OP to have concepts.

> it has become increasingly clear that LLMs can reason (as in derive knowable facts, extract logical conclusions, compare it to different alternatives

These are highly controversial claims. LLMs present conclusions textually that are implicit in the training data. To get from there to the claim that they can reason is a huge leap. Certainly we know (from studies by Anthropic and elsewhere) that the reasoning steps that LLMs claim to go through are not actual states of the LLM.

I'm not going to say more about this ... it has been discussed at length in the academic literature.

int_19h a day ago | parent | prev | next [-]

If that were true, homonyms would be an intractable challenge to LLMs, yet they can handle them just fine, and do so in tasks that require understanding of their semantics (e.g. give LLM a long text and require it to catalog all uses of the word "right" segregated into buckets according to meaning).

jibal 18 hours ago | parent [-]

That's a syntactic pattern matching operation.

empath75 a day ago | parent | prev [-]

Do you learn anything from reading books or is everything you know entirely derived from personal experience.

jibal 18 hours ago | parent [-]

Non sequitur. I learn things from LLMs because the LLM training data contains a vast amount of information. This has nothing to do with whether LLMs have concepts.