▲ | Jensson 5 days ago | |
> Have the LLM talk about what “truth” is and the nature of LLM hallucinations and it can cook up an explanation that demonstrates it completely understands the concepts. This isn't how LLM works. What an LLM understands has nothing to do with the words they say, it only has to do with what connections they have seen. If an LLM has only seen a manual but has never seen examples of how the product is used, then it can tell you exactly how to use the product by writing out info from the manual, but if you ask it to do those things then it wont be able to, since it has no examples to go by. This is the primary misconception most people have and make them over estimate what their LLM can do, no they don't learn by reading instructions they only learn by seeing examples and then doing the same thing. So an LLM talking about truth just comes from it having seen others talk about truth, not from it thinking about truth on its own. This is fundamentally different to how humans think about words. | ||
▲ | ninetyninenine 5 days ago | parent [-] | |
>This isn't how LLM works. I know how an LLM works. I've built one. At best we only know surface level stuff like the fact that it involves a feed forward network and is using token prediction. But the emergent effect of how it an LLM produces an overall statement that reflects high level conceptual understanding is something we don't know. So your claim of "This isn't how an LLM works" which was said which such confidence is utterly wrong. You don't know how it works, no one does. |