| ▲ | throwawaymaths 5 days ago |
| that's wrong. there is probably a categorical difference between making something up due to some sort of inferential induction from the kv cache context under the pressure of producing a token -- any token -- and actually looking something up and producing a token. so if you ask, "what is the capital of colorado" and it answers "denver" calling it a Hallucination is nihilistic nonsense that paves over actually stopping to try and understand important dynamics happening in the llm matrices |
|
| ▲ | saghm 5 days ago | parent | next [-] |
| > so if you ask, "what is the capital of colorado" and it answers "denver" calling it a Hallucination is nihilistic nonsense that paves over actually stopping to try and understand important dynamics happening in the llm matrices On the other hand, calling it anything other than a hallucination misrepresents the idea of truth as being something that these models have any ability to differentiate between their outputs based on whether they accurately reflect reality by conflating a fundamentally unsolved problem as an engineering tradeoff. |
| |
| ▲ | ComplexSystems 5 days ago | parent [-] | | It isn't a hallucination because that isn't how the term is defined. The term "hallucination" refers, very specifically, to "plausible but false statements generated by language models." At the end of the day, the goal is to train models that are able to differentiate between true and false statements, at least to a much better degree than they can now, and the linked article seems to have some very interesting suggestions about how to get them to do that. | | |
| ▲ | player1234 3 days ago | parent | next [-] | | Why use a word that you have to redefine the meaning of? The answer is to deceive. | |
| ▲ | throwawaymaths 5 days ago | parent | prev [-] | | your point is good and taken but i would amend slightly -- i dont think that "absolute truth" is itself a goal, but rather "how aware is it that it doesn't know something". this negative space is frustratingly hard to capture in the llm architecture (though almost certainly there are signs -- if you had direct access to the logits array, for example) |
|
|
|
| ▲ | mannykannot 5 days ago | parent | prev | next [-] |
| There is a way to state Parson's point which avoids this issue: hallucinations are just as much a consequence of the LLM working as designed as are correct statements. |
| |
| ▲ | throwawaymaths 5 days ago | parent [-] | | fine. which part is the problem? | | |
| ▲ | mannykannot 4 days ago | parent | next [-] | | I suppose you are aware that, for many uses of LLMs, the propensity for hallucinating is a problem (especially if this is not properly taken into account by the people hoping to use these LLMs), but this then leaves me puzzled about what you are asking here. | |
| ▲ | johnnyanmac 5 days ago | parent | prev [-] | | The part where it can't admit situations where there's not enough data/training to admit it doesn't know. I'm a bit surprised no one talks about this factor. It's like talking to a giant narcissist who can Google really fast but not understand what it reads. The ability to admit ignorance is a major factor of credibility, because none of us know everything all at once. | | |
|
|
|
| ▲ | littlestymaar 5 days ago | parent | prev [-] |
| > that's wrong. Why would anyone respond with so little nuance? > a Hallucination Oh, so your shift key wasn't broken all the time, then why aren't you using it in your sentences? |