| ▲ | ACCount37 2 days ago | |
There appears to be a degree of "introspection of its own confidence" in modern LLMs. They can identify their own hallucinations, at a rate significantly better than chance. So there must be some sort of "do I recall this?" mechanism built into them. Even if it's not exactly a reliable mechanism. Anthropic has discovered that this is definitely the case for name recognition, and I suspect that names aren't the only things subject to a process like that. | ||