▲ | ElFitz 5 days ago | ||||||||||||||||||||||
Even when instructed to say "I don’t know" it is just as likely to make up an answer instead, or say it "doesn’t know" when the data is actually present somewhere in its weights. | |||||||||||||||||||||||
▲ | codeflo 5 days ago | parent [-] | ||||||||||||||||||||||
That's because the architecture isn't built for it to know what it knows. As someone put it, LLMs always hallucinate, but for in-distribution data they mostly hallucinate correctly. | |||||||||||||||||||||||
|