▲ | shadowgovt 6 days ago | |||||||
It's fundamentally the wrong tool to get factual answers from because the training data doesn't have signal for factual answers. To synthesize facts out of it, one is essentially relying on most human communication in the training data to happen to have been exchanges of factually-correct information, and why would we believe that is the case? | ||||||||
▲ | astrange 5 days ago | parent | next [-] | |||||||
Because people are paying the model companies to give them factual answers, so they hire data labellers and invent verification techniques to attempt to provide them. Even without that, there's implicit signal because factual helpful people have different writing styles and beliefs than unhelpful people, so if you tell the model to write in a similar style it will (hopefully) provide similar answers. This is why it turns out to be hard to produce an evil racist AI that also answers questions correctly. | ||||||||
▲ | lblume 6 days ago | parent | prev [-] | |||||||
Empirically, there seems to be strong evidence for LLMs giving factual output for accessible knowledge questions. Many benchmarks test this. | ||||||||
|