▲ | jeremyjh 2 days ago | |||||||
It doesn’t know what it doesn’t know. | ||||||||
▲ | fallpeak 2 days ago | parent | next [-] | |||||||
It doesn't know that because it wasn't trained on any tasks that required it to develop that understanding. There's no fundamental reason an LLM couldn't learn "what it knows" in parallel with the things it knows, given a suitable reward function during training. | ||||||||
▲ | binarymax 2 days ago | parent | prev | next [-] | |||||||
Well sure. But maybe the token logprobs can be used to help give a confidence assessment. | ||||||||
| ||||||||
▲ | smt88 2 days ago | parent | prev [-] | |||||||
That's not true for all types of questions. You've likely seen a model decline to answer a question that requires more recent training data than it has, for example. |