| ▲ | hatefulmoron 5 days ago | |||||||
It does say something that the models simultaneously: a) "know" that they're not able to do it for the reason you've outlined (as in, you can ask about the limitations of LLMs for counting letters in words) b) still blindly engage with the query and get the wrong answer, with no disclaimer or commentary. If you asked me how many atoms there are in a chair, I wouldn't just give you a large natural number with no commentary. | ||||||||
| ▲ | Nevermark 4 days ago | parent [-] | |||||||
That is interesting. A factor might be that they are trained to behave like people who can see letters. During training they have no ability to not comply, and during inference they have no ability to choose to operate differently than during training. A pre-prompt or co-prompt that requested they only answer questions about sub-token information if they believed they actually had reason to know the answer, would be a better test. | ||||||||
| ||||||||