▲ | Nevermark 5 days ago | ||||||||||||||||
Since models are given tokens, not letters, to process, the famous issues with counting letters is not indicative of incompetence. They are just sub-sensory for the model. None of us can reliably count the e’s as someone talks to us, either. | |||||||||||||||||
▲ | hatefulmoron 5 days ago | parent [-] | ||||||||||||||||
It does say something that the models simultaneously: a) "know" that they're not able to do it for the reason you've outlined (as in, you can ask about the limitations of LLMs for counting letters in words) b) still blindly engage with the query and get the wrong answer, with no disclaimer or commentary. If you asked me how many atoms there are in a chair, I wouldn't just give you a large natural number with no commentary. | |||||||||||||||||
|