▲ | amelius 5 days ago | ||||||||||||||||
But humans we can feed ascii, whereas LLMs require token inputs. My original question was about that: why can't we just feed the LLMs ascii, and let it figure out how it wants to encode that internally, __implicitly__? I.e., we just design a network and feed it ascii, as opposed to figuring out an encoding in a separate step and feeding it tokens in that encoding. | |||||||||||||||||
▲ | cesarb 5 days ago | parent [-] | ||||||||||||||||
> But humans we can feed ascii, whereas LLMs require token inputs. To be pedantic, we can't feed humans ASCII directly, we have to convert it to images or sounds first. > My original question was about that: why can't we just feed the LLMs ascii, and let it figure out how it wants to encode that internally, __implicitly__? I.e., we just design a network and feed it ascii, as opposed to figuring out an encoding in a separate step and feeding it tokens in that encoding. That could be done, by having only 256 tokens, one for each possible byte, plus perhaps a few special-use tokens like "end of sequence". But it would be much less efficient. | |||||||||||||||||
|