▲ | eamsen 2 days ago | |
Completely agree with this statement. I would go further, and say we don't understand how next-token predictors work either. We understand the model structure, just as we do with the brain, but we don't have a complete map of the execution patterns, just as we do not with the brain. Predicting the next token can be as trivial as a statistical lookup or as complex as executing a learned reasoning function. My intuition suggests that my internal reasoning is not based on token sequences, but it would be impossible to convey the results of my reasoning without constructing a sequence of tokens for communication. |