▲ | imtringued 6 days ago | |
Look you don't have to lie at every opportunity you get. You are fully aware and know what you've written is bullshit. Tokens are a highly specific transformer exclusive concept. The human brain doesn't run a byte pair encoding (BPE) tokenizer [0] in their head. anything as tokens. It uses asynchronous time varying spiking analog signals. Humans are the inventors of human languages and are not bound to any static token encoding scheme, so this view of what humans do as "token prediction" requires either a gross misrepresentation of what a token is or what humans do. If I had to argue that humans are similar to anything in machine learning research specifically, I would have to argue that they extremely loosely follow the following principles: * reinforcement learning with the non-brain parts defining the reward function (primarily hormones and pain receptors) * an extremely complicated non-linear kalman filter that not only estimates the current state of the human body, but also "estimates" the parameters of a sensor fusing model * there is a necessary projection of the sensor fused result that then serves as available data/input to the reinforcement learning part of the brain Now here are two big reasons why the model I describe is a better fit: The first reason is that I am extremely loose and vague. By playing word games I have weaseled myself out of any specific technology and am on the level of concepts. The second reason is that the kalman filter concept here is general enough that it also includes predictor models, but the predictor model here is not the output that drives human action, because that would logically require the dataset to already contain human actions, which is what you did, you assume that all learning is imitation learning. In my model, any internal predictor model that is part of the kalman filter is used to collect data, not drive human action. Actions like eating or drinking are instead driven by the state of the human body, e.g. hunger is controlled through leptin and insulin and others. All forms of work, no matter how much of a detour it represents, ultimately has the goal of feeding yourself or your family (=reproduction). [0] A BPE tokenizer is a piece of human written software that was given a dataset to generate an efficient encoding scheme and the idea itself is completely independent of machine learning and neural networks. The fundamental idea behind BPE is that you generate a static compression dictionary and never change it. | ||
▲ | zahlman 5 days ago | parent [-] | |
> Look you don't have to lie at every opportunity you get. You are fully aware and know what you've written is bullshit. As much as I may agree with your subsequent claims, this is not how users are expected to engage with each other on HN. |