▲ | phailhaus 7 days ago | ||||||||||||||||||||||
Feels like, but isn't. When you are reasoning things out, there is a brain with state that is actively modeling the problem. AI does no such thing, it produces text and then uses that text to condition the next text. If it isn't written, it does not exist. Put another way, LLMs are good at talking like they are thinking. That can get you pretty far, but it is not reasoning. | |||||||||||||||||||||||
▲ | Enginerrrd 6 days ago | parent | next [-] | ||||||||||||||||||||||
The transformer architecture absolutely keeps state information "in its head" so to speak as it produces the next word prediction, and uses that information in its compute. It's true that if it's not producing text, there is no thinking involved, but it is absolutely NOT clear that the attention block isn't holding state and modeling something as it works to produce text predictions. In fact, I can't think of a way to define it that would make that untrue... unless you mean that there isn't a system wherein something like attention is updating/computing and the model itself chooses when to make text predictions. That's by design, but what you're arguing doesn't really follow. Now, whether what the model is thinking about inside that attention block matches up exactly or completely with the text it's producing as generated context is probably at least a little dubious, and its unlikely to be a complete representation regardless. | |||||||||||||||||||||||
| |||||||||||||||||||||||
▲ | double0jimb0 7 days ago | parent | prev [-] | ||||||||||||||||||||||
So exactly what language/paradigm is this brain modeling the problem within? | |||||||||||||||||||||||
|