| ▲ | mrob 10 hours ago | |
It's plausible that LLMs experience things during training, but during inference an LLM is equivalent to a lookup table. An LLM is a pure function mapping a list of tokens to a set of token probabilities. It needs to be connected to a sampler to make it "chat", and each token of that chat is calculated separately (barring caching, which is an implementation detail that only affects performance). There is no internal state. | ||
| ▲ | Kim_Bruning 10 hours ago | parent [-] | |
Right, no hidden internal state. Exactly. There's 0. And the weights are sitting there statically, which is absolutely true. But my current favorite frontier model has this 1 million token mutable state just sitting there. Holding natural language. Which as we know can encode emotions. (Which I imagine you might demonstrate on reading my words, and then wisely temper in your reply) | ||