▲ | hatefulmoron a day ago | |||||||
The LLM's output is chaotic relative to the input, but it's deterministic right? Same settings, same model, same input, .. same output? Where does the chain get broken here? | ||||||||
▲ | tsimionescu 18 hours ago | parent | next [-] | |||||||
Depends on what you mean specifically by the output. The actual neural network will produce deterministic outputs that could be interpreted as probability values for various tokens. But the interface you'll commonly see used in front of these models will then non-deterministiclaly choose a single next token to output based on those probabilities. Then, this single randomly chosen output is fed back into the network to produce another token, and this process repeats. I would ultimately call the result non-deterministic. You could make it deterministic relatively easily by having a deterministic process for choosing a single token from all of the outputs of the NN (say, always pick the one with the highest weight, and if there are multiple with the same weight, pick the first one in token index order), but no one normally does this, because the results aren't that great per my understanding. | ||||||||
| ||||||||
▲ | fc417fc802 a day ago | parent | prev [-] | |||||||
Now compare a human to an LSTM with persistent internal state that you can't reset. | ||||||||
|