▲ | cubefox 5 days ago | |||||||
His solution still relies on greedy (temperature 0) sampling, which is probably not optimal for model performance on various tasks. For example, Gemini 2.5 uses temperature 1 by default. But deterministic inference with temperature >0 can still be achieved by using pseudorandom sampling with a fixed seed. | ||||||||
▲ | red2awn 5 days ago | parent | next [-] | |||||||
Conceptually setting temperature to be >0 doesn't actually introduce any non-determinism. If your sampler is seeded then it will always choose the same next token. Higher temperature only flattens the logit distribution. | ||||||||
▲ | mynameismon 5 days ago | parent | prev [-] | |||||||
The point of the blog is that even at "supposed" deterministic generative sampling, non-determinism creeps in. This in turn has disastrous effects in very real experiments. | ||||||||
|