| ▲ | fooker 6 hours ago | |||||||
LLMs specifically are fine with random bits flipped for the results to be 'creative'. | ||||||||
| ▲ | jedberg 6 hours ago | parent [-] | |||||||
That's not exactly how LLM temperature works. :). Also that's on inference, not training. Presumably these would be used for training, the latency would be too high for inference. | ||||||||
| ||||||||