▲ | blutfink 5 days ago | |||||||||||||||||||||||||||||||||||||||||||
The LLM can also “say” “cat” with few bits. Note that the meaning of the word as stored in your brain takes more than 24 bits. | ||||||||||||||||||||||||||||||||||||||||||||
▲ | amelius 5 days ago | parent [-] | |||||||||||||||||||||||||||||||||||||||||||
No, an LLM really uses __much__ more bits per token. First, the embedding typically uses thousands of dimensions. Then, the value along each dimension is represented with a floating point number which will take 16 bits (can be smaller though with higher quantization). | ||||||||||||||||||||||||||||||||||||||||||||
|