| ▲ | JKCalhoun 7 hours ago | |
"16-bit BF16 69.4 GB" Is that (BF16) a 16-bit float? | ||
| ▲ | adrian_b 44 minutes ago | parent | next [-] | |
The IEEE standard FP16 is an older 16-bit format, which has balanced exponent and significand sizes. It has been initially supported by GPUs, where it is useful especially for storing the color components of pixels. For geometry data, FP32 is preferred. In CPUs, some support has been first added in 2012, in Intel Ivy Bridge. Better support is provided in some server CPUs, and since next year also in the desktop AMD Zen 6 and Intel Nova Lake. BF16 is a format introduced by Google, intended only for AI/ML applications, not for graphics, so initially it was implemented in some of the Intel server CPUs and only later in GPUs. Unlike FP16, which is balanced, BF16 has great dynamic range, but very low precision. This is fine for ML but inappropriate for any other applications. Nowadays, most LLMs are trained preponderantly using BF16, with a small number of parameters using FP32, for higher precision. Then from the biggest model that uses BF16, smaller quantized models are derived, which use 8 bits or less per parameter, trading off accuracy for speed. | ||
| ▲ | mtklein 6 hours ago | parent | prev | next [-] | |
Yes, it's a "Brain float", basically an ordinary 32-bit float with the low 16 mantissa bits cut off. Exact same range as fp32, lower precision, and not the same as the other fp16, which has less exponent and more mantissa. | ||
| ▲ | Gracana 6 hours ago | parent | prev | next [-] | |
https://en.wikipedia.org/wiki/Bfloat16_floating-point_format Yes, however it’s a different format from standard fp16, it trades precision for greater dynamic range. | ||
| ▲ | 6 hours ago | parent | prev | next [-] | |
| [deleted] | ||
| ▲ | WithinReason 6 hours ago | parent | prev [-] | |
yes, it has 8 exponent bits like float32 instead of 6 like float16 | ||