▲ | Havoc 2 days ago | |||||||||||||
Is there a reason why the 1.58 ones are always aimed at quite small ones? Think I’ve seen an 8B but that’s about it. Is there a technical reason for it or just research convenience ? | ||||||||||||||
▲ | londons_explore 2 days ago | parent | next [-] | |||||||||||||
I suspect because current GPU hardware can't efficiently train such low bit depth models. You end up needing activations to use 8 or 16 bits in all the data paths, and don't get any more throughput per cycle on the multiplications than you would have done with FP32. Custom silicon would solve that, but nobody wants to build custom silicon for a data format that will go out of fashion before the production run is done. | ||||||||||||||
| ||||||||||||||
▲ | yieldcrv 2 days ago | parent | prev [-] | |||||||||||||
They aren’t, there is a 1.58 version of deepseek that’s like 200gb instead of 700 | ||||||||||||||
|