▲ | ekelsen 5 days ago | |||||||||||||
The math on the input tokens is definitely wrong. It claims each instance (8 GPUs) can handle 1.44 million tokens/sec of input. Let's check that out. 1.44e6 tokens/sec * 37e9 bytes/token / 3.3e12 bytes/sec/GPU = ~16,000 GPUs And that's assuming a more likely 1 byte per parameter. So the article is only off by a factor of at least 1,000. I didn't check any of the rest of the math, but that probably has some impact on their conclusions... | ||||||||||||||
▲ | thatguysaguy 5 days ago | parent | next [-] | |||||||||||||
37 billion bytes per token? Edit: Oh assuming this is an estimate based on the model weights moving fromm HBM to SRAM, that's not how transformers are applied to input tokens. You only have to do move the weights for every token during generation, not during "prefill". (And actually during generation you can use speculative decoding to do better than this roofline anyways). | ||||||||||||||
| ||||||||||||||
▲ | GaggiX 5 days ago | parent | prev | next [-] | |||||||||||||
Your calculations make no sense. Why are you loading the model for each token independently? You can process all the input tokens at the same time as long as they can fit in memory. You are doing the calculation as they were output tokens on a single batch, it would not make sense even in the decode phase. | ||||||||||||||
| ||||||||||||||
▲ | 5 days ago | parent | prev | next [-] | |||||||||||||
[deleted] | ||||||||||||||
▲ | endtime 5 days ago | parent | prev | next [-] | |||||||||||||
> 37e9 bytes/token This doesn't quite sound right...isn't a token just a few characters? | ||||||||||||||
| ||||||||||||||
▲ | Lionga 5 days ago | parent | prev [-] | |||||||||||||
Well he asked some AI to do the math for him probably |