| ▲ | ru552 7 hours ago | |||||||
You won't like it, but the answer is Apple. The reason is the unified memory. The GPU can access all 32gb, 64gb, 128gb, 256gb, etc. of RAM. An easy way (napkin math) to know if you can run a model based on it's parameter size is to consider the parameter size as GB that need to fit in GPU RAM. 35B model needs atleast 35gb of GPU RAM. This is a very simplified way of looking at it and YES, someone is going to say you can offload to CPU, but no one wants to wait 5 seconds for 1 token. | ||||||||
| ▲ | samtheprogram 7 hours ago | parent | next [-] | |||||||
That estimate doesn't account for context, which is very important for tool use and coding. I used this napkin math for image generation, since the context (prompts) were so small, but I think it's misleading at best for most uses. | ||||||||
| ▲ | sliken 5 hours ago | parent | prev [-] | |||||||
> You won't like it, but the answer is Apple. Or strix halo. Seems rather over simplified. The different levels of quants, for Qwen3.6 it's 10GB to 38.5GB. Qwen supports a context length of 262,144 natively, but can be extended to 1,010,000 and of course the context length can always be shortened. Just use one of the calculators and you'll get much more useful number. | ||||||||
| ||||||||