| ▲ | utilize1808 7 hours ago | ||||||||||||||||
Obviously going to depend on your definition of "decent". My impression so far is that you will need between 90GB to 100GB of memory to run medium sized (31B dense or ~110B MoE) models with some quantization enabled. | |||||||||||||||||
| ▲ | cjbgkagh 7 hours ago | parent [-] | ||||||||||||||||
I’m running Gemma4 31B (Q8) on my 2 4090s (48GB) with no problem. | |||||||||||||||||
| |||||||||||||||||