▲ | JSR_FDED 4 days ago | |||||||||||||||||||||||||
Which of these can I run locally on a 64GB Mac Mini Pro? And how much does quantization affect the quality? | ||||||||||||||||||||||||||
▲ | simonw 4 days ago | parent [-] | |||||||||||||||||||||||||
I use a 64GB M2 MacBook Pro. I tend to find any model that's smaller than 32B works well (I can just about run a 70B but it's not worth it as I have to quit all other apps first). My current favorite to run on my machine is OpenAI's gpt-oss-20b because it only uses 11GB of RAM and it's designed to run at that quantization size. I also really like playing with the Qwen 3 family at various sizes and I'm fond of Mistral Small 3.2 as a vision LLM that works well. | ||||||||||||||||||||||||||
|