▲ | binary132 6 hours ago | |||||||
sure would be neat if these companies would release models that could run on consumer hardware | ||||||||
▲ | edude03 5 hours ago | parent | next [-] | |||||||
So there are two ways to look at this - both hinge on how your define "consumer": 1) We haven't managed to distill models enough to get good enough performance to fit in the typical gaming desktop (say, 7B-24b class models). Even then though - most consumers don't have high end desktops, so even a 3060 class GPU requirement would exclude a lot of people. 2) Nothing is stopping you/anyone from buying 24ish 5090s (a consumer hardware product) to get the required ~600GB-1TB of VRAM to run unquantized deepseek except time/money/know how. Sure, it's unreasonably expensive but it's not like labs are conspiring to prevent people from running these models, it's just expensive for everyone and the common person doesn't have the funding to get into it. | ||||||||
| ||||||||
▲ | __mharrison__ 5 hours ago | parent | prev [-] | |||||||
I'm using Qwen3Next on my MBP. It uses around 42GB of memory and, according to Aider benchmarks, has similar perf to GPT-4.1 https://huggingface.co/mlx-community/Qwen3-Next-80B-A3B-Inst... |