| ▲ | rvz 2 days ago | ||||||||||||||||
It is more than good enough and has effectively caught up with Opus 4.6 and GPT 5.4 according to the benchmarks. It's about 2 months behind GPT 5.5 and Opus 4.7. As long as it is cheap to run for the hosting providers and it is frontier level, it is a very competitive model and impressive against the others. I give it 2 years maximum for consumer hardware to run models that are 500B - 800B quantized on their machines. It should be obvious now why Anthropic really doesn't want you to run local models on your machine. | |||||||||||||||||
| ▲ | deaux 2 days ago | parent | next [-] | ||||||||||||||||
Vibes > Benchmarks. And it's all so task-specific. Gemini 3 has scored very well in benchmarks for very long but is poor at agentic usecases. A lot of people prefering Opus 4.6 to 4.7 for coding despite benchmarks, much more than I've seen before (4.5->4.6, 4->4.5). Doesn't mean Deepseek v4 isn't great, just benchmarks alone aren't enough to tell. | |||||||||||||||||
| ▲ | snovv_crash 2 days ago | parent | prev | next [-] | ||||||||||||||||
With the ability of the Qwen3.6 27B, I think in 2 years consumers will be running models of this capability on current hardware. | |||||||||||||||||
| ▲ | colordrops 2 days ago | parent | prev [-] | ||||||||||||||||
What's going to change in 2 years that would allow users to run 500B-800B parameter models on consumer hardware? | |||||||||||||||||
| |||||||||||||||||