| ▲ | root_axis 6 hours ago | ||||||||||||||||||||||
> At some point a beefy Mac Studio and the "right sized" model is going to be what people want. It's pretty clear that this isn't going to happen any time soon, if ever. You can't shrink the models without destroying their coherence, and this is a consistently robust observation across the board. | |||||||||||||||||||||||
| ▲ | sipjca 6 hours ago | parent [-] | ||||||||||||||||||||||
I don’t think it’s about literally shrinking the models via quantization, but rather training smaller/more efficient models from scratch Smaller models have gotten much more powerful the last 2 years. Qwen 3.5 is one example of this. The cost/compute requirements of running the same level intelligence is going down | |||||||||||||||||||||||
| |||||||||||||||||||||||