| ▲ | AussieWog93 13 hours ago | |
I've tried these small models and they're nowhere near as good as Claude or GPT-5. The new ones running on a 16GB M1 are maybe GPT-4 level (with decent performance to be fair). I wonder if it's possible to make some hyper-overturned model that, say, does nothing but program in Python get SOTA-ish performance in that narrow task. | ||