▲ | regularfry 5 hours ago | |
> 1) We haven't managed to distill models enough to get good enough performance to fit in the typical gaming desktop (say, 7B-24b class models). That really depends on what "good enough" means. Qwen3-30b runs absolutely fine at q4 on a 24GB card, although that's also stretching "typical gaming desktop". It's competent as a code completion or aider-type coding agent model in that scenario. But really we need both. Yes it would be nice to have things targeted to our own particular niche, but there are only so many labs cranking these things out. Small models will only get better from here. |