| ▲ | icedchai 2 hours ago | |||||||
The quality of local models is still abysmal compared to commercial SOTA models. You're not going to run something like Gemini or Claude locally. I have some "serious" hardware with 128G of VRAM and the results are still laughable. If I moved up to 512G, it still wouldn't be enough. You need serious hardware to get both quality and speed. If I can get "quality" at a couple tokens a second, it's not worth bothering. They are getting better, but that doesn't mean they're good. | ||||||||
| ▲ | _aavaa_ 2 hours ago | parent [-] | |||||||
Good by what standard? Compared to SOTA today? No they're not. But they are better than the SOTA in 2020, and likely 2023. We have a magical pseudo-thinking machine that we can run locally completely under our control, and instead the goal posts have moved to "but it's not as fast as the proprietary could". | ||||||||
| ||||||||