| ▲ | ffsm8 3 hours ago | |
Please don't oversell them. Eg Kimi k2.6 has a maximum context size of 270k, that's a quarter of opus. The model is fine, Ive switched to it entirely for a personal project, but it's not opus. And no, you're not running then locally unless you're a millionaire. You still need hundreds of GB (500+++) of VRAM on your graphics card - that's not at a level of consumer electronics. Sure you can run the quantized models, but then you're at Haiku performance. | ||
| ▲ | 2ndorderthought 3 hours ago | parent [-] | |
Qwen 3.6 runs in a single GPU. But I mostly agree with you except, just because a model has a given context doesn't mean it's all available or entirely reliable. | ||