Remix.run Logo
ffsm8 3 hours ago

Please don't oversell them. Eg Kimi k2.6 has a maximum context size of 270k, that's a quarter of opus.

The model is fine, Ive switched to it entirely for a personal project, but it's not opus.

And no, you're not running then locally unless you're a millionaire. You still need hundreds of GB (500+++) of VRAM on your graphics card - that's not at a level of consumer electronics.

Sure you can run the quantized models, but then you're at Haiku performance.

2ndorderthought 3 hours ago | parent [-]

Qwen 3.6 runs in a single GPU. But I mostly agree with you except, just because a model has a given context doesn't mean it's all available or entirely reliable.