| ▲ | refulgentis 2 days ago | |||||||
Reversing the X and Y axis, adding in a few other random models, and dropping all the small Qwens makes this worse than useless as a Qwen 3.5 comparison, it’s actively misleading. If you’re using AI, please don’t rush to copy paste output :/ EDIT: Lordy, the small models are a shadow of Qwen's smalls. See https://huggingface.co/Qwen/Qwen3.5-4B versus https://www.reddit.com/r/LocalLLaMA/comments/1salgre/gemma_4... | ||||||||
| ▲ | scrlk 2 days ago | parent | next [-] | |||||||
I transposed the table so that it's readable on mobile devices. I should have mentioned that the Qwen 3.5 benchmarks were from the Qwen3.5-122B-A10B model card (which includes GPT-5-mini and GPT-OSS-120B); apologies for not including the smaller Qwen 3.5 models. | ||||||||
| ||||||||
| ▲ | BloondAndDoom a day ago | parent | prev [-] | |||||||
Small qwen models are magical | ||||||||
| ||||||||