| ▲ | tjwebbnorfolk 16 hours ago | |
All of gemma's main competitors have larger models in the 80-240b range that take advantage of larger VRAM GPUs and dual-GPU setups. Personally I have 2x RTX 6000 PROs and right now am running the 235b-parameter Qwen model with very good results. I also occasionally use gpt-oss:120b. I would like to see a gemma model in the same range. Also many people are running these on Mac Minis now with 128GB+ of unified RAM. Aiming for the "runs on a single H100" tagline doesn't make a lot of sense to me, because most people do not have H100s anyway. | ||