| ▲ | philipkglass 2 days ago | |
Do you have plans to do a follow-up model release with quantization aware training as was done for Gemma 3? https://developers.googleblog.com/en/gemma-3-quantized-aware... Having 4 bit QAT versions of the larger models would be great for people who only have 16 or 24 GB of VRAM. | ||