| ▲ | meander_water 4 hours ago | |||||||
This looks like it uses Gemini Nano under the hood. But the latest Gemma4 E2B and E4B models appear to be much better, so you'd probably be better off deploying quantized versions through an extension for now. - Gemini Nano-1: 46% MMLU, 1.8B - Gemini Nano-2: 56% MMLU, 3.25B - Gemma4 E2B: 60.0% MMLU, 2.3B - Gemma4 E4B: 69.4% MMLU, 4.5B Sources: - https://huggingface.co/google/gemma-4-E2B-it - https://android-developers.googleblog.com/2024/10/gemini-nan... | ||||||||
| ▲ | domenicd 3 hours ago | parent [-] | |||||||
I no longer have any inside knowledge, but from my time on this team they were very quick about getting the latest small (Google) models into Chrome. I expect that if Gemma 4 (or its equivalent Gemini Nano) isn't already in Chrome, then it will be soon. Note that the article here was last updated 2025-09-21, and as of that time it was already on Gemini Nano 3. | ||||||||
| ||||||||