logoalt Hacker News

meander_wateryesterday at 9:02 AM3 repliesview on HN

This looks like it uses Gemini Nano under the hood. But the latest Gemma4 E2B and E4B models appear to be much better, so you'd probably be better off deploying quantized versions through an extension for now.

- Gemini Nano-1: 46% MMLU, 1.8B

- Gemini Nano-2: 56% MMLU, 3.25B

- Gemma4 E2B: 60.0% MMLU, 2.3B

- Gemma4 E4B: 69.4% MMLU, 4.5B

Sources:

- https://huggingface.co/google/gemma-4-E2B-it

- https://android-developers.googleblog.com/2024/10/gemini-nan...


Replies

domenicdyesterday at 9:18 AM

I no longer have any inside knowledge, but from my time on this team they were very quick about getting the latest small (Google) models into Chrome. I expect that if Gemma 4 (or its equivalent Gemini Nano) isn't already in Chrome, then it will be soon.

Note that the article here was last updated 2025-09-21, and as of that time it was already on Gemini Nano 3.

show 2 replies
ceejayozyesterday at 2:23 PM

> This looks like it uses Gemini Nano under the hood.

Yes; "With the Prompt API, you can send natural language requests to Gemini Nano in the browser."

Tepixyesterday at 2:38 PM

The Prompt API uses the model that's available in your browser. For Edge I believe it's Phi4.