Chrome is experimentally shipping with Gemini Nano, their smallest Large Language Model (LLM) baked right in, then offer APIs to use it.
In Chrome, these APIs are built to run inference against Gemini Nano with fine-tuning or an expert model. Designed to run locally on most modern devices, Gemini Nano is best for language-related use cases, such as summarization, rephrasing, or categorization.
It’s an API most, with methods you call and get responses. Raymond Camden had a look:
const model = await window.ai.createTextSession();
await model.prompt("Who are you?");
// I am a large language model, trained by Google.
Code language: JavaScript (javascript)
Using AI in this way means 1) it’s fast (no network trip) 2) it works offline 3) it’s private (maybe) 4) it’s free to use.
I admit that’s awfully compelling. I suspect this will happen and will be very highly used.
Don’t we need to think about standards here though? What if Apple ships window.ai.instantiateIntelligence()
with an .ask()
method? And Firefox ships navigator.llm('dolly').enqueueQuery()
? I’d just like to remind everyone that when browser just ship whatever and compete on proprietary features, everybody loses.