WebLLM

AI Just Got Standardized

A Boring Standard That Changes Everything

WebLLM is an open protocol that makes AI universal. Users choose their provider. Developers skip the infrastructure. The web gets AI everywhere.

Decentralized
Provider-agnostic
Privacy by design
Straightforward API
Agentic capabilities
User-controlled
Affordable

AI should be a utility.
Instead it's a mess.

💸

Users pay 5x

Same model, different apps, multiple bills

⚙️

Developers rebuild

Identical infrastructure, over and over

🏛️

Wikipedia can't afford

What individual users pay for

🔒

3 companies control

How we get access to AI

10 Reasons whyBrowser-Native AIis inevitable

1

End the AI Tax

Stop paying multiple times for the same technology

Users pay for GPT-4 five times across different apps. One subscription should work everywhere — like having one internet bill, not paying each website separately.

After: Pay once → Use everywhere

$75/mo for same model$20/mo everywhere
AI
ChatGPT$20
N
Notion AI$10
G
Grammarly$30
Cursor$15

Total: $75/mo

Same GPT-4 model

2

Democratize AI Access

Everyone deserves AI, not just the venture-funded

AI shouldn't be a luxury for funded startups. When users bring their own AI, every website can be intelligent.

Only if you have $10k/monthEvery website, any budget
W
Wikipedia8B views
Open SourceAI enabled
Schools$0 AI budget
Local BizSmart site

All AI-enabled now

3

We already own the hardware

Make use of the computers people already have

Millions of powerful GPUs and M-series Macs sit idle while their owners pay for cloud AI. That's backwards.

Privacy bonus: Your data never leaves your device

Speed bonus: Often faster than API calls

Your hardware finally wakes up and works for you

RTX 4090

82 TFLOPS AI

Before5% used
After80% used

M3 Max

40 TFLOPS AI

Before3% used
After70% used
4

Eliminate Infrastructure Redundancy

Stop rebuilding the same plumbing

Every company spends weeks building identical infrastructure: auth, billing, rate limiting, error handling. Collective waste in the millions of developer hours.

Ship features, not infrastructure

Before

auth
billing
rate limit
tokens
fallback
errors

4 weeks

After

navigator.llm

4 lines

5

Future-Proof Development

Write once, evolve forever

Stop hardcoding model names that become obsolete in 6 months. Describe what you need, let the system evolve.

When GPT-5 or Claude 4 launches, your code automatically uses it. No refactoring. No deployment.

Standards outlive products

Hardcoded approach:

gpt-3.5gpt-4gpt-4-turbogpt-4ogpt-5?

Constant refactoring

Standard approach:

navigator.llm
forever

Auto-upgrades to best model

6

Enable True Client-Side AI

AI at the speed of thought

Every AI call shouldn't require a round-trip to your server, which often just relay requests to other AI services. Tools should execute where the data lives — in the browser.

Instant tools: Manipulate DOM, update state immediately

Real-time: Forms that complete as you think

The backend was always unnecessary overhead

Traditional:

200ms
500ms
AI

~700ms round trip

With WebLLM:

direct
AI

~10ms instant

7

Guarantee Privacy by Design

Data sovereignty for users

Privacy shouldn't depend on corporate promises. When AI runs locally, sensitive data never leaves your control.

Privacy by architecture, not by policy

Your Device
🏥

Medical

Local only

💰

Financial

Local only

📄

Work docs

Local only

💭

Personal

Local only

upload to cloud
8

Restore User Agency

Users control their tools, not vice versa

You shouldn't need permission to use your preferred AI. Your tools should adapt to you, not lock you in.

Switch anytime: No app breakage when you change

Mix and match: Different AI for different tasks

Your preferences, your choice, everywhere

Your AI preferences:

C

Claude

Primary choice

G

GPT-4

Backup

G

Gemini

Available

Works on every website

9

Level the Playing Field

Small teams can compete with giants

Infrastructure advantage disappears when users bring their own AI. Now it's about ideas, not capital.

Open source: Compete with commercial products

Students: Build without credit cards

David and Goliath use the same AI

BigCorp

1000 devs

Solo Dev

1 person

Same AI capabilities

10

Standards Always Win

Proprietary products become universal utilities

The web thrives on open standards. AI should be no different — universal, interoperable, and user-controlled.

Standards boring. Standardization revolutionary.

Flash
<video>
ActiveX
fetch()
Silverlight
WebGL
APIs
navigator.llm

527

Developers building

10,241

Users connected

142

Commits this week

This isn't a product launch.
It's a protocol shift.

What Developers Are Saying

See what the community thinks about our LLM Gateway

Just integrated WebLLM into my app. One API, works with OpenAI, Claude, and local models. My users can pick their own provider. This is how AI should work on the web.
Been running Ollama locally + WebLLM extension. Zero API costs, complete privacy, and it just works on every website. The future of browser AI is here.
navigator.llm is the API the web needed. No more building auth, billing, rate limiting for every AI feature. Just call the API and let users bring their own models.
WebLLM is doing for AI what the web did for documents. A standard protocol that works everywhere. No vendor lock-in. User-controlled. This is big.
Switched from hardcoding OpenAI to WebLLM. Now my users can use GPT-4, Claude, Gemini, or even local Llama. Same code. Mind = blown.
WebLLM solves the open source AI problem. Projects can add AI features without paying for API calls. Users bring their own models. Everyone wins.
My M3 Mac was just sitting there while I paid OpenAI. Now with WebLLM + Ollama, I'm running AI locally for free. Why did I wait so long?
WebLLM follows the same pattern as other successful web APIs. Permission-based, provider-agnostic, user-controlled. This could be the next navigator API.

Works with your AI Providers

Ollama
OpenAI
Anthropic
Google AI
Meta
Mistral AI
Hugging Face
Perplexity
Ollama
OpenAI
Anthropic
Google AI
Meta
Mistral AI
Hugging Face
Perplexity

Plus Chrome's built-in Gemini Nano, and many more through OpenAI-compatible APIs

Partners

Sponsor WebLLM for a tight-knit partnership

Become a Partner
AlignableSourcegraphEcosiaVercelAnthropicOpenAI
Hugging FaceMozillaCloudflare

Significant sponsors, current and past (smaller logos)

They already use WebLLM

Discover how developers are building intelligent applications with browser-native AI.

I
Creativity

Inspire Cards

AI-powered card interpretation for brainstorming, idea generation, card games, and activities like tarot reading. Let users explore and discuss cards with conversational AI.

The Roadmap

Done

Chrome extension alpha

Core protocol design

10+ providers integrated

Popular SDK integration

Doing

Model routing system

Configuration console

Request playground

Demo projects

Next
After

Sharable computing tokens

Desktop Installer

Cloud compute program

Partners announcement

Image, audio & midi models

Better Tool calling

Model management & download tools

Future

WebGPU acceleration for local models

P2P resource pooling

Web standards alignements

Mobile SDK (iOS/Android)

standards always win

FAQ

Common Questions & Answers

Find out all the essential details about WebLLM and how it can serve your needs.

1

What is WebLLM and how does it work?

WebLLM is an open protocol that brings AI directly to your browser. Instead of each app having its own AI integration, WebLLM provides a standard SDK that lets users bring their own AI provider to any website.

2

Do I need to pay for multiple AI subscriptions?

No! That's exactly the problem WebLLM solves. Instead of paying $20/month to ChatGPT, $10 to Notion AI, $30 to Grammarly — all using the same underlying model — you pay once and use it everywhere through WebLLM.

3

Is my data private when using WebLLM?

Privacy is built into the architecture. You control which AI provider handles your data. For maximum privacy, you can run models locally on your own hardware — your data never leaves your device.

4

Does WebLLM work with Chrome's built-in AI?

Yes! Chrome 138+ includes Gemini Nano — free, on-device AI for desktop browsers. WebLLM automatically uses it when available, with graceful fallback to other providers. Users with Chrome AI get free local AI; everyone else uses their configured provider.

5

How do I get started as a developer?

It's just a few lines of code. Install the WebLLM SDK, create a session, and start making requests. No API keys to manage, no billing to set up — WebLLM handles it all through the user's configured provider.

6

Which AI providers are supported?

WebLLM supports 30+ providers including OpenAI, Anthropic, Google, Mistral, Groq, Together.ai, OpenRouter, Chrome's built-in Gemini Nano, and local options like Ollama and LM Studio. Users choose their preferred provider and can switch anytime.

7

Is WebLLM free to use?

WebLLM itself is completely free and open source. Users pay only for their chosen AI provider (or nothing if using Chrome AI or local models). Developers can add AI features to their apps without any infrastructure costs.

Help Build the Web

AI as a utility needs builders, users, advocates, and believers.
How will you help build the grid?

I Build Things
Implement WebLLM in your app.
It's easier than you think.

→ See Quick Start (4 lines of code)

→ View Examples

→ Join Discord

Start Building
I Use Things
Install WebLLM. Vote with your usage.
Make AI work like it should.

→ Add to Chrome

→ Configure Your AI

→ Browse Normally

Install Extension
I Influence Things
Ask sites to support WebLLM.
Share what's possible.

→ Tweet Template

→ Email Template

→ Success Stories

Share the Movement
I Believe in This
Fund the project.
Accelerate the future.

This is a community effort, privately funded.

We believe it has the potential to disrupt the market and shift value to small players and individual users.

Support WebLLM

Utilities aren't built by companies.
They're built by communities.

Are you in?