WebLLM Β· Mobile + Desktop Β· WebGPU

Local AI Chat

Run open-source LLMs entirely in your browser. No server, no API keys β€” powered by WebLLM.

Qwen3 Β· Llama Β· Gemma Streaming Markdown
β–Ύ
Initializing engine…
Model is cached after first download.
W
WebLLM Chat
Ready
Temp0.70
Top-P0.95
Tokens
Rep1.10

Start a conversation

Type a message below.
Everything runs locally in your browser.