llm

A small language model running entirely in your browser via WebGPU. No server, no API key. Your prompt never leaves your machine.

This first version is running WebLLM on WebGPU, with a SmolLM2-1.7B from HuggingFace. This chat is single turn, input->output.

← trucs.ai