How to Run a 1.7B Parameter LLM in Your Browser With WebGPU
You've got a 1.7 billion parameter model. You want it running locally. In a browser tab. No server, no API keys, no Docker containers.
Sounds impossible, right? A few months ago, I would've agreed with you. But 1-bit quantized models like Bonsai 1.7B...
alan-west.hashnode.dev7 min read