Tech News
← Back to articles

MDST Engine: run GGUF models in the browser with WebGPU/WASM

read original related products more articles

5 min read Share

TLDR: MDST brings GGUF to WebGPU, the most popular format for LLMs, so anyone can create, edit, and review any files and collaborate from their browser without being dependent on cloud LLM providers or complicated setups.

In 2026, more people want local models that they can actually run and trust, and the hardware and software are finally catching up. Better consumer GPUs, new models and better quantizations are making “local” feel normal and accessible as never before.

So we built a WASM/JS engine that can run GGUF on WebGPU. The GGUF format is one of the most popular LLM formats and supports various quantizations. Shipped in a single-file container, it is best for consumer-grade devices and easy to download, cache, and tune.

We believe this will open a new, bigger market for GGUF: fast, local inference for anyone who just wants it to work, right in the browser.