Tech News
← Back to articles

This is the fastest local AI I've tried, and it's not even close - how to get it

read original related products more articles

Jack Wallen / Elyse Betters Picaro / ZDNET

ZDNET's key takeaways

The gpt-oss:20b model is very fast.

You'll get blazing-fast answers to your queries with gpt-oss:20b.

With the latest version of Ollama installed, you can use this model.

Let's talk about local AI and speed. There are a lot of factors that go into getting the most speed out of your AI, such as:

Whether you have a dedicated GPU.

The context length you use (the smaller, the faster).

The complexity of your query.

The LLM you use.

... continue reading