This is the fastest local AI I've tried, and it's not even close - how to get it
Jack Wallen / Elyse Betters Picaro / ZDNET ZDNET's key takeaways The gpt-oss:20b model is very fast. You'll get blazing-fast answers to your queries with gpt-oss:20b. With the latest version of Ollama installed, you can use this model. Let's talk about local AI and speed. There are a lot of factors that go into getting the most speed out of your AI, such as: Whether you have a dedicated GPU. The context length you use (the smaller, the faster). The complexity of your query. The LLM you