Running local models on Macs gets faster with Ollama's MLX support
(arstechnica.com)
1.
2.
LLM Neuroanatomy II: Modern LLM Hacking and Hints of a Universal Language?
(news.ycombinator.com)
3.
Flash-MoE: Running a 397B Parameter Model on a Laptop
(news.ycombinator.com)
4.
MacBook M5 Pro and Qwen3.5 = Local AI Security System
(news.ycombinator.com)
5.
Qwen3.5-397B at 4.74 tok/s using 5.9GB RAM
(news.ycombinator.com)
6.
Qwen3.5 122B and 35B models offer Sonnet 4.5 performance on local computers
(news.ycombinator.com)
7.
Qwen3.5: Towards Native Multimodal Agents
(news.ycombinator.com)