61.
62.
Post-transformer inference: 224× compression of Llama-70B with improved accuracy
(news.ycombinator.com)
63.
Vsora Jotunn-8 5nm European inference chip
(news.ycombinator.com)
64.
Principles of Vasocomputation
(news.ycombinator.com)
65.
Cloud-Native Computing Is Poised To Explode
(slashdot.org)
67.
68.
Ovi: Twin backbone cross-modal fusion for audio-video generation
(news.ycombinator.com)
69.
Ovi
(news.ycombinator.com)
70.
Elixir 1.19
(news.ycombinator.com)
71.
Cerebras systems raises $1.1B Series G
(news.ycombinator.com)
72.
Cerebras Systems Raises $1.1B Series G at $8.1B Valuation
(news.ycombinator.com)
73.
GPT-OSS Reinforcement Learning
(news.ycombinator.com)
74.
Show HN: Run Qwen3-Next-80B on 8GB GPU at 1tok/2s throughput
(news.ycombinator.com)
75.
Defeating Nondeterminism in LLM Inference
(news.ycombinator.com)
76.
Some users report their Firefox browser is scoffing CPU power
(news.ycombinator.com)
77.
Token growth indicates future AI spend per dev
(news.ycombinator.com)
78.
Running GPT-OSS-120B at 500 tokens per second on Nvidia GPUs
(news.ycombinator.com)
79.
80.
My favorite use-case for AI is writing logs
(news.ycombinator.com)
81.
LLM Inference Handbook
(news.ycombinator.com)
82.
I extracted the safety filters from Apple Intelligence models
(news.ycombinator.com)
83.
Tools: Code Is All You Need
(news.ycombinator.com)
84.
The inference trap: How cloud providers are eating your AI margins
(venturebeat.com)
85.
How runtime attacks turn profitable AI into budget black holes
(venturebeat.com)
86.
87.
88.
OpenInfer raises $8M for AI inference at the edge
(venturebeat.com)