Nvidia just admitted the general-purpose GPU era is ending
(venturebeat.com)
61.
62.
Five Things to Know About Nvidia’s $20 Billion Licensing Deal
(feeds.content.dowjones.io)
63.
64.
65.
Post-transformer inference: 224× compression of Llama-70B with improved accuracy
(news.ycombinator.com)
66.
Vsora Jotunn-8 5nm European inference chip
(news.ycombinator.com)
67.
The Easiest Way to Build a Type Checker
(news.ycombinator.com)
68.
Principles of Vasocomputation
(news.ycombinator.com)
69.
Cloud-Native Computing Is Poised To Explode
(slashdot.org)
70.
Realizing value with AI inference at scale and in production
(technologyreview.com)
72.
73.
Ovi: Twin backbone cross-modal fusion for audio-video generation
(news.ycombinator.com)
74.
Ovi
(news.ycombinator.com)
75.
Elixir 1.19
(news.ycombinator.com)
76.
Cerebras systems raises $1.1B Series G
(news.ycombinator.com)
77.
Cerebras Systems Raises $1.1B Series G at $8.1B Valuation
(news.ycombinator.com)
78.
GPT-OSS Reinforcement Learning
(news.ycombinator.com)
79.
Show HN: Run Qwen3-Next-80B on 8GB GPU at 1tok/2s throughput
(news.ycombinator.com)
80.
Defeating Nondeterminism in LLM Inference
(news.ycombinator.com)
81.
Some users report their Firefox browser is scoffing CPU power
(news.ycombinator.com)
82.
Token growth indicates future AI spend per dev
(news.ycombinator.com)
83.
Running GPT-OSS-120B at 500 tokens per second on Nvidia GPUs
(news.ycombinator.com)
84.
85.
My favorite use-case for AI is writing logs
(news.ycombinator.com)
86.
LLM Inference Handbook
(news.ycombinator.com)
87.
I extracted the safety filters from Apple Intelligence models
(news.ycombinator.com)
88.
Tools: Code Is All You Need
(news.ycombinator.com)
89.
The inference trap: How cloud providers are eating your AI margins
(venturebeat.com)
90.
How runtime attacks turn profitable AI into budget black holes
(venturebeat.com)