DeepSeek-V4: Towards Highly Efficient Million-Token Context Intelligence
(news.ycombinator.com)
1.
2.
3.
MSA: Memory Sparse Attention
(news.ycombinator.com)
4.
DeepSeek tests “sparse attention” to slash AI processing costs
(arstechnica.com)
5.
Using uninitialized memory for fun and profit (2008)
(news.ycombinator.com)