Nvidia Rubin Nvidia
Follow ZDNET: Add us as a preferred source on Google.
ZDNET's key takeaways
Nvidia's new platform aims to reduce the cost of training LLMs.
It uses six AI chips to lower token costs and GPU requirements.
The first platforms will roll out to partners later in the year.
The last several years have been stupendous for Nvidia. When generative AI became all the rage, demand for the tech giant's hardware skyrocketed as companies and developers scrambled for its graphics cards to train their large language models (LLMs). During CES 2026, Nvidia held a press conference to unveil its latest innovation in the AI space: the Rubin platform.
Also: CES 2026: Biggest TV, smart glasses, phone news, and more we've seen so far
Nvidia announced what the technology can do, and it's all pretty dense, so to keep things concise, I'm only focusing on the highlights.
Rubin is an AI supercomputing platform designed to make "building, deploying, and securing the world's largest and most advanced AI systems at the lowest cost" possible. According to Nvidia, the platform can deliver up to a 10x reduction in inference token costs and requires four times fewer graphics cards to train mixture-of-experts (MoE) models compared to the older Blackwell platform.
... continue reading