Tech News
← Back to articles

Nvidia Launches Family of Open Reasoning AI Models: OpenReasoning Nemotron

read original related products more articles

GTC—NVIDIA today announced the open Llama Nemotron family of models with reasoning capabilities, designed to provide developers and enterprises a business-ready foundation for creating advanced AI agents that can work independently or as connected teams to solve complex tasks.

Built on Llama models, the NVIDIA Llama Nemotron reasoning family delivers on-demand AI reasoning capabilities. NVIDIA enhanced the new reasoning model family during post-training to improve multistep math, coding, reasoning and complex decision-making.

This refinement process boosts accuracy of the models by up to 20% compared with the base model and optimizes inference speed by 5x compared with other leading open reasoning models. The improvements in inference performance mean the models can handle more complex reasoning tasks, enhance decision-making capabilities and reduce operational costs for enterprises.

Leading agent AI platform pioneers — including Accenture , Amdocs , Atlassian, Box , Cadence , CrowdStrike , Deloitte, IQVIA , Microsoft, SAP and ServiceNow — are collaborating with NVIDIA on its new reasoning models and software.

“Reasoning and agentic AI adoption is incredible,” said Jensen Huang, founder and CEO of NVIDIA. “NVIDIA’s open reasoning models, software and tools give developers and enterprises everywhere the building blocks to create an accelerated agentic AI workforce.”

NVIDIA Post-Training Boosts Accuracy and Reliability for Enterprise Reasoning

Built to deliver production-ready AI reasoning, the Llama Nemotron model family is available as NVIDIA NIM™ microservices in Nano, Super and Ultra sizes — each optimized for different deployment needs.

The Nano model delivers the highest accuracy on PCs and edge devices, the Super model offers the best accuracy and highest throughput on a single GPU, and the Ultra model will provide maximum agentic accuracy on multi-GPU servers.

NVIDIA conducted extensive post-training on NVIDIA DGX™ Cloud using high-quality curated synthetic data generated by NVIDIA Nemotron™ and other open models, as well as additional curated datasets cocreated by NVIDIA.

The tools, datasets and post-training optimization techniques used to develop the models will be openly available, giving enterprises the flexibility to build their own custom reasoning models.

... continue reading