NVIDIA Nemotron
NVIDIA Nemotron™ is a family of open models with open weights, training data, and recipes, delivering leading efficiency and accuracy for building specialized AI agents.
NVIDIA Nemotron Models
Nemotron models are transparent—the training data used for these models, as well as their weights, are open and available on Hugging Face for you to evaluate before deploying them in production. The technical reports outlining the steps necessary to recreate these models are also freely available.
The new Nemotron 3 family provides the most efficient open models, powered by hybrid Mamba‑Transformer MoE with 1M-token context, delivering top accuracy for complex, high-throughput agentic AI applications.
Easily deploy models using open frameworks like vLLM, SGLang, Ollama and llama.cpp on any NVIDIA GPUs — from the edge and cloud to the data center. Endpoints are also available as NVIDIA NIM™ microservices for easy deployment on any GPU-accelerated system.
Nemotron reasoning models are optimized for various platforms:
Nano provides cost efficiency with high accuracy for targeted agentic tasks
Super delivers high accuracy for multi-agentic reasoning
Ultra is designed for applications demanding highest reasoning accuracy
Additionally, these models provide highest throughput, enabling agents to think faster and generate higher-accuracy response while lowering inference cost.
Nemotron 3 Nano 30B A3B
- Nemotron 3 Nano offers 4x faster throughput compared to Nemotron 2 Nano
- Leading accuracy for coding, reasoning, math and long context tasks
- Perfect for agents that need to deliver highest accuracy and efficiency for targeted tasks
Llama Nemotron Super 49B
- High in-class accuracy and throughput
- Great for efficient deep research agents
- Suitable for single data center GPU deployments
Llama Nemotron Ultra 253B
- Ideal for multi-agent enterprise workflows requiring highest accuracy, such as customer service automation, supply chain management, and IT security
- Suitable for data center-scale deployments
Nemotron Nano VL 12B
- Ideal for multi-agent enterprise workflows requiring highest accuracy, such as customer service automation, supply chain manageBest-in-class vision-language accuracyment, and IT security
- Designed for document intelligence and video understanding
Suitable for single data center GPU deployments
Nemotron RAG
- Industry-leading extraction, embedding, and reranking models
- Best-in-class accuracy for multimodal document intelligence, question answering, and passage retrieval
Nemotron Safety
- Safety models for advanced jailbreak detection, multilingual content safety with cultural nuance and reasoning capabilities, and topic control for secure and compliant LLM operations globally
NVIDIA Nemotron Datasets
Improve reasoning capabilities of large language models (LLMs) with one of the largest open collections of synthetic data for agentic AI. With over 9T tokens of pre- and post-training data, the collection spans across math, coding, scientific knowledge, function calling, instruction following, and multi-step reasoning tasks.
Generating, filtering, and curating this size of data is a huge undertaking, and by making the dataset openly available, researchers and developers can train, fine-tune, and evaluate models with greater transparency and build models faster.
Nemotron Pre- and Post-Training Dataset
NVIDIA provides over 9T tokens of multilingual reasoning, coding, and safety data to help the community build their custom models.
Llama Nemotron VLM Dataset
The Nemotron Nano 2 VL model provides full transparency with the compilation of high-quality post-training datasets for understanding, querying, and summarizing images.
Nemotron Safety Datasets
High-quality, curated datasets built to power multilingual content safety, advanced policy reasoning, and threat-aware AI—spanning moderation data and audio-based safety signals for modern AI assistants.
Nemotron RL Datasets
Train models with the same reinforcement learning (RL) data powering Nemotron, including multi-turn trajectories, tool calls, and preference signals across coding, math, reasoning, and agentic tasks to build adaptive, reliable real-world AI.
Feature Request Board
Shape the future of Nemotron. Upvote your favorite features or suggest new ones.
Developer Tools
NVIDIA NeMo
Simplify AI agent lifecycle management by fine-tuning, deploying, and continuously optimizing Nemotron models with NVIDIA NeMo™.
NVIDIA TensorRT-LLM
TensorRT™-LLM is an open-source library built to deliver high-performance, real-time inference optimization for large language models like Nemotron on NVIDIA GPUs. This open-source library is available on the TensorRT-LLM GitHub repo and includes a modular Python runtime, PyTorch-native model authoring, and a stable production API.
Open-Source Frameworks
Deploy Nemotron models using open-source frameworks such as Hugging Face transformers for development or vLLM for deployment and production use cases on all supported platforms.
Introductory Resources
Power Specialized AI Agents For Targeted Tasks With Efficient NVIDIA Nemotron 3 Nano Accuracy
NVIDIA Nemotron 3 Nano brings advanced reasoning and agentic capabilities with high efficiency using hybrid Transformer-Mamba MoE architecture and a configurable thinking budget—so you can dial accuracy, throughput, and cost to match your real‑world needs.
Build More Accurate and Efficient AI Agents With NVIDIA Llama Nemotron Super 1.5
AI agents now solve multi-step problems, write production-level code, and act as general assistants across multiple domains. But to reach their full potential, the systems need advanced reasoning models without being prohibitively expensive.
Open Dataset Preserves High-Value Math and Code, and Augments With Multilingual Reasoning
Build advanced reasoning models from carefully curated, high-signal web content and large-scale synthetic data.
Starter Kits
Start solving AI challenges by developing custom agents with NVIDIA Nemotron models for various use cases. Explore implementation scripts, explainer blogs, and more how-to documentation for various stages of AI development.
Build a Report Generation Agent With Nemotron
The workshop guides developers in building a report generation agent using NVIDIA Nemotron and LangGraph, focusing on four core considerations of AI agents: model, tools, memory and state, and routing.
Tutorial Video: Building a Report Generation Agent With NVIDIA Nemotron Nano v2
NVIDIA Launchable: Build an Agent Workshop
Learning Path: How to Build an AI Agent
Build a RAG Agent With Nemotron
In this self-paced workshop, gain a deep understanding of agentic retrieval-augmented generation (RAG) core principles, including the NVIDIA Nemotron model family, and learn how to build your own customized, shareable agentic RAG system using LangGraph within a turnkey, portable development environment.
Tutorial Video: Build a RAG Agent With NVIDIA Nemotron
On-Demand Livestream: Build a RAG Agent With NVIDIA Nemotron | Nemotron Labs
NVIDIA Launchable: Build an Agent Workshop
Learning Path: How To Build an Agent RAG Application
Build a Bash Computer Use Agent With Nemotron
In this self-paced workshop, gain a deep understanding of agentic retrieval-augmented generation (RAG) core principles, including the NVIDIA Nemotron model family, and learn how to build your own customized, shareable agentic RAG system using LangGraph within a turnkey, portable development environment.
Tutorial Video: Create a Bash Agent in One Hour
On-Demand Livestream: Build a Bash Computer Operator Agent | Nemotron Labs
Nemotron 3 Nano 30B A3B
Below are the resources that outline exactly how NVIDIA Research Teams trained the NVIDIA Nemotron 3 Nano model. From pretraining to final model checkpoint—everything is open and available for you to use and learn from.
Models: Nemotron 3 Model Collection
Datasets: Pretraining, Post training, and RL Dataset
Whitepaper: Nemotron 3 Whitepaper
Llama Nemotron Super 1.5 49B
Below are a set of resources that outline the process the NVIDIA Research Teams used to produce Llama 3.3 Nemotron Super 49B V1.5.
More Resources
Ethical Considerations
NVIDIA believes Trustworthy AI is a shared responsibility, and we have established policies and practices to enable development for a wide array of AI applications. When downloading or using this model in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
NVIDIA has collaborated with Google DeepMind to watermark generated videos from the NVIDIA API catalog.
For more detailed information on ethical considerations for this model, please see the System Card, Model Card++ Explainability, Bias, Safety & Security, and Privacy Subcards. Please report security vulnerabilities or NVIDIA AI concerns here.