Agentic AI / Generative AI
Dec 17, 2025
Real-Time Decoding, Algorithmic GPU Decoders, and AI Inference Enhancements in NVIDIA CUDA-Q QEC
Real-time decoding is crucial to fault-tolerant quantum computers. By enabling decoders to operate with low latency concurrently with a quantum processing unit...
6 MIN READ
Dec 17, 2025
Simulate Robotic Environments Faster with NVIDIA Isaac Sim and World Labs Marble
Building realistic 3D environments for robotics simulation has traditionally been a labor-intensive process, often requiring weeks of manual modeling and setup....
10 MIN READ
Dec 16, 2025
Accelerating Long-Context Inference with Skip Softmax in NVIDIA TensorRT-LLM
For machine learning engineers deploying LLMs at scale, the equation is familiar and unforgiving: as context length increases, attention computation costs...
6 MIN READ
Dec 16, 2025
AI Factories, Physical AI, and Advances in Models, Agents, and Infrastructure That Shaped 2025
2025 was another milestone year for developers and researchers working with NVIDIA technologies. Progress in data center power and compute design, AI...
4 MIN READ
Dec 15, 2025
Inside NVIDIA Nemotron 3: Techniques, Tools, and Data That Make It Efficient and Accurate
Agentic AI systems increasingly rely on collections of cooperating agents—retrievers, planners, tool executors, verifiers—working together across large...
10 MIN READ
Dec 15, 2025
How to Train Scientific Agents with Reinforcement Learning
The scientific process can be repetitive and tedious, with researchers spending hours digging through papers, managing experiment workflows, or wrangling...
13 MIN READ
Dec 12, 2025
Enabling Horizontal Autoscaling of Enterprise RAG Components on Kubernetes
Today’s best AI agents rely on retrieval-augmented generation (RAG) to enable more accurate results. A RAG system facilitates the use of a knowledge base to...
24 MIN READ
Dec 12, 2025
How to Build Privacy-Preserving Evaluation Benchmarks with Synthetic Data
Validating AI systems requires benchmarks—datasets and evaluation workflows that mimic real-world conditions—to measure accuracy, reliability, and safety...
11 MIN READ
Dec 11, 2025
NVIDIA Blackwell Enables 3x Faster Training and Nearly 2x Training Performance Per Dollar than Previous-Gen Architecture
AI innovation continues to be driven by three scaling laws: pre-training, post-training, and test-time scaling. Training is foundational to building smarter...
7 MIN READ
Dec 08, 2025
Optimizing Inference for Long Context and Large Batch Sizes with NVFP4 KV Cache
Quantization is one of the strongest levers for large-scale inference. By reducing the precision of weights, activations, and KV cache, we can reduce the memory...
10 MIN READ
Dec 05, 2025
NVIDIA Kaggle Grandmasters Win Artificial General Intelligence Competition
NVIDIA researchers on Friday won a key Kaggle competition many in the field treat as a real-time pulse check on humanity’s progress toward artificial general...
3 MIN READ
Dec 04, 2025
Optimize Data Center Efficiency for AI and HPC Workloads with Power Profiles
Exponentially growing computational demand is driving power usage higher and pushing data centers to their limits. With facilities power constrained, extracting...
7 MIN READ
Dec 03, 2025
How to Enhance 3D Gaussian Reconstruction Quality for Simulation
Building truly photorealistic 3D environments for simulation is challenging. Even with advanced neural reconstruction methods such as 3D Gaussian Splatting...
7 MIN READ
Dec 02, 2025
NVIDIA-Accelerated Mistral 3 Open Models Deliver Efficiency, Accuracy at Any ScaleÂ
The new Mistral 3 open model family delivers industry-leading accuracy, efficiency, and customization capabilities for developers and enterprises. Optimized...
6 MIN READ
Dec 01, 2025
Train Small Orchestration Agents to Solve Big Problems
Using the right tool and model for a task is a challenging and ever-present engineering problem in agent design. At NVIDIA Research, we're making fast progress...
7 MIN READ
Dec 01, 2025
Build Efficient Financial Data Workflows with AI Model Distillation
Large language models (LLMs) in quantitative finance are increasingly being used for alpha generation, automated report analysis, and risk prediction. Yet...
11 MIN READ