AI Foundation Models
Sep 05, 2024
Low Latency Inference Chapter 1: Up to 1.9x Higher Llama 3.1 Performance with Medusa on NVIDIA HGX H200 with NVLink Switch
As large language models (LLMs) continue to grow in size and complexity, multi-GPU compute is a must-have to deliver the low latency and high throughput that...
5 MIN READ
Aug 21, 2024
Mistral-NeMo-Minitron 8B Foundation Model Delivers Unparalleled Accuracy
Last month, NVIDIA and Mistral AI unveiled Mistral NeMo 12B, a leading state-of-the-art large language model (LLM). Mistral NeMo 12B consistently outperforms...
5 MIN READ
Aug 13, 2024
New NIM Available: Mistral Large 2 Instruct LLM
The new model by Mistral excels at a variety of complex tasks including text summarization, multilingual translation and reasoning, programming, question and...
1 MIN READ
Jul 26, 2024
Power Text-Generation Applications with Mistral NeMo 12B Running on a Single GPU
NVIDIA collaborated with Mistral to co-build the next-generation language model that achieves leading performance across benchmarks in its class. With a growing...
6 MIN READ
Jul 25, 2024
Revolutionizing Code Completion with Codestral Mamba, the Next-Gen Coding LLM
In the rapidly evolving field of generative AI, coding models have become indispensable tools for developers, enhancing productivity and precision in software...
5 MIN READ
Jul 23, 2024
Supercharging Llama 3.1 across NVIDIA Platforms
Meta's Llama collection of large language models are the most popular foundation models in the open-source community today, supporting a variety of use cases....
8 MIN READ
Jul 02, 2024
Phi-3-Medium: Now Available on the NVIDIA API Catalog
Phi-3-Medium accelerates research with logic-rich features in both short (4K) and long (128K) context.
1 MIN READ
Jul 01, 2024
StarCoder2-15B: A Powerful LLM for Code Generation, Summarization, and Documentation
Trained on 600+ programming languages, StarCoder2-15B is now packaged as a NIM inference microservice available for free from the NVIDIA API catalog.
1 MIN READ
Jul 01, 2024
Google's New Gemma 2 Model Now Optimized and Available on NVIDIA API Catalog
Gemma 2, the next generation of Google Gemma models, is now optimized with TensorRT-LLM and packaged as NVIDIA NIM inference microservice.
1 MIN READ
Jun 26, 2024
Generate High-Quality, Context-Aware Responses for Chatbots and Search Engines with Llama 3-ChatQA
Experience and test Llama3-ChatQA models at scale with performance optimized NVIDIA NIM inference microservice using the NVIDIA API catalog.
1 MIN READ
Jun 17, 2024
Simplify and Accelerate Programming Tasks with Mistral’s Codestral GenAI Model
Experience Codestral, packaged as an NVIDIA NIM inference microservice for code completion, writing tests, and debugging in over 80 languages using the NVIDIA...
1 MIN READ
Jun 10, 2024
SOLAR-10.7B: Optimized Model Tailored Instruction Following, Reasoning, and Mathematical Tasks
Enhance efficiency and performance in instruction-based NLP tasks with SOLAR-10.7B, especially in following instructions, reasoning, and mathematical tasks.
1 MIN READ
Jun 03, 2024
Breeze-7B: LLM Specialized for Traditional Chinese
The model demonstrates strong performance for tasks such as Q&A, multi-round chat, and summarization in both traditional Chinese and English.
1 MIN READ
Jun 03, 2024
BGE-M3: Advanced Multilingual Text Retrieval Model
Experience the versatile embedding model designed for multilingual, multi-functional, and multi-granularity text retrieval tasks, excelling in dense,...
1 MIN READ
May 30, 2024
Convert Natural Language to Code with CodeGemma
Experience the advanced LLM API for code generation, completion, mathematical reasoning, and instruction following with free cloud credits.
1 MIN READ
May 30, 2024
Personalized Learning with Gipi, NVIDIA TensortRT-LLM, and AI Foundation Models
Over 1.2B people are actively learning new languages, with over 500M learners on digital learning platforms such as Duolingo. At the same time, a significant...
6 MIN READ