How to Deploy a Wide Variety of LLMs with a Single NVIDIA NIM

Просмотров: 1, 750   |   Загружено: 6 дн
icon
NVIDIA Developer
icon
40
icon
Скачать
iconПодробнее о видео
This video walks you through how to deploy different LLMs from Hugging Face with a single LLM-agnostic NVIDIA NIM container.

NIM microservices are your go-to for deploying the latest AI models from NVIDIA and the community quickly and reliably. They leverage NVIDIA TensorRT-LLM to accelerate and optimize inference alongside popular community backends like vLLM and SGLang to seamlessly deploy a massive range of community LLMs.

Get ready for high-performance inference on NVIDIA GPUs with rapid, reliable deployment for a broad range of LLMs.

➡️ Try out the developer example:

➡️ Technical Deep Dive:

➡️ Join the NVIDIA Developer Program:

➡️ Read and subscribe to the NVIDIA Technical Blog:

00:00:00 - Introduction and Overview
00:00:36 - Setting Up and Launching the Model
00:02:50 - Model Profiles and Compatibility
00:05:02 - Example

LLM, inference, agentic AI, generative AI, AI engineer

Похожие видео

Добавлено: 55 год.
Добавил:
  © 2019-2021
  How to Deploy a Wide Variety of LLMs with a Single NVIDIA NIM - RusLar.Me