NVIDIA AI

Instantly Deploy Generative AI With NVIDIA NIM

Explore the latest community-built AI models with an API optimized and accelerated by NVIDIA, then deploy anywhere with NVIDIA NIM inference microservices.

Experience Leading Open Models Now

Integrations

Accelerated AI is Just an API Call Away

Get up and running quickly with familiar APIs.

Seamless Compatibility With Popular Libraries

Use NVIDIA APIs from your existing tools and applications with as little as three lines of code.

Build With the Tools You Love

Work with your favorite LLM programming frameworks, including LangChain and LlamaIndex, and easily deploy your applications.


Run Anywhere

Accelerate Your AI Deployment With NVIDIA NIM

Part of NVIDIA AI Enterprise, NVIDIA NIM is a set of easy-to-use inference microservices for accelerating the deployment of foundation models on any cloud or data center and helping to keep your data secure.

Accelerate Your AI Deployment With NVIDIA NIM

Deploy NIM

Deploy NIM for your model with a single command. You can also easily run NIM with fine tuned-models.

Run Inference

Get NIM up and running with the optimal runtime engine based on your NVIDIA-accelerated infrastructure.

Build

Developers can integrate self-hosted NIM endpoints in just a few lines of code.

Deploy
Run
Build
docker run nvcr.io/nim/publisher_name/model_name
curl -X 'POST' \ 'http://0.0.0.0:8000/v1/completions' \ -H 'accept: application/json' \ -H 'Content-Type: application/json' \ -d '{ "model" : "model_name", "prompt" : "Once upon a time", "max_tokens" : 64 }'
import openai client = openai.OpenAI( base_url = "YOUR_LOCAL_ENDPOINT_URL", api_key="YOUR_LOCAL_API_KEY" ) chat_completion = client.chat.completions.create( model="model_name", messages=[{"role" : "user" , "content" : "Write me a love song" }], temperature=0.7 )

Deploy With Security and Confidence

Rely on production-grade runtimes, including ongoing security updates, and run your business applications with stable APIs backed by enterprise-grade support with NVIDIA NIM.

Launch Locally or Scale With Kubernetes

Lower the operational cost of running the models in production with AI runtimes that are continuously optimized for performance on NVIDIA accelerated infrastructure.

Lower Costs and Your Carbon Footprint

Lower the operational cost of running models in production with AI runtimes that are continuously optimized for low latency and high throughput on NVIDIA-accelerated infrastructure.

NIM Delivers 5X Higher Throughput

Configuration: Llama 3-70b-instruct, input token length: 7,000, output token length: 1,000. Concurrent client requests: 100, on 4 x H100 SXM NVLink. NIM OFF: FP16, TTFT: ~120s, ITL: ~180ms; NIM ON: FP8, TTFT: ~4.5s, ITL: ~70ms.

See NVIDIA NIMs in Action

How To Buy

Get Started

Start Prototyping for Free

Get started with easy-to-use, NVIDIA-managed serverless APIs or self-host NIM for development through the NVIDIA Developer Program.

  • Access fully-accelerated AI infrastructure.
  • Your data isn't used for model training.
  • Use 1,000 inference credits to get started or join the NVIDIA Developer Program to self-host NIM for development at no cost.

Deploy on Your Own Infrastructure

Use NVIDIA NIM to self-host AI models in the cloud or data center of your choice.

  • Ensure that data never leaves your secure enclave.
  • Seamlessly transition from cloud endpoints to self-hosted APIs without code change.
  • Buy an NVIDIA AI Enterprise license for production.

Use Cases

Ignite Your Innovation

See how NVIDIA APIs support industry use cases and jump-start your AI development with curated examples.

Digital Humans

Bring game characters to life or create interactive virtual avatars to enhance customer service, empowering your application to connect more deeply with users.

Content Generation

Generate highly relevant, bespoke, and accurate content, grounded in the domain expertise and proprietary IP of your enterprise.

Biomolecular Generation

Biomolecular generative models and the computational power of GPUs efficiently explore the chemical space, rapidly generating diverse sets of small molecules tailored to specific drug targets or properties.

Ecosystem

Take Your Enterprise AI Farther, Faster

Join leading partners to develop your AI applications with models, toolkits, vector databases, frameworks, and infrastructure from our ecosystem.

Resources

Unlock, Upskill, and Upscale

LaunchPad

Unlock AI With a Hands-On Lab

Experience the power of AI with end-to-end solutions through guided hands-on labs for RAG-based chatbots, drug discovery, and route optimizations.

NVIDIA Developer Program

Accelerate Your AI Applications

Get free access to NIM and technical learning resources for application development, research and testing through the NVIDIA Developer Program.

AI Workbench

Start Small. Scale Big

NVIDIA AI Workbench gives developers the flexibility to run API-enabled models on local or remote GPU-powered containers, allowing for interactive project workflows from experimentation to prototyping to proof of concept.

News

Explore NVIDIA NIM in the News

Check out the latest NVIDIA press releases to see how NIM and generative AI are impacting industries, partners, customers, and more.

Documentation

Explore technical documentation to start prototyping and building your enterprise AI applications with NVIDIA APIs or scale on your own infrastructure with NVIDIA NIM.