Explore the latest community-built AI models with an API optimized and accelerated by NVIDIA, then deploy anywhere with NVIDIA NIM inference microservices.
Integrations
Get up and running quickly with familiar APIs.
Use NVIDIA APIs from your existing tools and applications with as little as three lines of code.
Work with your favorite LLM programming frameworks, including LangChain and LlamaIndex, and easily deploy your applications.
Run Anywhere
Part of NVIDIA AI Enterprise, NVIDIA NIM is a set of easy-to-use inference microservices for accelerating the deployment of foundation models on any cloud or data center and helping to keep your data secure.
Deploy NIM for your model with a single command. You can also easily run NIM with fine tuned-models.
Get NIM up and running with the optimal runtime engine based on your NVIDIA-accelerated infrastructure.
Developers can integrate self-hosted NIM endpoints in just a few lines of code.
Rely on production-grade runtimes, including ongoing security updates, and run your business applications with stable APIs backed by enterprise-grade support with NVIDIA NIM.
Lower the operational cost of running the models in production with AI runtimes that are continuously optimized for performance on NVIDIA accelerated infrastructure.
Lower the operational cost of running models in production with AI runtimes that are continuously optimized for low latency and high throughput on NVIDIA-accelerated infrastructure.
Configuration: Llama 3-70b-instruct, input token length: 7,000, output token length: 1,000. Concurrent client requests: 100, on 4 x H100 SXM NVLink. NIM OFF: FP16, TTFT: ~120s, ITL: ~180ms; NIM ON: FP8, TTFT: ~4.5s, ITL: ~70ms.
How To Buy
Use Cases
See how NVIDIA APIs support industry use cases and jump-start your AI development with curated examples.
Ecosystem
Join leading partners to develop your AI applications with models, toolkits, vector databases, frameworks, and infrastructure from our ecosystem.
Resources
Explore technical documentation to start prototyping and building your enterprise AI applications with NVIDIA APIs or scale on your own infrastructure with NVIDIA NIM.