NVIDIA NIM boosts developer productivity by offering a standardized way to embed generative AI into applications.
What is NVIDIA NIM?NVIDIA NIM (NVIDIA Inference Microservices) is a platform designed to enhance the deployment and integration of generative AI models. It provides a standardized approach for embedding generative AI into various applications, significantly boosting developer productivity and optimizing infrastructure investments.
Features of NVIDIA NIMKey features and capabilities of NVIDIA NIM include:
Nearly 200 technology partners, including Cadence, Cloudera, Cohesity, DataStax, NetApp, Scale AI, and Synopsys, are incorporating NVIDIA NIM to speed up generative AI deployments tailored for specific applications like copilots, coding assistants, and digital human avatars. Additionally, Hugging Face is introducing NVIDIA NIM, beginning with Meta Llama 3.
Jensen Huang, NVIDIA’s founder and CEO, stressed the accessibility and significance of NVIDIA NIM, saying, “Every enterprise aims to integrate generative AI into its operations, but not all have a team of dedicated AI researchers.” NVIDIA NIM makes it possible for almost any organization to utilize generative AI.
Businesses can implement AI applications using NVIDIA NIM via the NVIDIA AI Enterprise software platform. Starting next month, NVIDIA Developer Program members will have free access to NVIDIA NIM for research, development, and testing on their chosen infrastructure.
NVIDIA NIM containers are designed to streamline model deployment for GPU-accelerated inference and come equipped with NVIDIA CUDA software, NVIDIA Triton Inference Server, and NVIDIA TensorRT-LLM software. Over 40 models, such as Databricks DBRX, Google’s Gemma, Meta Llama 3, Microsoft Phi-3, and Mistral Large, are accessible as NIM endpoints on ai.nvidia.com.
Through the Hugging Face AI platform, developers can easily access NVIDIA NIM microservices for Meta Llama 3 models. This integration allows them to efficiently run Llama 3 NIM using Hugging Face Inference Endpoints powered by NVIDIA GPUs.
Numerous platform providers, including Canonical, Red Hat, Nutanix, and VMware, support NVIDIA NIM on both open-source KServe and enterprise solutions. AI application companies like Hippocratic AI, Glean, Kinetica, and Redis are utilizing NIM to drive generative AI inference. Leading AI tools and MLOps partners, such as Amazon SageMaker, Microsoft Azure AI, Dataiku, DataRobot, and others, have integrated NIM into their platforms, enabling developers to create and deploy domain-specific generative AI applications with optimized inference.
Get your free NVIDIA Game Pass offer now
Global system integrators and service delivery partners, including Accenture, Deloitte, Infosys, Latentview, Quantiphi, SoftServe, TCS, and Wipro, have developed expertise in NIM to assist enterprises in rapidly developing and implementing production AI strategies. Enterprises can run NIM-enabled applications on NVIDIA-Certified Systems from manufacturers like Cisco, Dell Technologies, Hewlett-Packard Enterprise, Lenovo, and Supermicro, as well as on servers from ASRock Rack, ASUS, GIGABYTE, Ingrasys, Inventec, Pegatron, QCT, Wistron, and Wiwynn. Additionally, NIM microservices are integrated into major cloud platforms, including Amazon Web Services, Google Cloud, Azure, and Oracle Cloud Infrastructure.
Leading companies are leveraging NIM for diverse applications across industries. Foxconn uses NIM for domain-specific LLMs in AI factories, smart cities, and electric vehicles. Pegatron employs NIM for Project TaME to advance local LLM development for various industries. Amdocs uses NIM for a customer billing LLM, significantly reducing costs and latency while improving accuracy. ServiceNow integrates NIM microservices within its Now AI multimodal model, providing customers with fast and scalable LLM development and deployment.
How to try NVIDIA NIM?Using NVIDIA NIM to deploy and experiment with generative AI models is straightforward. Here is a step-by-step guide to help you get started:
Here is an example of using the model:
Featured image credit: Kerem Gülen/Midjourney