Back toBlog

3 Ways NVIDIA NIM and GPT-in-a-Box 2.0 Accelerate Your Enterprise GenAI

By Mike Barmonde

June 4, 2024 | min

The speed at which you can deploy and use GenAI is now faster than ever, thanks to the latest announcements from NVIDIA and Nutanix.

At Computex 2024, NVIDIA announced the availability of NIM microservices for developers, offering a cost-free way for developers to accelerate their GenAI efforts if they’re part of the NVIDIA Developer Program. This announcement aligns closely with the latest GenAI updates from Nutanix at .NEXT 2024, where Nutanix revealed ongoing efforts to integrate NVIDIA NIM into the upcoming Nutanix GPT-in-a-Box 2.0 solution. This integration will enable customers to effortlessly deploy GenAI in minutes, without the need for specialized infrastructure or architecture.

Key diagram of how Nutanix GPT-in-a-Box 2.0 works with NVIDIA NIM

These recent announcements aren’t just superficial hype. The combination of NVIDIA AI Enterprise NIM and Nutanix GPT-in-a-Box 2.0 not only promises unprecedented speed for your GenAI deployments but also unparalleled adaptability to a rapidly evolving AI landscape. 

But is adaptability critical for GenAI success? Absolutely, and here are three reasons why.

1. Quick Access to New and Updated GenAI Models

GenAI models are constantly evolving, and today’s top performers might not meet your needs tomorrow. Consistency, support, and viability for a large-language model (LLM) are also key points to consider. Like code projects before them, a dead end for an LLM could stall or derail your GenAI offering.

Hugging Face Open LLM Leaderboard Benchmark as of May, 29, 2024

With NVIDIA NIM and Nutanix GPT-in-a-Box, the workflow to update or replace a model will be seamless. 

NVIDIA NIM provides a library of LLMs curated for enterprise GenAI. These models are tied to APIs that developers will leverage to connect GenAI apps to models. A model change or replacement with NVIDIA NIM is invisible to the end-user and they’ll see the results quickly. 

Emerging models are becoming increasingly more powerful by their addition of multiple types of generated AI content beyond text, like pictures, audio, or even video. These multimodal models create a single interface for GenAI content creation.

LLM Multimodality abstract concept from NVIDIA

Nutanix GPT-in-a-Box creates a platform to manage models, data sets, and infrastructure for GenAI. The upcoming 2.0 release - now in development - will integrate NVIDIA NIM.  These models and data will sit closest to their users from the edge to the public cloud ensuring performance. The infrastructure of GPT-in-a-Box is based on hyperconvergence that combines all the typical hardware elements like compute, networking, storage, and GPUs into a single form factor that can be scaled up, or down, from a small closet to a massive datacenter. 

NVIDIA NIM combined with GPT-in-a–Box 2.0 will create a model ‘factory’ where choice, adaptability, and speed remove the pressure of picking the best model, and instead allow you to choose the right model.

2. Private Control over GenAI Models and Data

Running GenAI on Nutanix GPT-in-a-Box gives you native data services that are resilient and scalable and will provide NVIDIA AI Enterprise NIM users a solid foundation for the storage of GenAI workflows. 

GenAI workloads and models mainly run on containers. NVIDIA NIM is itself a set of microservices and containers. For many GenAI teams that need enterprise-grade data availability, simple data management, backup, and recovery, GPT-in-a-Box provides comprehensive container data services: Nutanix Data Services for Kubernetes.

Many customers demand security and privacy for GenAI, and both are built into the data fabric Nutanix provides. Your data is better protected wherever you run your GenAI workloads, even in public cloud environments, with the Nutanix Cloud Clusters platform.  

3. Software-Defined Infrastructure Speeds Gen AI Success 

GenAI is evolving rapidly, demanding a broad range of compute performance and storage capacity to run, protect, and store both Gen AI models and data. Nutanix and NVIDIA are working together to simplify access to new and updated models running on a software-defined architecture that easily integrates new GPUs while providing consistent data services across block, file, and object storage.   

The Nutanix Cloud Platform with GPT-in-a-Box running NVIDIA AI Enterprise is currently validated to run on the Nutanix AHV hypervisor. This solution combines a scalable infrastructure platform with a scalable GenAI framework, enabling fast results that can adapt to your business, anywhere you run your AI workloads.

The Time for AI Starts Now

There has never been a better time to determine your AI strategy. NVIDIA NIM offers developers a cost-free way to get into GenAI with turn-key APIs and models, while Nutanix GPT-in-a-Box is under development to accelerate the deployment of NVIDIA NIM with a software-defined infrastructure for all your data to scale anywhere. 

So, what’s your next move? 

  • Go to ai.nvidia.com to explore NVIDIA NIM for yourself and get started today

  • Learn more about Getting Started with GenAI on Nutanix GPT-in-a-Box 2.0

https://www.nutanix.com/blog/gpt-in-a-box-2-is-here

About Nutanix

Nutanix is a global leader in cloud software, offering organizations a single platform for running apps and data across clouds. With Nutanix, organizations can reduce complexity and simplify operations, freeing them to focus on their business outcomes. Building on its legacy as the pioneer of HCI, Nutanix is trusted by companies worldwide to power hybrid multicloud environments consistently, simply, and cost-effectively. Learn more at www.nutanix.com or follow us on social media @nutanix.

© 2024 Nutanix, Inc. All rights reserved. Nutanix, the Nutanix logo, and all Nutanix product and service names mentioned herein are registered trademarks or unregistered trademarks of Nutanix, Inc. (“Nutanix”) in the United States and other countries. Kubernetes® is a registered trademark of the Linux Foundation. Other brand names or marks mentioned herein are for identification purposes only and may be the trademarks of their respective holder(s). This blog is for informational purposes only and nothing herein constitutes a warranty or other binding commitment by Nutanix. This blog contains express and implied forward-looking statements, including but not limited to statements regarding our plans and expectations relating to new product features and technology that are under development, the capabilities of such product features and technology and our plans to release product features and technology in the future. Such statements are not historical facts and are instead based on our current expectations, estimates and beliefs. The accuracy of such statements involves risks and uncertainties and depends upon future events, including those that may be beyond our control, and actual results may differ materially and adversely from those anticipated or implied by such statements. Any forward-looking statements included herein speak only as of the date hereof and, except as required by law, we assume no obligation to update or otherwise revise any of such forward-looking statements to reflect subsequent events or circumstances. Any future product or product feature information is intended to outline general product directions, and is not a commitment, promise or legal obligation for Nutanix to deliver any functionality.  This information should not be used when making a purchasing decision.