New NVIDIA Storage Companion Validation Program Streamlines Enterprise AI Deployment

A pointy improve in generative AI deployments is driving enterprise innovation for enterprises throughout industries. However it’s additionally posing vital challenges for his or her IT groups, as slowdowns from lengthy and complicated infrastructure deployment cycles forestall them from shortly spinning up AI workloads utilizing their very own information.

To assist overcome these obstacles, NVIDIA has launched a storage associate validation program for NVIDIA OVX computing programs. The high-performance storage programs main the best way in finishing the NVIDIA OVX storage validation are DDN, Dell PowerScale, NetApp, Pure Storage and WEKA.

NVIDIA OVX servers mix high-performance, GPU-accelerated compute with high-speed storage entry and low-latency networking to deal with a variety of advanced AI and graphics-intensive workloads. Chatbots, summarization and search instruments, for instance, require giant quantities of information, and high-performance storage is vital to maximise system throughput.

To assist enterprises pair the correct storage with NVIDIA-Licensed OVX servers, the brand new program supplies a standardized course of for companions to validate their storage home equipment. They will use the identical framework and testing that’s wanted to validate storage for the NVIDIA DGX BasePOD reference structure.

To realize validation, companions should full a set of NVIDIA assessments measuring storage efficiency and enter/out scaling throughout a number of parameters that signify the demanding necessities of varied enterprise AI workloads. This contains combos of various I/O sizes, various numbers of threads, buffered I/O vs. direct I/O, random reads, re-reads and extra.

Every take a look at is run a number of occasions to confirm the outcomes and collect the required information, which is then audited by NVIDIA engineering groups to find out whether or not the storage system has handed.

This system gives prescriptive steering to make sure optimum storage efficiency and scalability for enterprise AI workloads with NVIDIA OVX programs. However the total design stays versatile, so prospects can tailor their system and storage selections to suit their current information middle environments and produce accelerated computing to wherever their information resides.

Generative AI use circumstances have essentially totally different necessities than conventional enterprise purposes, so IT groups should rigorously take into account their compute, networking, storage and software program selections to make sure excessive efficiency and scalability.

NVIDIA-Licensed Programs are examined and validated to offer enterprise-grade efficiency, manageability, safety and scalability for AI workloads. Their versatile reference architectures assist ship quicker, extra environment friendly and less expensive deployments than independently constructing from the bottom up.

Powered by NVIDIA L40S GPUs, OVX servers embrace NVIDIA AI Enterprise software program with NVIDIA Quantum-2 InfiniBand or NVIDIA Spectrum-X Ethernet networking, in addition to NVIDIA BlueField-3 DPUs. They’re optimized for generative AI workloads, together with coaching for smaller LLMs (for instance, Llama 2 7B or 70B), fine-tuning current fashions and inference with excessive throughput and low latency.

NVIDIA-Licensed OVX servers are actually out there and delivery from world system distributors, together with GIGABYTE, Hewlett Packard Enterprise and Lenovo. Complete, enterprise-grade help for these servers is supplied by every system builder, in collaboration with NVIDIA.


Validated storage options for NVIDIA-Licensed OVX servers are actually out there, and reference architectures might be revealed over the approaching weeks by every of the storage and system distributors. Be taught extra about NVIDIA OVX Programs.

Leave a Reply

Your email address will not be published. Required fields are marked *