By William Parks, Luke Congdon, Aashica Amrith
The promise of AI is massive, but the infrastructure required to power it often forces a difficult choice: do you prioritize the raw speed of bare-metal hardware or the security and management benefits of a virtualized environment? For too long, organizations have been forced to choose. Now, Nutanix is addressing this trade-off, delivering the high-performance AI compute you need without sacrificing the control and security your enterprise demands.
How do we do this? We power GPU-dense platforms, such as the NVIDIA HGX systems, with the Nutanix AHV hypervisor and leveraging Topology-Aware Scheduling with AI optimized System-Defined VM profiles. Topology-aware scheduling improves AI workload performance by ensuring that virtualized resources (vCPUs, memory) are physically aligned with the hardware accelerators (GPUs, NICs) they utilize. AI-optimized VM profiles improve performance by automating complex, hardware-specific tuning that would otherwise require deep manual expertise.
The integration between Nutanix AHV, Nutanix Flow and NKP allows platform engineers to automate the deployment of AI workloads on complex, GPU-dense passthrough servers using these AI-optimized VM profile-based virtual machines as GPU-dense Kubernetes worker nodes for their AI workloads.
Capable of operating in fully air-gapped FIPS-compliant environments, this solution is designed to deliver the multitenancy and security of virtualization. It also provides the throughput and performance required for massive LLM training, transforming the Nutanix Cloud Platform into a foundational engine for Enterprise Sovereign AI.
We’ve optimized Nutanix AHV to help ensure the hypervisor enables highly performant placement of VMs for your AI workloads.
The integration between Nutanix AHV and NKP allows platform engineers to automate the deployment of complex, GPU-dense environments with ease.
Agentic AI demands a high-performance, zero trust foundation that maximizes throughput and secures data flow across the AI Factory.
| Feature | Impact on AI Workloads |
|---|---|
| Traffic Segmentation | Nutanix Flow splits traffic to maximize throughput—dedicating InfiniBand/Ethernet to GPU training while BlueField-3 DPUs handle data ingest. |
| GPU Direct RDMA | Allows GPUs across different hosts to exchange data directly over the network, bypassing the host CPU to significantly reduce latency. |
| Infrastructure Offload | By utilizing BlueField-3 DPUs to offload the network dataplane, host CPU and memory reduction deliver higher-performance for AI processing. |
Nutanix Agentic AI delivers the foundational platform for AI. At its core, Nutanix AHV delivers high-performance by leveraging Topology-Aware Scheduling and 8-GPU passthrough so massive models get the raw compute they need without sacrificing the strict multi-tenancy, security, and lifecycle governance of a virtualized environment.
Working in tandem with AHV to address bottlenecks, Nutanix Flow provides advanced software-defined networking and security. Flow intelligently segments the network traffic into high-performance lanes, dedicating InfiniBand (via ConnectX-7) strictly for East-West GPU-to-GPU training communication, while routing North-South data ingest through NVIDIA BlueField-3 DPUs. Combined with the orchestration of AI workloads with NKP, organizations building Enterprise AI Factories or Sovereign Neoclouds no longer need to compromise between bare-metal speed and secure control.
With this seamlessly integrated stack, platform engineering teams are fully equipped to deploy, secure, and scale the next generation of intelligent, compliant applications without sacrificing performance.
©2026 Nutanix, Inc. All rights reserved. Nutanix, the Nutanix logo and all Nutanix product and service names mentioned are registered trademarks or trademarks of Nutanix, Inc. in the United States and other countries. All other brand names mentioned are for identification purposes only and may be the trademarks of their respective holder(s).