Kubernetes Embraces Multi-Node NVLink for Enhanced AI Workloads



Timothy Morano
Nov 10, 2025 06:48

NVIDIA’s GB200 NVL72 introduces ComputeDomains for efficient AI workload management on Kubernetes, facilitating secure, high-bandwidth GPU connectivity across nodes.



Kubernetes Embraces Multi-Node NVLink for Enhanced AI Workloads

NVIDIA has unveiled a significant advancement in AI infrastructure with the introduction of the GB200 NVL72, which enhances the deployment and scaling of AI workloads on Kubernetes. This innovation is set to redefine how large-language models are trained and scalable, low-latency inference workloads are managed, according to NVIDIA.

ComputeDomains: A New Abstraction

The core of this development lies in a new Kubernetes abstraction called ComputeDomains. This abstraction is designed to simplify the complexity of ensuring secure GPU-to-GPU memory operations across nodes using a multi-node NVLink fabric. ComputeDomains are integrated into the NVIDIA DRA driver for GPUs, bridging low-level GPU constructs like NVIDIA NVLink and IMEX with Kubernetes-native scheduling concepts.

ComputeDomains address the limitations of static, manually defined NVLink setups by dynamically creating and managing IMEX domains as workloads are scheduled. This flexibility enhances security isolation, fault tolerance, and cost efficiency, making it a robust solution for modern AI infrastructure.

Advancements in GPU System Design

The evolution from single-node to multi-node GPU computing has been pivotal. Earlier NVIDIA DGX systems were limited to intra-node scaling. However, with NVIDIA’s Multi-Node NVLink (MNNVL), GPUs across different servers can communicate at full NVLink bandwidth, transforming an entire rack into a unified GPU fabric. This enables seamless performance scaling and forms the basis for ultra-fast distributed training and inference.

ComputeDomains capitalize on this advancement by providing a Kubernetes-native way to support multi-node NVLink, already forming the basis for several higher-level components in NVIDIA’s Kubernetes stack.

Implementation and Benefits

The NVIDIA DRA driver for GPUs now offers ComputeDomains, which dynamically manage IMEX domains as workloads are scheduled and completed. This dynamic management ensures that each workload gets its own isolated IMEX domain, facilitating secure GPU-to-GPU communication while maintaining high resource utilization.

ComputeDomains allow for seamless integration and management across nodes, dynamically adjusting as workloads grow or shrink. This not only enhances security and fault isolation but also maximizes resource utilization, particularly in multi-tenant environments.

Future Outlook

The latest release of the NVIDIA DRA driver for GPUs, version 25.8.0, includes significant improvements for ComputeDomains. These enhancements aim to provide more flexible scheduling and ease of use, addressing current limitations such as single pod per node constraints and increasing resource utilization.

As NVIDIA continues to push the boundaries of AI infrastructure, ComputeDomains are poised to become a cornerstone for scalable, topology-aware AI orchestration on platforms like the GB200 NVL72. These innovations promise to streamline multi-node training and inference, making distributed workloads simpler to deploy and manage on Kubernetes.

Image source: Shutterstock


Source: https://blockchain.news/news/kubernetes-embraces-multi-node-nvlink-ai-workloads