James Ding
Nov 10, 2025 06:41
NVIDIA’s Dynamo platform now integrates with Kubernetes to streamline AI inference management, offering improved performance and reduced costs for data centers, according to NVIDIA’s latest updates.
NVIDIA has announced a significant enhancement to its AI inference capabilities through the integration of its Dynamo platform with Kubernetes. This collaboration aims to streamline the management of both single- and multi-node AI inference, according to NVIDIA.
Enhanced Performance through Disaggregated Inference
The NVIDIA Dynamo platform now supports disaggregated serving, a method that optimizes performance by intelligently assigning AI inference tasks to independently optimized GPUs. This approach alleviates resource bottlenecks by separating the processing of input prompts from output generation. As a result, NVIDIA claims that models such as DeepSeek-R1 can achieve greater efficiency and performance.
Recent benchmarks have shown that disaggregated serving with NVIDIA Dynamo on GB200 NVL72 systems offers the lowest cost per million tokens for complex reasoning models. This integration allows AI providers to reduce manufacturing costs without additional hardware investments.
Scaling AI Inference in the Cloud
With NVIDIA Dynamo now integrated into managed Kubernetes services from major cloud providers, enterprise-scale AI deployments can scale efficiently across NVIDIA Blackwell systems. This integration ensures performance, flexibility, and reliability for large-scale AI applications.
Cloud giants like Amazon Web Services, Google Cloud, and Oracle Cloud Infrastructure are leveraging NVIDIA Dynamo to enhance their AI inference capabilities. For instance, AWS accelerates generative AI inference with NVIDIA Dynamo integrated with Amazon EKS, while Google Cloud offers a recipe for optimizing large language model inference using NVIDIA Dynamo.
Simplifying AI Inference with NVIDIA Grove
To further simplify AI inference management, NVIDIA has introduced NVIDIA Grove, an API within the Dynamo platform. Grove enables users to provide a high-level specification of their inference systems, allowing for seamless coordination of various components such as prefill and decode phases across GPU nodes.
This innovation allows developers to build and scale intelligent applications more efficiently, as Grove handles the intricate coordination of scaling components, maintaining ratios and dependencies, and optimizing communication across the cluster.
As AI inference becomes increasingly complex, the integration of NVIDIA Dynamo with Kubernetes and NVIDIA Grove offers a cohesive solution for managing distributed AI workloads effectively.
Image source: Shutterstock
Source: https://blockchain.news/news/nvidia-enhances-ai-inference-dynamo-kubernetes