NVIDIA’s GB200 NVL72 Boosts AI Model Evaluation at UC Berkeley’s LMArena



Tony Kim
Jun 18, 2025 15:16

LMArena at UC Berkeley utilizes NVIDIA’s GB200 NVL72 to enhance AI model evaluation, thanks to collaborations with NVIDIA and Nebius, improving the ranking of large language models.



NVIDIA's GB200 NVL72 Boosts AI Model Evaluation at UC Berkeley's LMArena

LMArena, a research initiative at the University of California, Berkeley, has significantly advanced its ability to evaluate large language models (LLMs) with the aid of NVIDIA’s GB200 NVL72 systems, as reported by NVIDIA. This collaboration, alongside Nebius, has enabled LMArena to refine its model ranking capabilities, providing insights into which LLMs excel in particular tasks such as math, coding, and creative writing.

Enhancing Model Evaluation with P2L

The core of LMArena’s advancements lies in the Prompt-to-Leaderboard (P2L) model, which collects human votes to determine the best-performing AI in various domains. According to Wei-Lin Chiang, LMArena’s co-founder and a doctoral student at Berkeley, the process involves applying Bradley-Terry coefficients to user preferences. This helps identify the most effective models for specific tasks, offering a nuanced understanding beyond a single overall score.

LMArena’s collaboration with NVIDIA DGX Cloud and Nebius AI Cloud has been crucial in deploying P2L at scale. The use of NVIDIA’s GB200 NVL72 allows for scalable, production-ready AI workloads in the cloud. This partnership has fostered a cycle of rapid feedback and co-learning, enhancing both P2L and the DGX Cloud platform.

Technical Advancements and Deployment

In February, LMArena successfully deployed P2L on the NVIDIA GB200 NVL72, hosted by Nebius via NVIDIA DGX Cloud. This deployment was facilitated by a shared sandbox environment developed by NVIDIA and Nebius, enabling early adopters to test the NVIDIA Blackwell platform efficiently.

The GB200 NVL72 platform, integrating 36 Grace CPUs and 72 Blackwell GPUs, provides high-bandwidth, low-latency performance, and is equipped with up to 30 TB of fast, unified memory. This infrastructure supports demanding AI tasks and promotes efficient resource allocation.

Open Source Enablement

The DGX Cloud team, in collaboration with Nebius and LMArena, ensured a seamless deployment process for open-source developers targeting GB200 NVL72. This involved compiling and optimizing key AI frameworks, such as PyTorch and Hugging Face Transformers, for the Arm64 and CUDA environment.

This comprehensive support allowed developers to leverage state-of-the-art tools without compatibility issues, focusing on building products rather than porting libraries. The project demonstrated impressive performance improvements, completing training runs significantly faster than previous configurations.

For a detailed look at the collaboration and technological advancements, visit the NVIDIA blog.

Image source: Shutterstock


Source: https://blockchain.news/news/nvidia-gb200-nvl72-boosts-ai-model-evaluation-lmarena