MLPerf Training v5.1 NVIDIA Dominates while AMD Has a Strong Showing and Cisco Silicon

0
NVIDIA DGX Station GB300 Grace With SOCAMM
NVIDIA DGX Station GB300 Grace With SOCAMM

MLPerf Training v5.1 was released this week and a lot of the results are about what you would expect. NVIDIA dominated the results. The only real other accelerator in the results was AMD, and AMD certainly had a strong showing. Perhaps the most interesting is that Cisco Silicon One G200 made an appearance in the list. We dug into the configurations and found something very notable.

MLPerf Training v5.1 NVIDIA Dominates, while AMD Has a Strong Showing and Cisco Silicon

Starting off, NVIDIA showed Blackwell Ultra results, saying it is 4-5x the performance of the H100 generation. If you have not been keeping up, B200 is Blackwell and B300 is Blackwell Ultra.

NVIDIA MLPerf Training V5.1 Performance Growth
NVIDIA MLPerf Training V5.1 Performance Growth

For several years, I have been calling MLPerf NVIDIA’s benchmark. NVIDIA made a chart of how often they have won compared to Google’s TPU and others that have submitted.

NVIDIA MLPerf Training V5.1 Winds Versus Google
NVIDIA MLPerf Training V5.1 Winds Versus Google

For its part, AMD did not win the benchmarks, but it showed relatively competitive numbers. We did not measure the pixels on the bar charts here, but the message was clear. AMD is saying that its current Instinct MI355X solutions are in the ballpark of Blackwell.

AMD MLPerf Training V5.1 Results Slide
AMD MLPerf Training V5.1 Results Slide

Reading through the configuration notes, we saw that Cisco submitted results using the Cisco Silicon One G200 networking. That is important since Cisco Silicon One joined NVIDIA Spectrum-X.

NVIDIA MLPerf Training V5.1 Cisco Silicon One G200
NVIDIA MLPerf Training V5.1 Cisco Silicon One G200

We also saw in the MangoBoost results that they had AMD Instinct GPUs and something interesting in networking. “MangoBoost GPUBoost RDMA” is listed in the networking.

NVIDIA MLPerf Training V5.1 MangoBoost GPUBoost RDMA
NVIDIA MLPerf Training V5.1 MangoBoost GPUBoost RDMA

It turns out that MangoBoost has a DPU solution. We tried finding a block diagram since someone mentioned they have an ASIC. We could not find one. It seems like the MangoBoost folks are STH readers based on this FPGA DPU presentation we found (the previous slide’s Xilinx SolarFlare headline was ours too. Given that we are about to work on theĀ Xsight Labs E1 DPU, maybe we will have to dig into this one as well.

Final Words

NVIDIA took top marks in benchmarks, which is what seems to happen just about every MLPerf training and inference benchmark. AMD showing up is likely good, but we are not sure if this is going to sell a lot of GPUs for AMD. Perhaps the most interesting thing was the networking side with Cisco using its switch silicon and MangoBoost using its DPUs.

In the Substack, we went into an analysis of the CPUs used for submissions, and some of the AMD EPYC versus Intel Xeon trends, along with the use of NVIDIA Grace. While doing that, we found that the configuration information, although apparently peer-reviewed, was incorrect for at least one cluster.

With MLPerf Training 5.1 Configurations in Question, We Look at the Server CPUs by Patrick Kennedy

Taking a CPU-centric look at the MLPerf Training benchmark

Read on Substack

LEAVE A REPLY

Please enter your comment!
Please enter your name here

This site uses Akismet to reduce spam. Learn how your comment data is processed.