NVIDIA ConnectX-8 C8240 800G Dual 400G NIC Review This is a SuperNIC

1

Solving the 800G NIC in a PCIe Gen5 Server Challenge

NVIDIA Grace, Intel Xeon 6, AMD EPYC Turn, and Ampere AmpereOne (M) servers are PCIe Gen5. As a result, a Gen5 x16 root that is found on the card is only capable of driving a 400G network link. While that matched theĀ NVIDIA ConnectX-7 400GbE NICs well, with ConnectX-8, we need twice the host bandwidth.

NVIDIA ConnectX 8 24
NVIDIA ConnectX 8 24

The answer was the connector at the rear of the card. While you might glance at this and see a MCIO x16 connector, it turns out that the MCIO x16 connector found on many server motherboards is too wide to fit on the low-profile ConnectX-8 card.

NVIDIA ConnectX 8 16
NVIDIA ConnectX 8 16

As a result, NVIDIA sells a PCIe auxiliary card kit. We have been trying to get one for months, and finally, a STH reader let us borrow one for this review.

NVIDIA ConnectX 8 Aux Kit With Cable 1
NVIDIA ConnectX 8 Aux Kit With Cable 1

The NVIDIA ConnectX-8 side terminates one end of the cable.

NVIDIA ConnectX 8 C8180 Connected To Aux X16 Kit With Cable 1
NVIDIA ConnectX 8 C8180 Connected To Aux X16 Kit With Cable 1

That card has an expansion card that goes into a PCIe x16 (riser) slot and connects using a cable connection to the rear of the SuperNIC.

NVIDIA ConnectX 8 C8180 Connected To Aux X16 Kit With Cable 3
NVIDIA ConnectX 8 C8180 Connected To Aux X16 Kit With Cable 3

While this is a great solution, we ran into a practical challenge. We had only so many PCIe Gen5 x16 slots, and modern server CPUs generally have more lanes than PCIe slots. We obtained the Lenovo kit that goes from the special connector on the back of the ConnectX-8 NIC to two MTK x8 connectors in compatible Lenovo ThinkSystem V4 servers.

NVIDIA ConnectX 8 With Lenovo MTK X8 Cable 1
NVIDIA ConnectX 8 With Lenovo MTK X8 Cable 1

Supermicro uses more industry-standard MCIO x8 connectors, so the Lenovo MTK kit did not help. Instead, custom 2x MCIO x8 to the ConnectX-8 NIC cables were built. These allowed us to connect additional MCIO connections beyond the x16 slots in the server so we could use a single x16 slot and get 32 lanes of PCIe Gen5 to it. While this may sound trivial, the folks working on this went through a monumental effort to make this happen.

NVIDIA ConnectX 8 With Custom Dual MCIO X8 Cable 1
NVIDIA ConnectX 8 With Custom Dual MCIO X8 Cable 1

Now that the physical connectivity was done, we have a quite different configuration. There are instructions on how to do these, but the mlxconfig options are wild.

NVIDIA ConnectX 8 SuperNIC Mlxconfig
NVIDIA ConnectX 8 SuperNIC Mlxconfig

Also, we have mentioned how the setup looks a lot more like the NVIDIA GB10 networking. This is a great example with two NVIDIA C8240’s installed in a single system, yet we have four RDMA devices per and tied to different NUMA nodes.

NVIDIA ConnectX 8 SuperNIC 2x C8240 In A System Mst
NVIDIA ConnectX 8 SuperNIC 2x C8240 In A System Mst

Before we move on, here is a fun one. In ethtool we see 400G speeds, 200G, 100G, and 50G, but not lower speeds. If you were to look at something like theĀ NVIDIA ConnectX-7 Quad Port 50GbE SFP56 adapter we reviewed, that goes from 50G to 1G speeds.

NVIDIA ConnectX 8 Supernicro C8240 Supported Link Modes Ethtool
NVIDIA ConnectX 8 Supernicro C8240 Supported Link Modes Ethtool

Next, let us get to the fun part, the performance.

1 COMMENT

LEAVE A REPLY

Please enter your comment!
Please enter your name here

This site uses Akismet to reduce spam. Learn how your comment data is processed.