Inspur i24 Review a 2U4N Server Platform for Scale

3

Inspur i24 Power Consumption to Baseline

One of the other, sometimes overlooked, benefits of the 2U4N form factor is power consumption savings. We ran our standard STH 80% CPU utilization workload, which is a common figure for a well-utilized virtualization server, and ran that in the sandwich between the 1U servers and the Inspur i24. With dense servers, heat is a concern, so we replicate what one would likely see in the field. This is the only way to get useful comparison information for 2U4N servers.

STH 2U 4 Node Power Comparison Test Setup
STH 2U 4 Node Power Comparison Test Setup

Here is what we saw in terms of performance compared to our baseline nodes.

Inspur I24 Power Consumption To Baseline
Inspur I24 Power Consumption To Baseline

As you can see, we are getting a small but noticeable power efficiency improvement with the i24’s 2U4N design. Had Inspur used fully independent nodes with their own cooling, we would expect to see higher power consumption here than in our control set.

Inspur I24 Rear Chassis View
Inspur I24 Rear Chassis View

This is about a 1.8% power consumption improvement. Part of that may be running power supplies at higher efficiency levels. Part of it is the shared cooling. When one combines power consumption savings along with increasing density, that can be a big win for data center operators.

STH Server Spider: Inspur i24

In the second half of 2018, we introduced the STH Server Spider as a quick reference to where a server system’s aptitude lies. Our goal is to start giving a quick visual depiction of the types of parameters that a server is targeted at.

STH Server Spider Inspur I24
STH Server Spider Inspur I24

The Inspur i24 is a very interesting mix of capabilities. One gets eight CPUs in a 2U form factor along with 64 DIMM slots. This is while providing a solid mix of hot-swap storage options. Each node can utilize an OCP NIC 2.0 along with up to two PCIe expansion cards making the nodes surprisingly flexible.

There are severs that can fit more capability into each 2U of space such as higher-end GPU or storage servers. Even with that backdrop, the Inspur i24 excels in a number of areas while providing a solid consolidation and efficiency story.

Final Words

The Inspur i24 (NS5162M5) is a system we did not know what to expect when reviewing. It is a system that we have not seen often at trade shows. We also have reviewed many 2U4N servers from a number of different vendors. As such, we have a very thorough process to find differences between the solutions.

Inspur I24 Rear Nodes And PSU View
Inspur I24 Rear Nodes And PSU View

One of the biggest results from our testing is that we saw a performance baseline within testing variances while getting around 1.8% lower power consumption in the “STH Sandwich” test configuration we use to simulate dense racks of these servers. That is a very impressive result as it adds an additional facet to TCO calculations when looking to consolidate from 1U 1-node to 2U 4-node platforms.

Perhaps the highlight of the Inspur i24 is its design. The use of a dedicated chassis management controller, or CMC, is becoming standard. At the same time, using high-density connectors instead of PCB gold fingers is something we find mostly on higher-end systems. Weeks after starting our testing of this system, the one item that clearly makes an impact is the rear PCIe riser design. Many vendors, including large vendors such as Dell EMC, have a series of screws to secure the PCIe risers into 2U4N sleds. Here, Inspur is using thicker gauge steel on the compute sleds which allows them to make what is likely the industry’s most easily serviceable PCIe risers.

Inspur I24 Node PCIe Riser Out
Inspur I24 Node PCIe Riser Out

We have covered several times how Inspur is now the third-largest server vendor. Unlike many of the other large server vendors, Inspur is building its platforms with industry-standard BMCs like what hyper-scalers use. Like these larger vendors, the Inspur i24 is a great example of how it is doing a great job designing its server platforms, in some cases even providing clear industry leadership in terms of server design.

3 COMMENTS

  1. The units should also support the new Series 200 Intel Optane DCPMMs, but only in AppDirect (not Memory) mode.

LEAVE A REPLY

Please enter your comment!
Please enter your name here