ASUS RS700-E11-RS12U Review A New 1U Intel Xeon Sapphire Rapids Server

5

ASUS RS700-E11-RS12U Block Diagram and Topology

Here is the block diagram for the server. Something that is really nice that ASUS does is that it shows the PCIe Gen5 x16 lanes that are not being used in this 1U server.

ASUS RS700 E11 RS12U Block Diagram
ASUS RS700 E11 RS12U Block Diagram

Here is a quick look at the topology. As configured when we got the system, CPU0 had all of the peripherals and CPU1 just had cores.

ASUS RS700 E11 RS12U Topology
ASUS RS700 E11 RS12U Topology

Our system only came with two 64GB DIMMs, one per CPU. While this worked for screenshots, we replaced these with 8x DDR5 RDIMMs in each socket for our performance testing.

ASUS ASMB11-iKVM Management

With this generation of servers, we have the new ASPEED AST2600 BMC. That means we also get a newer ASUS management solution.

ASUS RS700 E11 RS12U Rear IO Area With M.2 MicroSD GenZ PCIe Gen5 X16 Connectors And ASPEED AST2600 3
ASUS RS700 E11 RS12U Rear IO Area With M.2 MicroSD GenZ PCIe Gen5 X16 Connectors And ASPEED AST2600 3

The new solution appears to be MegaRAC based and is called the ASUS ASMB11-iKVM.

ASUS ASMB11 IKVM Logon
ASUS ASMB11 IKVM Logon

This has all of the sensor readings and logs that one would expect, as well as not just a web GUI management interface but IPMI and Redfish capabilities.

ASUS ASMB11 IKVM Sensors
ASUS ASMB11 IKVM Sensors

One nice feature is that ASUS has the HTML5 iKVM with remote media as standard. We have seen other vendors charge for this capability, so it is nice that ASUS includes it here.

ASUS ASMB11 IKVM IKVM
ASUS ASMB11 IKVM IKVM

There are also standard features like being able to power cycle the server. One item we wish was here was the ability to boot to BIOS.

ASUS ASMB11 IKVM Power Control
ASUS ASMB11 IKVM Power Control

Next, let us get to the performance.

ASUS RS700-E11-RS12U Performance

The CPUs installed in this unit are Intel Xeon Platinum 8452Y units. We configured them to P1 36 cores, 300W TDP, and a 2.0GHz base clock which is the maximum core count for these Intel CPUs. We have been testing a large series of servers with these processors, and that led to an interesting finding, especially since this was the 1U platform.

ASUS RS700 E11 RS12U Intel Xeon Platinum 8452Y Performance To Baseline
ASUS RS700 E11 RS12U Intel Xeon Platinum 8452Y Performance To Baseline

This server, despite only being 1U, performed similarly to the Dell PowerEdge R760 2U server we reviewed as well as the PhoenixNAP Bare Metal Cloud server.

This was notable since ASUS only has 1U of height to work with. In this generation cooling CPUs is a challenge, and so 2U servers can be a bit better in terms of performance just because they can utilize larger coolers. If you want to know why we spent extra time on the cooler design in this review, it was because of this result.

ASUS RS700 E11 RS12U CPU Heatsink 1
ASUS RS700 E11 RS12U CPU Heatsink 1

What was perhaps more interesting, is how ASUS did this while still using a reasonable amount of power. That is what we are going to look at next.

5 COMMENTS

  1. It’s amazing how fast the power consumption is going up in servers…My ancient 8-core per processor 1U HP DL 360 Gen9 in my homelab has 2x 500 Watt platinum power supplies (well also 8x 2.5″ SATA drive slots versus 12 slots in this new server and no on-board M.2 slots and less RAM slots).

    So IF someone gave me one of these new beasts for my basement server lab, would my wife notice that I hijacked the electric dryer circuit? Hmmm.

  2. @carl, you dont need 1600W PSUs to power these servers. honestly, i dont see a usecase when this server uses more than 600W, even with one GPU – i guess ASUS just put the best rated 1U PSU they can find

  3. Data Center (DC) “(server room) Urban renewal” vs “(rack) Missing teeth”: When I first landed in a sea of ~4000 Xeon servers in 2011, until I powered-off my IT career in said facility in 2021, pre-existing racks went from 32 servers per rack to many less per rack (“missing teeth”).

    Yes the cores per socket went from Xeon 4, 6, 8, 10, 12, Eypc Rome 32’s. And yes with each server upgrade I was getting more work done per server, but less servers per rack in the circa 2009 original server rooms in this corporate DC after maxing out power/cooling.

    Yes we upped our power/cooling game at the 12 core Xeon era with immersion cooling, as we built out a new server room. My first group of vats had 104 servers (2U/4node chassis) per 52U vat…The next group of vats with the 32-core Romes, we could not fill (yes still more core per vat though)….So again losing ground on a real estate basis.

    ….

    So do we just agree that as this hockey stick curve of server power usage grows quickly, we live with a growing “missing teeth” issue over upgrade cycles, perhaps start to look at 5 – 8 year “urban renewal” cycles (rebuild of the given server room’s power/cooling infrastructure at great expense) instead of 2010-ish perhaps 10 – 15 year cycles?

    For anyone running their own data center, this will greatly effect their TCO spreadsheets.

  4. @altmind… I am not sure why you can’t imagine it using more than 600w when it used much much more (+200w @70% load, +370w at peak) in the review, all without a gpu.

    @Carl, I think there is room for innovation in the DC space, I don’t see the power/heat density changing and it is not exactly easy to “just run more power” to an existing DC let alone cooling.

  5. Which leads to the current nominal way to get out of the “Urban renewal” vs “Missing teeth” dilemma as demand for compute rises as new compute devices power/cooling needs per unit rise: “Burn down more rain forest” (build more data centers as our cattle herd grows).

    But I’m not sure everyplace wants to be Northern Virginia, nor want to devote a growing % of their energy grid to facilities that use a lot of power (thus requiring more hard-to-site power generation facilities).

    As for “I think there is room for innovation in the DC space”, this seems to be a basic physics problem that I don’t see any solution for on the horizon. Hmmm.

LEAVE A REPLY

Please enter your comment!
Please enter your name here

This site uses Akismet to reduce spam. Learn how your comment data is processed.