SK hynix AI Memory at Hot Chips 2023

0
SK Hynix Memory Centric Computing With DSM HC35_Page_07
SK Hynix Memory Centric Computing With DSM HC35_Page_07

At Hot Chips 35 (2023) SK hynix is applying its expertise in memory to the big computing problem of today, AI. At the show, it is showing its work on Memory Centric Computing with Domain Specific Memory. The company is looking at ways to alleviate one of the biggest challenges with AI compute today, memory capacity and bandwidth in relation to the compute resources available. Let us get to the talk.

Please note, we are doing this live at the conference. Please excuse the typos. It is a crazy rush of a conference for coverage with back-to-back presentations all day.

SK hynix AI Memory at Hot Chips 2023

Here is SK hynix’s problem definition. Generative AI inference costs are enormous. It is not just the AI compute. It is also the power, interconnects, and memory that also drives a lot of the costs.

SK Hynix Memory Centric Computing With DSM HC35_Page_03
SK Hynix Memory Centric Computing With DSM HC35_Page_03

With large transformer models, memory is a major challenge. These models need a ton of data, and therefore are often memory capacity and bandwidth bound.

SK Hynix Memory Centric Computing With DSM HC35_Page_04
SK Hynix Memory Centric Computing With DSM HC35_Page_04

SK hynix is setting up that the industry needs more than just memory, it needs different types of memory including domain-specific memory with compute built-in. Both Samsung and SK hynix have been working towards becoming in memory compute providers because that is how they move up the value chain.

SK Hynix Memory Centric Computing With DSM HC35_Page_05
SK Hynix Memory Centric Computing With DSM HC35_Page_05

On that note, we are going to hear about Accelerator-in-Memory or SK hynix AiM.

SK Hynix Memory Centric Computing With DSM HC35_Page_06
SK Hynix Memory Centric Computing With DSM HC35_Page_06

Here is a look at the GDDR6 memory where there are banks of memory each with its own 1GHz processing unit capable of 512GB/s of internal bandwidth.

SK Hynix Memory Centric Computing With DSM HC35_Page_08
SK Hynix Memory Centric Computing With DSM HC35_Page_08

SK hynix discussed how it plans to do GEMV in memory for AI compute. Weight matrix data is sourced from banks while vector data comes from the global buffer.

SK Hynix Memory Centric Computing With DSM HC35_Page_09
SK Hynix Memory Centric Computing With DSM HC35_Page_09

There are specific AiM memory commands for the in-memory compute.

SK Hynix Memory Centric Computing With DSM HC35_Page_10
SK Hynix Memory Centric Computing With DSM HC35_Page_10

SK hynix is showing how memory scales and the need for AiM in-memory compute resources are required for large language models.

SK Hynix Memory Centric Computing With DSM HC35_Page_12
SK Hynix Memory Centric Computing With DSM HC35_Page_12

Here is how these scale out for large language models (LLMs):

SK Hynix Memory Centric Computing With DSM HC35_Page_13
SK Hynix Memory Centric Computing With DSM HC35_Page_13

One of the big challenges is that using this type of AiM is that it requires mapping from the software side, hardware architected for AiM, and then an interface. This is one of the other big barriers for adoption.

SK Hynix Memory Centric Computing With DSM HC35_Page_14
SK Hynix Memory Centric Computing With DSM HC35_Page_14

Here is how SK hynix is looking at doing the mapping from problems to AiM.

SK Hynix Memory Centric Computing With DSM HC35_Page_15
SK Hynix Memory Centric Computing With DSM HC35_Page_15

The system architecture needs to handle scale-up and scale-out.

SK Hynix Memory Centric Computing With DSM HC35_Page_16
SK Hynix Memory Centric Computing With DSM HC35_Page_16

Key components of the AIM architecture are the AiM controller, Scalable Multicasting Interconnect, Router, Compute Unit (ALU), and an Instruction Sequencer.

SK Hynix Memory Centric Computing With DSM HC35_Page_17
SK Hynix Memory Centric Computing With DSM HC35_Page_17

Matrix vector accumulate functions are key to AI workloads. SK hynix AiM uses a CISC-like instruction set to manage this.

SK Hynix Memory Centric Computing With DSM HC35_Page_18
SK Hynix Memory Centric Computing With DSM HC35_Page_18

The next step of this is optimization. With a new architecture, often there are nuances that can be exploited to get better performance.

SK Hynix Memory Centric Computing With DSM HC35_Page_19
SK Hynix Memory Centric Computing With DSM HC35_Page_19

SK hynix did not just talk about AiM in the abstract. Instead, it showed a proof of concept GDDR6 AiM solution using two FPGAs.

SK Hynix Memory Centric Computing With DSM HC35_Page_21
SK Hynix Memory Centric Computing With DSM HC35_Page_21

It also showed its software stack for AiM.

SK Hynix Memory Centric Computing With DSM HC35_Page_22
SK Hynix Memory Centric Computing With DSM HC35_Page_22

It does not sound like SK hynix is looking to sell these cards, instead, these seem like they are being used to prove out the concept.

SK Hynix Memory Centric Computing With DSM HC35_Page_24
SK Hynix Memory Centric Computing With DSM HC35_Page_24

SK hynix is still in the evaluation stage doing different types of analysis on the solution versus more traditional solutions.

SK Hynix Memory Centric Computing With DSM HC35_Page_25
SK Hynix Memory Centric Computing With DSM HC35_Page_25

So this is being looked at for the future.

Final Words

Both SK hynix and Samsung have been talking about in-memory compute for years. It will be interesting to see if a big customer adopts this in the future. For now, it seems like the next generations of AI compute will be more traditional in nature, but perhaps this is one of the areas that will take off in a few years.

LEAVE A REPLY

Please enter your comment!
Please enter your name here

This site uses Akismet to reduce spam. Learn how your comment data is processed.