Files
YaeBlog/source/posts/hpc-2025-heterogeneous-system.md
jackfiled 462fbb28ac
Some checks failed
Build blog docker image / Build-Blog-Image (push) Failing after 14s
feat: rewrite about page for 2026. (#21)
Signed-off-by: jackfiled <xcrenchangjun@outlook.com>
Reviewed-on: #21
2026-03-03 09:09:49 +00:00

2.2 KiB

title, date, tags
title date tags
High Performance Computing 25 SP Heterogeneous Computing 2025-05-10T00:36:20.5391570+08:00
高性能计算
学习资料

Heterogeneous Computing is on the way!

GPU Computing Ecosystem

CUDA: NVIDIA's Architecture for GPU computing.

image-20250417195644624

Internal Buses

HyperTransport:

Primarily a low latency direct chip to chip interconnect, supports mapping to board to board interconnect such as PCIe.

PCI Expression

Switched and point-to-point connection.

NVLink

image-20250417200241703

OpenCAPI

Heterogeneous computing was in the professional world mostly limited to HPC, in the consumer world is a "nice to have".

But OpenCAPI is absorbed by CXL.

CPU-GPU Arrangement

image-20250424184701573

First Stage: Intel Northbrige

image-20250424185022360

Second Stage: Symmetric Multiprocessors:

image-20250424185048036

Third Stage: Nonuniform Memory Access

And the memory controller is integrated directly in the CPU.

image-20250424185152081

So in such context, the multiple CPUs is called NUMA:

image-20250424185219673

And so there can be multi GPUs:

image-20250424185322963

Fourth Stage: Integrated PCIe in CPU

image-20250424185354247

And there is such team integrated CPU, which integrated a GPU into the CPU chipset.

image-20250424185449577

And the integrated GPU can work with discrete GPUs:

image-20250424185541483

Final Stage: Multi GPU Board

image-20250424190159059