Intel Sapphire Rapids-SP Xeon Server CPU Detailed: Quad-Tile Chiplet Design With EMIB, 56 Cores, 112 Threads, CXL 1.1, DDR5, HBM & PCIe 5.0 Support

Hassan Mujtaba
Intel Xeon W9-3495 Sapphire Rapids HEDT CPU Spotted - Rocks 56 Cores, 112 Threads & a 1.80 GHz ES Clock Speed

Intel has officially detailed its next-generation Sapphire Rapids-SP CPU lineup which will be part of the 4th Gen Xeon Scalable family. The Intel Sapphire Rapids-SP lineup will consist of a range of new technologies with the most important being the seamless integration of multiple chiplets or 'Tiles', as Intel refers to them, through their EMIB technology.

Intel Fully Details Next-Gen Sapphire Rapids-SP Xeon CPUs, Multi-Tile Chiplet Design Based on 'Intel 7' Process Node

The Sapphire Rapids-SP family will be replacing the Ice Lake-SP family and will go all on board with the 'Intel 7' process node (formerly 10nm Enhanced SuperFin) that will be making its formal debut later this year in the Alder Lake consumer family. The server lineup will feature the performance-optimized Golden Cove core architecture which delivers a 20% IPC improvement over Willow Cove core architecture. Several cores are featured on multiple tiles and packaged together through the use of EMIB.

Related Story Intel Arc Battlemage “Xe2” BMG-10 & BMG-21 GPUs Confirmed: Pre Qualification Samples For Gaming Graphics Cards

Sapphire Rapids: Combining Intel’s Performance-cores with new accelerator engines, Sapphire Rapids sets the standard for next-generation data center processors. At the heart of Sapphire Rapids is a tiled, modular SoC architecture that delivers significant scalability while still maintaining the benefits of a monolithic CPU interface thanks to Intel’s EMIB multi-die interconnect packaging technology and advanced mesh architecture.

intel-architecture-day-2021_pressdeck_final_embargo-compressed-115
intel-sapphire-rapids-sp-xeon-cpu-architecture-day-2021-_2
intel-architecture-day-2021_pressdeck_final_embargo-compressed-116
intel-architecture-day-2021-5-kottapalli-custom

For Sapphire Rapids-SP, Intel is using a quad multi-tile chiplet design which will come in HBM and non-HBM flavors. While each tile is its own unit, the chip itself acts as one singular SOC and each thread has full access to all resources on all tiles, consistently providing low-latency & high cross-section bandwidth across the entire SOC. Each tile is further composed of three main IP blocks & which are detailed below:

Compute IP

  • Cores
  • Acceleration Engines

I/O IP

  • CXL 1.1
  • PCIe Gen 5
  • UPI 2.0

Memory IP

  • DDR5
  • Optane
  • HBM
intel-architecture-day-2021_pressdeck_final_embargo-compressed-117
intel-architecture-day-2021_pressdeck_final_embargo-compressed-118
intel-architecture-day-2021_pressdeck_final_embargo-compressed-119
intel-architecture-day-2021_pressdeck_final_embargo-compressed-120
intel-architecture-day-2021_pressdeck_final_embargo-compressed-121
intel-architecture-day-2021_pressdeck_final_embargo-compressed-122
intel-architecture-day-2021_pressdeck_final_embargo-compressed-123

We have already taken an in-depth look at the P-Core over here but some of the key changes that will be offered to the data center platform will include AMX, AiA, FP16, and CLDEMOTE capabilities. The Accelerator Engines will increase the effectiveness of each core by offloading common-mode tasks to these dedicated accelerator engines which will increase performance & decrease the time taken to achieve the necessary task.

intel-architecture-day-2021_pressdeck_final_embargo-compressed-125
intel-architecture-day-2021_pressdeck_final_embargo-compressed-126

In terms of I/O advancements, Sapphire Rapids-SP Xeon CPUs will introduce CXL 1.1 for accelerator and memory expansion in the data center segment. There's also an improved multi-socket scaling via Intel UPI, delivering up to 4 x24 UPI links at 16 GT/s and a new 8S-4UPI performance-optimized topology. The new tile architecture design also boosts the cache beyond 100 MB along with Optane Persistent Memory 300 series support.

intel-architecture-day-2021_pressdeck_final_embargo-compressed-127
intel-architecture-day-2021_pressdeck_final_embargo-compressed-128
intel-architecture-day-2021_pressdeck_final_embargo-compressed-129

Intel has also detailed its Sapphire Rapids-SP Xeon CPUs with HBM memory. From what Intel has shown, their Xeon CPUs will house up to four HBM packages, all offering significantly higher DRAM bandwidth versus a baseline Sapphire Rapids-SP Xeon CPU with 8-channel DDR5 memory. This is going to allow Intel to offer a chip with both increased capacity and bandwidth for customers that demand it. The HBM SKUs can be used in two modes, an HBM Flat mode & an HBM caching mode.

Intel also showed a demo of their Sapphire Rapids-SP Xeon CPUs running an internal GEMM Kernel with and without AMX instructions. The AMX enabled solution delivered a 7.8x improvement over the non-AMX solution. This demo was also from early silicon so final performance may further improve. Intel didn't disclose any additional details regarding the test platform.

Intel Sapphire Rapids-SP Xeon CPU Platform

The Sapphire Rapids lineup will make use of 8 channel DDR5 memory with speeds of up to 4800 Mbps & support PCIe Gen 5.0 on the Eagle Stream platform. The Eagle Stream platform will also introduce the LGA 4677 socket which will be replacing the LGA 4189 socket for Intel's upcoming Cedar Island & Whitley platform which would house Cooper Lake-SP and Ice Lake-SP processors, respectively. The Intel Sapphire Rapids-SP Xeon CPUs will also come with CXL 1.1 interconnect that will mark a huge milestone for the blue team in the server segment.

Coming to the configurations, the top part is started to feature 56 cores with a TDP of 350W. What is interesting about this configuration is that it is listed as a low-bin split variant which means that it will be using a tile or MCM design. The Sapphire Rapids-SP Xeon CPU will be composed of a 4-tile layout with each tile featuring 14 cores each.

Following are the leaked configurations:

  • Sapphire Rapids-SP 24 Core / 48 Thread / 45.0 MB / 225W
  • Sapphire Rapids-SP 28 Core / 56 Thread / 52.5 MB / 250W
  • Sapphire Rapids-SP 40 Core / 48 Thread / 75.0 MB / 300W
  • Sapphire Rapids-SP 44 Core / 88 Thread / 82.5 MB / 270W
  • Sapphire Rapids-SP 48 Core / 96 Thread / 90.0 MB / 350W
  • Sapphire Rapids-SP 56 Core / 112 Thread / 105 MB / 350W

It looks like AMD will still hold the upper hand in the number of cores & threads offered per CPU with their Genoa chips pushing for up to 96 cores whereas Intel Xeon chips would max out at 56 cores if they don't plan on making SKUs with a higher number of tiles. Intel will have a wider and more expandable platform that can support up to 8 CPUs at once so unless Genoa offers more than 2P (dual-socket) configurations, Intel will have the lead in the most number of cores per rack with an 8S rack packing up to 448 cores and 896 threads.

The Intel Saphhire Rapids CPUs will contain 4 HBM2 stacks with a maximum memory of 64 GB (16GB each). The presence of memory so near to the die would do absolute wonders for certain workloads that require huge data sets and will basically act as an L4 cache.

AMD has been taking away quite a few wins from Intel as seen in the recent Top500 charts from ISC '21. Intel would really have to up their game in the next couple of years to fight back the AMD EPYC threat. Intel is expected to launch Sapphire Rapids-SP in 2022 followed by HBM variants that are expected to launch around 2023.

Which server lineup do you think will offer the best feature set for enterprise markets?

Intel Xeon CPU Families (Preliminary):

Family BrandingDiamond RapidsClearwater ForestGranite RapidsSierra ForestEmerald RapidsSapphire RapidsIce Lake-SPCooper Lake-SPCascade Lake-SP/APSkylake-SP
Process NodeIntel 20A?Intel 18AIntel 3Intel 3Intel 7Intel 710nm+14nm++14nm++14nm+
Platform NameIntel Mountain Stream
Intel Birch Stream
Intel Mountain Stream
Intel Birch Stream
Intel Mountain Stream
Intel Birch Stream
Intel Mountain Stream
Intel Birch Stream
Intel Eagle StreamIntel Eagle StreamIntel WhitleyIntel Cedar IslandIntel PurleyIntel Purley
Core ArchitectureLion Cove?DarkmontRedwood CoveSierra GlenRaptor CoveGolden CoveSunny CoveCascade LakeCascade LakeSkylake
MCP (Multi-Chip Package) SKUsYesTBDYesYesYesYesNoNoYesNo
SocketLGA 4677 / 7529LGA 4677 / 7529LGA 4677 / 7529LGA 4677 / 7529LGA 4677LGA 4677LGA 4189LGA 4189LGA 3647LGA 3647
Max Core CountUp To 144?Up To 288Up To 136?Up To 288Up To 64?Up To 56Up To 40Up To 28Up To 28Up To 28
Max Thread CountUp To 288?Up To 288Up To 272?Up To 288Up To 128Up To 112Up To 80Up To 56Up To 56Up To 56
Max L3 CacheTBDTBD480 MB L3108 MB L3320 MB L3105 MB L360 MB L338.5 MB L338.5 MB L338.5 MB L3
Memory SupportUp To 12-Channel DDR6-7200?TBDUp To 12-Channel DDR5-6400Up To 8-Channel DDR5-6400?Up To 8-Channel DDR5-5600Up To 8-Channel DDR5-4800Up To 8-Channel DDR4-3200Up To 6-Channel DDR4-3200DDR4-2933 6-ChannelDDR4-2666 6-Channel
PCIe Gen SupportPCIe 6.0 (128 Lanes)?TBDPCIe 5.0 (136 Lanes)PCIe 5.0 (TBD Lanes)PCIe 5.0 (80 Lanes)PCIe 5.0 (80 lanes)PCIe 4.0 (64 Lanes)PCIe 3.0 (48 Lanes)PCIe 3.0 (48 Lanes)PCIe 3.0 (48 Lanes)
TDP Range (PL1)Up To 500W?TBDUp To 500WUp To 350WUp To 350WUp To 350W105-270W150W-250W165W-205W140W-205W
3D Xpoint Optane DIMMDonahue Pass?TBDDonahue PassTBDCrow PassCrow PassBarlow PassBarlow PassApache PassN/A
CompetitionAMD EPYC VeniceAMD EPYC Zen 5CAMD EPYC TurinAMD EPYC BergamoAMD EPYC Genoa ~5nmAMD EPYC Genoa ~5nmAMD EPYC Milan 7nm+AMD EPYC Rome 7nmAMD EPYC Rome 7nmAMD EPYC Naples 14nm
Launch2025?202520242024202320222021202020182017
Share this story

Deal of the Day

Comments