Today finally marks the official launch of Intel’s 3rd Gen Ice Lake-SP Xeon CPUs which will be the company’s first server lineup to be made on the 10nm process node. The Ice Lake-SP lineup aims at the server segment and will be competing against AMD’s EPYC Milan family which launched last month.
Intel 3rd Gen Ice Lake-SP Xeon CPUs Officially Launched – 10nm Comes To Servers With Up To 40 Cores, 270W TDP, 64 PCIe Gen 4.0 Lanes & More
Starting off with the details, Intel’s 3rd Gen Ice Lake-SP Xeon CPUs will be based on the 10nm+ process node and utilize the Sunny Cove core architecture. The Intel Sunny Cove x86 architecture has been around since 2019 and was first featured on Intel’s 10th Gen Ice Lake processors for the notebook segment. Intel has since moved over to Tiger Lake which is based on the Willow Cove x86 architecture and makes use of the 10nm SuperFin process node.
Some of the major upgrades that Intel’s 10nm+ for Ice Lake-SP Xeon CPU will deliver include:
- 2.7x density scaling vs 14nm
- Self-aligned Quad-Patterning
- Contact Over Active Gate
- Cobalt Interconnect (M0, M1)
- 1st Gen Foveros 3D Stacking
- 2nd Gen EMIB
Some of the major highlights of today’s Intel Xeon announcement include:
- New 3rd Gen Intel Xeon Scalable processors — combined with Intel’s portfolio of Intel® Optane™ persistent memory and storage, Ethernet adapters, FPGAs, and optimized software solutions — deliver performance and workload optimizations across hybrid-cloud, high-performance computing (HPC), networking, and intelligent edge applications.
- New 3rd Gen Intel Xeon Scalable processors feature a flexible architecture with integrated artificial intelligence (AI) acceleration with Intel® DL Boost technology, plus advanced security capabilities to protect data and application code with Intel® Software Guard Extension (Intel® SGX) and Intel Crypto Acceleration.
- New 3rd Gen Intel Xeon Scalable processors are rapidly ramping, having shipped more than 200,000 units for revenue in the first quarter of 2021 with broad industry adoption across all market segments, including the world’s top cloud service providers set to deploy services; more than 250 design wins within 50 unique OxM partners; more than 15 major telecom equipment manufacturers and communications service providers readying POCs and networking deployments, and over 20 HPC labs and HPC-as-a-service environments leveraging our latest Xeon Scalable processors.
Intel’s Ice Lake-SP will ship in two die configurations, XCC (Extreme Core Count) and HCC (High Core Count). The XCC SKUs will feature 16, 18, 28, 32, 36, 38, and up to 40 cores. The HCC SKUs will feature 8, 12, 16, 18, 20, 24, 26, and up to 28 cores. The TDPs will range from 105, 135, 150, 165, 185, 205, 220, 235, 250, and all the way up to 270W for the flagship SKU. The XCC variants with 32, 36, 38, and 40 cores will be configurated at around 205-270W TDPs.
Intel’s Xeon Ice Lake-SP lineup is split between Platinum, Gold, and Silver SKUs. The top SKU is said to be the Xeon Platinum 8380 which would feature 40 cores, a base clock of 2.30 GHz, and a 270W TDP. Compared to AMD’s flagship, we are looking at lower cores and a lower base clock however, Intel’s architecture and process nodes are tuned for higher clock speeds so that might mark a small win for the blue team. Intel is also placing bets on its AVX-512 instruction set and has shown CPU benchmarks where it handily beats AMD’s existing Rome CPUs but only when running applications that fully leverage AVX-512 instructions. Standard applications would not see the same benefit on Intel CPUs.
Intel Xeon SP Families:
|Family Branding||Skylake-SP||Cascade Lake-SP/AP||Cooper Lake-SP||Ice Lake-SP||Sapphire Rapids||Emerald Rapids||Granite Rapids||Diamond Rapids|
|Process Node||14nm+||14nm++||14nm++||10nm+||10nm SuperFin?||10nm Enhanced SuperFin?||7nm?||sub-7nm?|
|Platform Name||Intel Purley||Intel Purley||Intel Cedar Island||Intel Whitley||Intel Eagle Stream||Intel Eagle Stream||Intel Mountain Stream
Intel Birch Stream
|Intel Mountain Stream
Intel Birch Stream
|MCP (Multi-Chip Package) SKUs||No||Yes||No||No||TBD||TBD||TBD (Possibly Yes)||TBD (Possibly Yes)|
|Socket||LGA 3647||LGA 3647||LGA 4189||LGA 4189||LGA 4677||LGA 4677||LGA 4677||TBD|
|Max Core Count||Up To 28||Up To 28||Up To 28||Up To 40||TBD||TBD||TBD||TBD|
|Max Thread Count||Up To 56||Up To 56||Up To 56||Up To 80||TBD||TBD||TBD||TBD|
|Max L3 Cache||38.5 MB L3||38.5 MB L3||38.5 MB L3||60 MB L3||TBD||TBD||TBD||TBD|
|Memory Support||DDR4-2666 6-Channel||DDR4-2933 6-Channel||Up To 6-Channel DDR4-3200||Up To 8-Channel DDR4-3200||Up To 8-Channel DDR5-4800||Up To 8-Channel DDR5-5200?||TBD||TBD|
|PCIe Gen Support||PCIe 3.0 (48 Lanes)||PCIe 3.0 (48 Lanes)||PCIe 3.0 (48 Lanes)||PCIe 4.0 (64 Lanes)||PCIe 5.0||PCIe 5.0||PCIe 6.0?||PCIe 6.0?|
|3D Xpoint Optane DIMM||N/A||Apache Pass||Barlow Pass||Barlow Pass||Crow Pass||Crow Pass?||Donahue Pass?||Donahue Pass?|
|Competition||AMD EPYC Naples 14nm||AMD EPYC Rome 7nm||AMD EPYC Rome 7nm||AMD EPYC Milan 7nm+||AMD EPYC Genoa ~5nm||AMD Next-Gen EPYC (Post Genoa)||AMD Next-Gen EPYC (Post Genoa)||AMD Next-Gen EPYC (Post Genoa)|
The Ice Lake-SP CPUs will be supported by the Whitley platform that makes use of the LGA 4189 socket (socket P+). The Whitley platform features 2-way CPU support which will be interconnected with a UPI (Ultra Path Interconnect). The CPUs will be connected to the Intel C620A chipset through DMI & the chipset itself will feature up to 20 PCIe Gen 3 lanes, 10 USB 3.0 ports, and 14 SATA Gen 3 ports. As for the CPUs, they will offer up to 8-channel memory support in DDR4-3200 (1 DPC) or DDR4-2933 (2 DPC) modes. The Ice Lake-SP processors will feature up to 64 PCIe Gen 4.0 lanes.
- Built-in AI acceleration: The latest 3rd Gen Intel Xeon Scalable processors deliver the AI performance, productivity, and simplicity that enable customers to unlock more valuable insights from their data. As the only data center CPU with built-in AI acceleration, extensive software optimizations, and turnkey solutions, the new processors make it possible to infuse AI into every application from edge to network to cloud. The latest hardware and software optimizations deliver 74% faster AI performance compared with the prior generation and provide up to 1.5 times higher performance across a broad mix of 20 popular AI workloads versus AMD EPYC 7763 and up to 1.3 times higher performance on a broad mix of 20 popular AI workloads versus Nvidia A100 GPU.
- Built-in security: With hundreds of research studies and production deployments, plus the ability to be continuously hardened over time, Intel SGX protects sensitive code and data with the smallest potential attack surface within the system. It is now available on 2-socket Xeon Scalable processors with enclaves that can isolate and process up to 1 terabyte of code and data to support the demands of mainstream workloads. Combined with new features, including Intel® Total Memory Encryption and Intel Platform Firmware Resilience, the latest Xeon Scalable processors address today’s most pressing data protection concerns.
- Built-in crypto acceleration: Intel Crypto Acceleration delivers breakthrough performance across a host of important cryptographic algorithms. Businesses that run encryption-intensive workloads, such as online retailers who process millions of customer transactions per day, can leverage this capability to protect customer data without impacting user response times or overall system performance.
The 2nd Gen Optane DC Persistent memory would be supported on both platforms (Barlow Pass), offering up to 3200 MT/s and 15% bandwidth improvement in a 15W DIMM. Whitley would be able to support up to 3200 MT/s and 4 TB capacity per socket.
The lineup starts with 8 cores and goes all the way up to 40 cores. The lineup is quite confusing if compared to the AMD EPYC parts so you can see the table below to get a sense of what Intel will be offering in its 3rd Gen Xeon Ice Lake-SP stack:
Intel Xeon Ice Lake-SP Server CPU Lineup (Preliminary):
|CPU Name||Cores / Threads||Base Clock||Boost Clock||L3 Cache||L2 Cache||TDP|
|Xeon Platinum 8380||40 / 80||2.30 GHz||TBA||60 MB||50.00 MB||270W|
|Xeon Platinum 8368||38 / 76||2.40 GHz||TBA||57 MB||47.50 MB||270W|
|Xeon Platinum 8360Y||36 / 72||2.40 GHz||TBA||54 MB||45.00 MB||250W|
|Xeon Platinum 8358||32 / 64||2.65 GHz||TBA||48 MB||40.00 MB||250W|
|Xeon Platinum 8358P||32 / 64||2.60 GHz||TBA||48 MB||40.00 MB||240W|
|Xeon Platinum 8352S||32 / 64||2.20 GHz||3.40 GHz||48 MB||40.00 MB||205W|
|Xeon Platinum 8352Y||32 / 64||2.20 GHz||3.40 GHz||48 MB||40.00 MB||205W|
|Xeon Platinum 8352V||36 / 72||2.10 GHz||TBA||54 MB||45.00 MB||195W|
|Xeon Platinum 8351N||36 / 72||2.40 GHz||3.60 GHz||54 MB||45.00 MB||225W|
|Xeon Gold 6354||18 / 36||3.10 GHz||TBA||27 MB||22.50 MB||205W|
|Xeon Gold 6348||28 / 56||2.80 GHz||TBA||42 MB||35.00 MB||235W|
|Xeon Gold 6346||16 / 32||3.10 GHz||TBA||24 MB||20.00 MB||205W|
|Xeon Gold 6342||24 / 48||2.70 GHz||TBA||36 MB||30.00 MB||220W|
|Xeon Gold 6338||32 / 64||2.00 GHz||TBA||48 MB||40.00 MB||205W|
|Xeon Gold 6336Y||24 / 48||2.40 GHz||TBA||36 MB||30.00 MB||185W|
|Xeon Gold 6334||8 / 16||3.50 GHz||TBA||12 MB||10.00 MB||165W|
|Xeon Gold 6330||28 / 56||2.00 GHz||TBA||42 MB||35.00 MB||205W|
|Xeon Gold 6326||16 / 32||2.80 GHz||TBA||24 MB||20.00 MB||185W|
|Xeon Gold 5320||26 / 52||2.20 GHz||TBA||39 MB||16.25 MB||185W|
|Xeon Gold 5318Y||24 / 48||2.00 GHz||TBA||36 MB||30.00 MB||165W|
|Xeon Gold 5317||12 / 24||2.80 GHz||TBA||12 MB||15.00 MB||150W|
|Xeon Gold 5315Y||8 / 16||3.00 GHz||TBA||12 MB||10.00 MB||150W|
|Xeon Silver 4316||20 / 40||2.30 GHz||TBA||30 MB||25.00 MB||150W|
|Xeon Silver 4314||16 / 32||2.30 GHz||TBA||24 MB||20.00 MB||135W|
|Xeon Silver 4310||12 / 24||2.10 GHz||TBA||12 MB||15.00 MB||135W|
|Xeon Silver 4309Y||8 / 16||2.60 GHz||TBA||12 MB||10.00 MB||105W|