Loading AI tools
2024 AMD 4-nanometer processor microarchitecture From Wikipedia, the free encyclopedia
Zen 5 is the name for a CPU microarchitecture by AMD, shown on their roadmap in May 2022,[3] launched for mobile in July 2024 and for desktop in August 2024.[4] It is the successor to Zen 4 and is currently fabricated on TSMC's N4X process.[5] Zen 5 is also planned to be fabricated on the N3E process in the future.[6]
General information | |
---|---|
Launched | Mobile July 17, 2024 Desktop August 8, 2024 Server October 10, 2024 |
Designed by | AMD |
Common manufacturer | |
CPUID code | Family 1Ah |
Cache | |
L1 cache | 80 KB (per core):
|
L2 cache | 1 MB (per core) |
L3 cache |
|
Architecture and classification | |
Technology node | TSMC N4X (Zen 5 CCD) TSMC N3E (Zen 5c CCD) TSMC N6 (IOD) TSMC N4P (Mobile) |
Instruction set | AMD64 (x86-64) |
Extensions | |
Physical specifications | |
Cores |
|
Memory (RAM) | |
Sockets |
|
Products, models, variants | |
Product code names | |
Brand names | |
History | |
Predecessor | Zen 4 |
Successor | Zen 6 |
The Zen 5 microarchitecture powers Ryzen 9000 series desktop processors (codenamed "Granite Ridge"), Epyc 9005 server processors (codenamed "Turin"),[7] and Ryzen AI 300 thin and light mobile processors (codenamed "Strix Point").[8]
Zen 5 was first officially mentioned during AMD's Ryzen Processors: One Year Later presentation on April 9, 2018.[9]
A roadmap shown during AMD's Financial Analyst Day on June 9, 2022 confirmed that Zen 5 and Zen 5c would be launching in 3nm and 4nm variants in 2024.[10] The earliest details on the Zen 5 architecture promised a "re-pipelined front end and wide issue" with "integrated AI and Machine Learning optimizations".
During AMD's Q4 2023 earnings call on January 30, 2024, AMD CEO Lisa Su stated that Zen 5 products would be "coming in the second half of the year".[11]
Zen 5 is a ground-up redesign of Zen 4 with a wider front-end, increased floating point throughput and more accurate branch prediction.[12]
Zen 5 was designed with both 4nm and 3nm processes in mind. This acted as an insurance policy for AMD in the event that TSMC's mass production of its N3 nodes were to face delays, significant wafer defect issues or capacity issues. One industry analyst estimated early N3 wafer yields to be at 55% while others estimated yields to be similar to those of N5 at between 60-80%.[13][14] Additionally, Apple, as TSMC's largest customer, is given priority access to the latest process nodes. In 2022, Apple was responsible for 23% of TSMC's $72 billion in total revenue.[15] After N3 began ramping at the end of 2022, Apple bought up the entirety of TSMC's early N3B wafer production capacity to fabricate their A17 and M3 SoCs.[16] Zen 5 desktop and server processors continue to use the N6 node for the I/O die fabrication.[17]
Zen 5 CCDs are fabricated on TSMC's N4X node which is intended to accommodate higher frequencies for high-performance computing (HPC) applications.[18] Zen 4-based mobile processors were fabricated on the N4P node which is targeted more towards power efficiency. N4X maintains IP compatibility with N4P and offers a 6% frequency gain over N4P at the same power but comes with the trade-off of moderate leakage.[19] Compared to the N5 node used to produce Zen 4 CCDs, N4X can enable up to 15% higher frequencies while running at 1.2V.[20]
The Zen 5 CCD, codenamed "Eldora", has a die size of 70.6mm2, a 0.5% reduction in area from Zen 4's 71mm2 CCD while achieving a 28% increase in transistor density due to the N4X process node.[21] Zen 5's CCD contains 8.315 billion transistors compared to the Zen 4 CCD's 6.5 billion transistors.[22] The size of an individual Zen 5 core is actually larger than a Zen 4 core but the CCD has been reduced via shrinking the L3 cache. The monolithic die used by "Strix Point" mobile processors, fabricated on TSMC's lower power N4P node, measures 232.5mm2 in area.[21]
Zen 5's changes to branch prediction are the most significant divergence from any previous Zen microarchitecture. The branch predictor in a core tries to predict the outcome when there are diverging code paths. Zen 5's branch predictor is able to operate two-ahead where it can try to predict two code paths ahead before they are executed rather predicting one code path, waiting for it to be executed, then predicting the next one.[23] Two-ahead branch predictors have been discussed in academic research dating back to André Seznec et al.'s 1996 paper "Multiple-block ahead branch predictors".[24] 28 years after it was first proposed in academic research, AMD's Zen 5 architecture became the first microarchitecture to fully implement two-ahead branch prediction. Increased data prefetching assists the branch predictor.
Zen 5 contains 6 Arithmetic Logic Units (ALUs), up from 4 ALUs in prior Zen architectures. A greater number of ALUs that handle common integer operations can increase per-cycle scalar integer throughput by 50%.[25]
The vector engine in Zen 5 features 4 floating point pipes compared to 3 pipes in Zen 4. Zen 4 introduced AVX-512 instructions. AVX-512 capabilities have been expanded with Zen 5 with a doubling of the floating point pipe width to a native 512-bit floating point datapath. The AVX-512 datapath is configurable depending on the product. Ryzen 9000 series desktop processors and EPYC 9005 server processors feature the full 512-bit datapath but Ryzen AI 300 mobile processors feature a 256-bit datapath in order to reduce power consumption. AVX-512 instruction has been extended to VNNI/VEX instructions. Additionally, there is greater bfloat16
throughput which is beneficial for AI workloads.
The wider front end in the Zen 5 architecture necessitates larger caches and higher memory bandwidth in order to keep the cores fed with data. The L1 cache per core is increased from 64 KB to 80 KB per core. The L1 instruction cache remains the same at 32 KB but the L1 data cache is increased from 32 KB to 48 KB per core. Furthermore, the bandwidth of the L1 data cache for 512-bit floating point unit pipes has also been doubled. The L1 data cache's associativity has increased from 8-way to 12-way in order to accommodate its larger size.
The L2 cache remains at 1 MB but its associativity has increased from 8-way to 16-way. Zen 5 also has a doubled L2 cache bandwidth of 64 bytes per clock.
The L3 cache is filled from L2 cache victims and in-flight misses. Latency for accessing the L3 cache has been reduced by 3.5 cycles.[26] A Zen 5 Core Complex Die (CCD) contains 32 MB of L3 cache shared between the 8 cores. In Zen 5 3D V-Cache CCDs, a piece of silicon containing 64 MB of extra L3 cache is placed under the cores rather than on top like in prior generations for a total of 96 MB.[27] This allows for increased core frequency compared to previous generation 3D V-Cache implementations which were sensitive to higher voltages. The Zen 5-based Ryzen 7 9800X3D has a 500 MHz increased base frequency over the Zen 4-based Ryzen 7 7800X3D and allows overclocking for the first time.[28]
Ryzen AI 300 APUs, codenamed "Strix Point", features 24 MB of total L3 cache which is split into two separate cache arrays. 16 MB of dedicated L3 cache is shared the 4 Zen 5 cores and 8 MB is shared by the 8 Zen 5c cores.[29] Zen 5c cores are not able to access the 16 MB L3 cache array and vice versa.[30]
Cache | Zen 4 | Zen 5 | |
---|---|---|---|
L1 Data |
Size | 32 KB | 48 KB |
Associativity | 8-way | 12-way | |
Bandwidth | 32B/clk | 64B/clk | |
L1 Instructions |
Size | 32 KB | 32 KB |
Associativity | 8-way | 8-way | |
Bandwidth | 64B/clk | 64B/clk | |
L2 | Size | 1 MB | 1 MB |
Associativity | 8-way | 16-way | |
Bandwidth | 32B/clk | 64B/clk | |
L3 | Size | 32 MB | 32 MB |
Associativity | 16-way | 16-way | |
Bandwidth | 32B/clk Read 16B/clk Write |
32B/clk Read 16B/clk Write | |
Other features and changes in the Zen 5 architecture, compared to Zen 4, include:
Attribute | Zen 4 | Zen 5 |
---|---|---|
L1/L2 BTB | 1.5K/7K | 16K/8K |
Return Address Stack | 32 | 52 |
ITLB L1/L2 | 64/512 | 64/2048 |
Fetched/Decoded Instruction Bytes/cycle | 32 | 64 |
Op Cache associativity | 12-way | 16-way |
Op Cache bandwidth | 9 macro-ops | 12 inst or fused inst |
Dispatch bandwidth (macro-ops/cycle) | 6 | 8 |
AGU Scheduler | 3x24 ALU/AGU | 56 |
ALU Scheduler | 1x24 ALU | 88 |
ALU/AGU | 4/3 | 6/4 |
Int PRF (red/flag) | 224/126 | 240/192 |
Vector Reg | 192 | 384 |
FP Pre-Sched Queue | 64 | 96 |
FP Scheduler | 2x32 | 3x38 |
FP Pipes | 3 | 4 |
Vector Width | 256 | 256b/512b |
ROB/Retire Queue | 320 | 448 |
LS Mem Pipes support Load/Store | 3/1 | 4/2 |
DTLB L1/L2 | 72/3072 | 96/4096 |
AMD announced an initial lineup of four models of Ryzen 9000 processors on June 3, 2024, including one Ryzen 5, one Ryzen 7 and two Ryzen 9 models. Manufactured on a 4 nm process, the processors feature between 6 and 16 cores.[34] Ryzen 9000 processors were released in August 2024.
Common features of Ryzen 9000 desktop CPUs:
Branding and Model | Cores (threads) |
Clock rate (GHz) | L3 cache (total) |
TDP | Chiplets | Core config[lower-roman 1] |
Release date |
Launch price[lower-alpha 1] | ||
---|---|---|---|---|---|---|---|---|---|---|
Base | Boost | |||||||||
Ryzen 9 | 9950X[35][36] | 16 (32) | 4.3 | 5.7 | 64 MB | 170 W | 2 × CCD 1 × I/OD |
2 × 8 | August 15, 2024 | US $649 |
9900X[35][36] | 12 (24) | 4.4 | 5.6 | 120 W | 2 × 6 | US $499 | ||||
Ryzen 7 | 9800X3D[37][38] | 8 (16) | 4.7 | 5.2 | 96 MB | 1 × CCD 1 × I/OD |
1 × 8 | November 7, 2024 | US $479 | |
9700X[35][36] | 3.8 | 5.5 | 32 MB | 65 W[lower-alpha 2] | August 8, 2024 | US $359 | ||||
Ryzen 5 | 9600X[35][36] | 6 (12) | 3.9 | 5.4 | 1 × 6 | US $279 |
The Ryzen AI 300 series of high-performance ultrathin notebook processors were announced on June 3, 2024. Codenamed Strix Point, these processors are named under a new model numbering system similar to Intel's Core and Core Ultra model numbering. Strix Point features a 3rd gen Ryzen AI engine based on XDNA 2, providing up to 50 TOPS of neural processing unit performance. The integrated graphics is upgraded to RDNA 3.5, and top end models have 16 CUs of GPU and 12 cores of CPU, an increase from the maximum of 8 CPU cores on previous generation Ryzen ultrathin mobile processors.[39] Notebooks featuring Ryzen AI 300 series processors were released on July 17.[40]
Common features of Ryzen AI 300 notebook APUs:
Branding and Model | CPU | GPU | NPU (Ryzen AI) |
TDP | Release date | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Cores (threads) | Clock (GHz) | L3 cache (total) |
Model | Clock (GHz) | ||||||||
Total | Zen 5 | Zen 5c | Base | Boost[lower-alpha 1] | ||||||||
Ryzen AI 9 | (PRO) HX 375 |
12 (24) | 4 (8) | 8 (16) | 2.0 | 5.1 | 24 MB | 890M 16 CUs |
2.9 | 55 TOPS | 15–54 W | June 2, 2024 [42] |
(PRO) HX 370[43] |
50 TOPS | |||||||||||
365[43] | 10 (20) | 6 (12) | 5.0 | 880M 12 CUs | ||||||||
Ryzen AI 7 | PRO 360[44][45] | 8 (16) | 3 (6) | 5 (10) | 16 MB | October 10, 2024 [46] |
Alongside Granite Ridge desktop and Strix Point mobile processors, the Epyc 9005 series of high-performance server processors, codenamed Turin, were also announced at Computex on June 3, 2024. It uses the same SP5 socket as the previous Epyc 9004 series processors, and will pack up to 128 cores and 256 threads on the top-end model. Turin will be built on a TSMC 4 nm process.[47]
Common features of EPYC 9000 server processers:
Branding and Model | Cores (threads) |
Clock rate (GHz) | L3 cache (total) |
TDP | Chiplets | Core config[lower-roman 1] |
Release date |
Launch price[lower-alpha 1] | ||
---|---|---|---|---|---|---|---|---|---|---|
Base | Boost | |||||||||
Epyc | 9755 | 128 (256) | 2.7 | 4.1 | 512 MB | 500 W | 16 × CCD 1 × I/OD |
16 × 8 | October 10, 2024 | US $12,984 |
9655 | 96 (192) | 2.6 | 4.5 | 384 MB | 400 W | 12 × CCD 1 × I/OD |
12 × 8 | US $11,852 | ||
9565 | 72 (144) | 3.15 | 4.3 | 12 × 6 | US $10,486 | |||||
9575F | 64 (128) | 3.3 | 5.0 | 256 MB | 8 × CCD 1 × I/OD |
8 × 8 | US $11,791 | |||
9555 | 3.2 | 4.4 | 360 W | US $9,826 | ||||||
9535 | 2.4 | 4.3 | 300 W | US $8,992 | ||||||
9475F | 48 (96) | 3.65 | 4.8 | 400 W | 8 × 6 | US $7,592 | ||||
9455 | 3.15 | 4.4 | 300 W | US $5,412 | ||||||
9365 | 36 (72) | 3.4 | 4.3 | 192 MB | 6 × CCD 1 × I/OD |
6 × 6 | US $4,341 |
A variant of Epyc 9005 using Zen 5c cores was also shown off at Computex. It will feature a maximum of 192 cores and 384 threads, and be manufactured on a 3 nm process.[47]
Common features of EPYC Dense 9000 server processers:
Branding and Model | Cores (threads) |
Clock rate (GHz) | L3 cache (total) |
TDP | Chiplets | Core config[lower-roman 1] |
Release date |
Launch price[lower-alpha 1] | ||
---|---|---|---|---|---|---|---|---|---|---|
Base | Boost | |||||||||
Epyc | 9965 | 192 (384) | 2.25 | 3.7 | 384 MB | 500 W | 12 × CCD 1 × I/OD |
12 × 16 | October 10, 2024 | US $14,813 |
9845 | 160 (360) | 2.1 | 320 MB | 390 W | 10 × CCD 1 × I/OD |
10 × 16 | US $13,564 | |||
9825 | 144 (288) | 2.2 | 384 MB | 12 × CCD 1 × I/OD |
12 × 12 | US $13,006 | ||||
9745 | 128 (256) | 2.4 | 256 MB | 400 W | 8 × CCD 1 × I/OD |
8 × 16 | US $12,141 | |||
9645 | 96 (192) | 2.3 | 320 W | 8 × 12 | US $11,048 |
Zen 5c is a compact variant of the Zen 5 core, primarily targeted at hyperscale cloud compute server customers.[48] It will succeed the Zen 4c core.
Seamless Wikipedia browsing. On steroids.
Every time you click a link to Wikipedia, Wiktionary or Wikiquote in your browser's search results, it will show the modern Wikiwand interface.
Wikiwand extension is a five stars, simple, with minimum permission required to keep your browsing private, safe and transparent.