A Broadwell Retrospective Review in 2020: Is eDRAM Still Worth It?by Dr. Ian Cutress on November 2, 2020 11:00 AM EST
Intel’s first foray into 14nm was with its Broadwell product portfolio. It launched into the mobile market with a variety of products, however the desktop offering in 2015 was extremely limited - only two socketed desktop processors ever made it to retail, and in limited quantities. This is despite users waiting for a strong 14nm update to Haswell, but also because of the way Intel built the chip. Alongside the processor was 128 MB of eDRAM, a sort of additional cache between the CPU and the main memory. It caused quite a stir, and we’re retesting the hardware in 2020 to see if the concept of eDRAM is still worth the effort.
eDRAM: The Savior
In recent years, Intel has pushed hard its infamous ‘Pyramid of Optane’, designed to showcase the tradeoff between small amounts of cache memory close to the CPU being low latency, out to the large offline storage offered for at a significant ping time. When a processor requires data and instructions, it navigates this hierarchy, with the goal to have as much of what is required as close to the CPU (and therefore as fast) as possible.
Traditional modern x86 processors contain three levels of caches, each growing in size and latency, before reaching main memory, and then out to storage. What eDRAM does is add a fourth layer between the last L3 cache on the processor. Whereas the L3 is measured in single digit megabytes, the eDRAM is in the 10s-100s of megabytes, and DRAM measures in gigabytes. Whereas the L3 cache is located on the processor die and low latency, the eDRAM is slightly higher latency, and the main memory is on modules outside the processor socket at the highest latency. Intel enabled an ‘eDRAM’ layer as a separate piece of silicon with the processor package, up to 128 MiB, offering latency and bandwidth between the L3 and main memory.
This piece of silicon was built on Intel’s 22nm IO manufacturing process, rather than 22nm SoC or 14nm, due to Intel’s ability to drive higher 22nm frequencies at the time.
By keeping the eDRAM as a separate piece of silicon, it allowed Intel to adjust stock levels based on demand – if the product failed, there would still be plenty of smaller CPU die for packaging. Even today, processors made with extra eDRAM use the same die as seen back in 2013-2015, showing the longevity of the product. The first eDRAM products were mobile under the 22nm Haswell microarchitecture, but Broadwell saw it come to desktop.
On the Broadwell processors, this resulted in a memory access layer with the following performance:
|Broadwell Cache Structure|
|L1 Cache||32 KiB / core||Private||4-cycle||880 GiB/s|
|L2 Cache||256 KiB / core||Private||12-cycle||350 GiB/s|
|L3 Cache||6 MiB||Shared||26-50 cycle||175 GiB/s|
|eDRAM||128 MiB||Shared||< 150 cycle||50 GiB/s|
|DDR3-1600||Up to 16 GiB||Shared||200+ cycle||25.6 GiB/s|
The simplistic view of this eDRAM was as a ‘level 4’ cache layer – this is ultimately how it was described to us at the time, with the eDRAM layer acting as a victim cache accepting L3 evictions but enabled through a shadow tag system accessed through the L3. Data needed from the eDRAM would have to be moved back into L3 before going anywhere else, including the graphics or the other IO or main memory. In order to do this, these shadow tags required approximately 0.5 MiB/core of the L3 cache, reducing the L3 usefulness in exchange for lower latency extending out to 128 MiB. This is why Broadwell only had 1.5 MiB/core of L3 cache, rather than the full 2.0 MiB/core that the die shot suggested it should have.
Haswell/Broadwell eDRAM Layout
The eDRAM could be dynamically split on the fly for CPU or GPU requests, allowing it to be used in CPU-only mode when the integrated graphics are not in use, or full for the GPU when texture caching is required. The interface was described to us at the time as a narrow double-pumped serial interface capable of 50 GiB/s bi-directional bandwidth (100 GiB/s aggregate), running at a peak 1.6 GHz.
In this configuration, in combination with the graphics drivers, allowed for more granular control of the eDRAM, suggesting that the system could pull from both the eDRAM and the DDR memory simultaneously, potentially giving a peak memory bandwidth of 75.6 GiB/s, at a time when mid-range graphics cards such as the GT650M had a bandwidth around 80 GiB/s.
The second generation of the eDRAM design, as found in Skylake and future processors, moved the eDRAM out of the purview of the L3 cache, and enabled it as a purely transparent buffer between the system agent and the main DRAM memory controller, making it invisible to CPU/GPU accesses or IO accesses. This allows the cache to be accessed by all DRAM requests, enabling full coherency (although the drivers still allow it to be bypassed for textures larger than the eDRAM size), as well as removing the 0.5 MiB/core L3 cache reduction for shadow tags.
Skylake-and-beyond eDRAM Layout
There are arguments to be made about whether the eDRAM as an L4 victim cache or as a transparent buffer to DRAM is the correct direction to go – as a victim cache, Intel stated it allowed a cache hit rate over 95%, however in a number of scenarios in order to get the best performance it required software intervention, and a lot of software was not aware of such a configuration. As a buffer, it enabled seamless integration that all software can take advantage of, but it is not necessarily as optimizable as an L4 victim cache.
‘Go Big or Go Home’
For Broadwell’s eDRAM products, Intel enabled a 128 MiB implementation, quadruple that found on Xbox One silicon at the time. At the time, Intel said that a 32 MiB eDRAM L4 victim cache enabled substantial hit rates, but the company wanted the design to be futureproof as well as a long-term option in Intel’s product stack, so it was doubled, and doubled again just to be sure. The term was ‘go big or go home’, and in our initial review of the first Broadwell eDRAM products, Anand noted that it was very rare to see Intel be so ‘liberal’ with die area.
The eDRAM silicon was built on the 22nm SoC process, as mentioned, one node behind Intel’s leading edge CPU designs. The 128 MiB design came in at a die size of ~77 mm2, contributing to over a third of the total die area used in the 14nm Broadwell Iris Pro quad-core processor package (182mm2 + 77mm2 = 259 mm2).
In the subsequent next generation Skylake generation, eDRAM models with 64 MiB were also offered.
Under certain constraints, the system could save power by disabling the main memory controller entirely if all the data required over a period of time is available in the eDRAM. As part of the initial Broadwell launch, Intel described the extra power consumption of the eDRAM as under 1 watt at idle, moving up to a peak of 5 watts when operating at full bandwidth. Ultimately this means that at a chip level, less power is available to the cores should it be needed, but the trade-off will be better performance in memory limited scenarios. The power is meant to be tracked by the on-die PCU, or Power Control Unit, that can shift power budget between the CPU, GPU, eDRAM, as needed by performance counters or thermals.
As part of this review, we are able to give at least some insights into this number. In our testing, we saw idle package power numbers for the following processors:
- Core i7-4790S (22nm Haswell 4 core 6 MiB L3): 6.01 W
- Core i7-5775C (14nm Broadwell 4 core 6 MiB L3 + 128 MiB eDRAM) 9.71 W
- Core i7-6700K (14nm Skylake 4 core 8 MiB L3): 6.46 W
These numbers would suggest that the effect of the eDRAM, at idle, is more akin to 3.3-3.7 watts, not the sub 1-watt that Intel suggested. Perhaps that sub 1-watt value was more for mobile processors? When running at a steady-state full load, the processors reported power values of their TDP, which doesn’t enable any insight.
Broadwell’s eDRAM Flop?
Intel had somewhat backed itself into a corner with its Broadwell launch. Due to the delays of Intel’s 14nm process at the time, the company had decided to follow its popular Haswell-based 22nm Core i7-4770K high-end processor with the launch of a higher binned ‘Devil’s Canyon’ processor, the Core i7-4790K. This processor offered +500 MHz, which at the time was a substantial jump in performance, despite the processors being launched 12 months apart.
Devil’s Canyon Review: Intel Core i7-4790K and i5-4690K
Because Broadwell ‘wasn’t ready’, Devil’s Canyon was designed to be a stop-gap measure to appease Intel’s ever-hungry consumers and high-end enthusiasts. From the consumer point of view, Devil’s Canyon was at least a plus, but it gave Intel a significant headache.
By bumping the clock speed of its leading consumer processor by a significant margin, Intel now had a hill to climb – the goal of a new product generation is that it should be better than what came before. By boosting its previous best to be even better, it meant the next generation had to do even more. This is difficult to do when the upcoming process node isn’t working quite right. This meant that in the land of the desktop processor, Intel’s reluctance to launch Broadwell with eDRAM was painful to see, and the company had to change strategy.
Intel almost made Broadwell for desktops a silent launch, with very little fanfare. After the announcement, there was almost zero stock on shelves. At the time, Intel did not sample the processors for review – we were able to obtain units from other sources a few days in advance for our launch day coverage.
The Intel Broadwell Desktop Review: Core i7-5775C and Core i5-5675C Tested (Part 1)
The Intel Broadwell Review Part 2: Overclocking, IPC and Generational Analysis
By launching Broadwell Core i7 as a 65 W processor rather than an 84-88 W processor, it meant that the lower frequency Broadwell wasn’t necessarily a direct comparison to Devil’s Canyon. It came out of the gate with a frequency deficit, however the presence of the eDRAM would enable some very careful wins in memory limited scenarios, and perhaps most importantly, gaming.
Ultimately the stunted launch of desktop Broadwell in June 2nd 2015 was very quickly followed by launch of Skylake on August 5th 2015, and the top Core i7 processor was once again an 88+ watt unit and a true like-for-like competitor to Devil’s Canyon. Skylake also enabled DDR4 in the market, which was a significant upgrade on the memory front.
Unfortunately Intel had another conundrum – the older Broadwell processors, due to the eDRAM, actually offered slightly better gaming performance than Skylake! It was title, resolution, and quality dependent, and some might argue there was only a few percentage points in it, but for those that wanted the best at gaming, Skylake wasn’t necessarily the answer. For pretty much all CPU tasks though, Skylake was the answer.
Broadwell Still Available Today
Ultimately, Intel’s foray into socketed Broadwell processors with eDRAM was a momentary blip in its line of consumer-focused Core products. At the time, the processors were hard to find for sale, and were quickly made old by the arrival of Skylake and DDR4. There were six different Broadwell processors that were socketable, two mainstream Core products and four Xeon E3 parts.
|Intel Broadwell eDRAM Socketable CPUs|
|i7-5775C||4C / 8T||3300||3700||48 EUs||1150||65 W|
|i5-5675C *||4C / 4T||3100||3600||48 EUs||1100||65 W|
|* Sometimes listed as Core i7-5675C as some ES had an incorrect CPUID string|
|Enterprise Xeon E3 v4|
|E5-1285 v4||4C / 8T||3500||3800||48 EUs||1150||95 W|
|E5-1285L v4||4C / 8T||3400||3800||48 EUs||1150||65 W|
|E3-1270L v4||4C / 8T||3000||3600||-||-||45 W|
|E3-1265L v4||4C / 8T||2300||3300||48 EUs||1050||35 W|
We were able to also review three of the Xeons at the time.
The Intel Broadwell Xeon E3 v4 Review: 95W, 65W and 35W with eDRAM
Most of these processors are actually very easy to purchase today. The best place to find them are either on Aliexpress, or eBay, for as little as $104.
Broadwell in 2020
The main highlight of these processors was the high-speed eDRAM, coming up to 50 GiB/s bidirectional, at a time when the DDR3-1600 memory solution in dual channel could only offer 25.6 GiB/s. At some point in the future, it would be expected for the speed of normal DRAM to surpass this bandwidth offered, even if it can’t exactly match that latency.
We actually reached that mark very recently.
- Intel’s best consumer-grade processor is the Intel Core i9-10900K, offering 10 cores up to a peak 5.3 GHz, but most importantly the memory side has official support for DDR4-2933, which in dual channel mode would enable 46.9 GiB/s.
- Current AMD Zen 2 processors have a peak supported frequency of DDR4-3200, which in dual channel mode would enable 51.2 GiB/s bandwidth.
- Intel’s mobile Tiger Lake processors support LPDDR4X-4266, which when fully populated would provide 68.2 GiB/s bandwidth.
- With the introduction of DDR5 set to come in the next couple of years, we are expecting to see DDR5-4800 as a possible entry point. This would enable 38.4 GiB/s per 64-bit channel, or 76.8 GiB/s in a standard consumer configuration.
Perhaps it is difficult to wrap your head around the fact that only in 2020 are we matching bandwidth levels that were enabled back in 2015 by the addition of a simple piece of silicon. It might make you question why Broadwell was the only family of Intel’s socketable processors to get this innovation – all future eDRAM products were all for mobile devices that rely on integrated graphics, despite the benefits observed for discrete graphics configurations.
It should be noted that because eDRAM offers a latency benefit in memory accesses from 6 MiB to 128 MiB, then as we approach the situation where a single core has access to 128 MiB of L3 cache, this benefit would also disappear. For consumer processors, we’re not there quite yet – while Intel processors offer up to 20 MiB (or 24 MiB in upcoming Tiger Lake 8-core processors), AMD’s future Zen 3 processors will offer access to 32 MiB of L3 for each core within a CCX. By that metric, we’re still very far behind.
For this review, because we recently tested Intel’s Tiger Lake quad-core processors and graphics, I wanted to probe exactly where Broadwell will finally sit in the hierarchy of CPU performance and graphics performance. We recently announced a new benchmark and gaming suite, and Broadwell is always one of the interesting products to put on a new test suite.
All integrated gaming tests (as well as gaming tests with an RTX 2080 Ti) will be under the respective game pages.
Pages In This Review
- Analysis and Competition
- Test Setup and #CPUOverload Benchmarks
- Power Consumption
- CPU Tests: Office and Science
- CPU Tests: Simulation
- CPU Tests: Rendering
- CPU Tests: Encoding
- CPU Tests: Legacy and Web Tests
- CPU Tests: Synthetics
- CPU Tests: SPEC
- CPU Tests: Microbenchmarks
- Gaming: Chernobylite
- Gaming: Civilization VI
- Gaming: Deus Ex: MD
- Gaming: Final Fantasy XIV
- Gaming: Final Fantasy XV
- Gaming: World of Tanks
- Gaming: Borderlands 3
- Gaming: F1 2019
- Gaming: Far Cry 5
- Gaming: Gears Tactics
- Gaming: GTA 5
- Gaming: Red Dead Redemption 2
- Gaming: Strange Brigade
- Conclusions and Final Words
Post Your CommentPlease log in or sign up to comment.
View All Comments
Billy Tallis - Wednesday, November 4, 2020 - linkIan already said he tests at JEDEC speeds, which includes the latency timings. Using modules that are capable of faster timings does not prevent running them at standard timings.
Quantumz0d - Tuesday, November 3, 2020 - linkDon't even bother Ian with these people.
Nictron - Wednesday, November 4, 2020 - linkI appreciate the review and context over a period of time. Having a baseline comparison is important and it is up to us the reader to determine the optimal environment we would like to invest in. As soon as we do the price starts to skyrocket and comparisons are difficult.
Reviews like this also show that a well thought out ecosystem can deliver great value. Companies are here to make money and I appreciate reviewers that provide baseline compatible testing over time for us to make informed decisions.
Thank you and kind regards,
GeoffreyA - Tuesday, November 3, 2020 - linkThanks, Ian. I thoroughly enjoyed the article and the historical perspective especially. And the technical detail: no other site can come close.
eastcoast_pete - Tuesday, November 3, 2020 - linkIan, could you comment on the current state of the art of EDRAM? How fast can it be, how low can the latency go? Depending on those parameters and difficulty of manufacturing, there might be a number of uses that make sense.
One where it could is to possibly allow Xe graphics to use cheaper and lower power LPDDR-4 or -5 RAM without taking a large performance hit vs. GDDR6. 128 or 256 MB EDRAM cache might just do that, and still keep costs lower. Pure speculation, of course.
DARK_BG - Tuesday, November 3, 2020 - linkHi , what I'm wondering is where the 30% gap between the 5770C and 4790K in Games came from , compared to your original review and all other reviews out there of 5770C. Since I'm with a Z97 platform and 4.0GHz Xeon , moving to 4770k or 4790K doesn't make any sense given their second hand prices but 5770C on this review makes alot of sense.
So is it the OS,the drivers , some BIOS settings or on the older reviews the systems were just GPU limited failing to explore the CPU performance?
jpnex - Friday, January 8, 2021 - linkLol, no, the I7 5775c is just stronger than an i7 4790k, this is a known fact. Other benchmarks show the same thing. Old benchmarks don't show It because back then people didn't know that deactivating the iGPU would give a performance boost.
DARK_BG - Wednesday, July 20, 2022 - linkI forgot back then to reply back , based on this review I've sourced 5775C (for a little less than 100$ this days going for 140-150$) coupled with Asus Z97 Pro and after some tweaking (CPU at 4.1GHz , eDRAM at 2000MHz and some other minor stuff that I already forgot) the difference compared to the Xeon 4.0GHz in games was mind blowing.Later I was able to source and 32GB Corsair Dominator DDR3 2400MHz CL10 just for fun to make it top spec config. :)
It is a very capable machine but this days I'll swap it for Ryzen 5800X3D to get the final train on the fastest Windows 7 capable gaming system.Yeah i know it is OLD OS but everything I need runs flawessly for more than a decade with only reainstall 7 years ago due to an SSD failure. It is my only personal Intel System for the past 22 years since it was the for a first time the best price performance second hand platform for a moment , all the rest were AMD based and I keep them all in working condition.
BTW I was able to run Windows XP 64bit on the Z97 platform , I just need to swap the GTX 1070 for GTX 980/980 Ti to be fully functional everything else runs like a charm under XP i was able to hack the driver to install as an GTX 960 so I have a 2D hardware acceleration under XP on GTX 1070 since nvidia havent changed anything in regard to 2D compared to the previous generation
dew111 - Tuesday, November 3, 2020 - linkRocket lake should have been the comet lake processor with eDRAM. Instead they'll be lucky to beat comet lake at all.
erotomania - Tuesday, November 3, 2020 - linkThanks, Ian. I enjoyed this article from a NUC8i7BEH that has 128MB of coffee-flavored eDRAM. Also, thanks Ganesh for the recent reminder that Bean > Frost.