Would 3D V-Cache Help Intel CPUs? 14th-gen Cores vs. Cache

แชร์
ฝัง
  • เผยแพร่เมื่อ 1 มิ.ย. 2024
  • Download Opera for free using opr.as/Opera-browser-hardware...
    Thanks Opera for sponsoring this video!
    Support us on Patreon: / hardwareunboxed
    Join us on Floatplane: www.floatplane.com/channel/Ha...
    Buy relevant products from Amazon, Newegg and others below:
    Intel Core i3-13100: geni.us/DfDsan
    Intel Core i5-13400: geni.us/lQZFpt
    Intel Core i5-13500: geni.us/goRuq2
    Intel Core i5-13600K: geni.us/lJgCa
    Intel Core i7-13700K: geni.us/Gpuqh3
    Intel Core i9-13900K: geni.us/AieGiL
    AMD Ryzen 7 8700G: geni.us/rTptD
    AMD Ryzen 7 7800X3D: geni.us/FxBWD
    AMD Ryzen 9 7950X3D: geni.us/BV9I9nH
    AMD Ryzen 5 7600: geni.us/qGJVbZ7
    AMD Ryzen 5 7600X: geni.us/Hn7m9
    AMD Ryzen 7 7700: geni.us/6VP8wD
    AMD Ryzen 7 7700X: geni.us/fAfH
    AMD Ryzen 9 7900: geni.us/RRL3Fa
    AMD Ryzen 9 7900X: geni.us/Rjlq
    AMD Ryzen 9 7950X: geni.us/lTmCsO
    AMD Ryzen 7 5800X3D: geni.us/wZd1
    AMD Ryzen 9 5950X: amzn.to/32gHUAN
    AMD Ryzen 9 5900X: amzn.to/3kUa1gq
    AMD Ryzen 7 5800X: amzn.to/32cX9uB
    AMD Ryzen 5 5600X: amzn.to/3evOFn9
    AMD Ryzen 5 5600: geni.us/dLiI
    Video Index
    00:00 - Welcome to Hardware Unboxed
    00:25 - Ad Spot
    01:30 - Cores vs. Cache
    04:19 - Assassin's Creed Mirage
    05:12 - Helldivers 2
    05:36 - Ratchet & Clank: Rift Apart
    06:00 - Marvel’s Spider-Man Remastered
    06:39 - Cyberpunk 2077: Phantom Liberty
    07:06 - Hogwarts Legacy
    07:32 - Horizon Forbidden West
    07:55 - Dragon’s Dogma II
    08:29 - Baldur's Gate 3
    08:49 - The Last of Us Part I
    09:38 - Starfield
    09:58 - 5.7 GHz Results
    11:05 - Final Thoughts
    Would 3D V-Cache Help Intel CPUs? 14th gen Cores vs. Cache
    Disclaimer: Any pricing information shown or mentioned in this video was accurate at the time of video production, and may have since changed
    Disclosure: As an Amazon Associate we earn from qualifying purchases. We may also earn a commission on some sales made through other store links
    FOLLOW US IN THESE PLACES FOR UPDATES
    Twitter - / hardwareunboxed
    Facebook - / hardwareunboxed
    Instagram - / hardwareunboxed
    Outro music by David Vonk/DaJaVo
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 709

  • @alltheotherhandlesaretaken
    @alltheotherhandlesaretaken 14 วันที่ผ่านมา +472

    Where’s the 5800X3D results?

    • @Hardwareunboxed
      @Hardwareunboxed  14 วันที่ผ่านมา +250

      Haha 😄

    • @user-wq9mw2xz3j
      @user-wq9mw2xz3j 14 วันที่ผ่านมา +92

      intel 5800k3d

    • @Collin_J
      @Collin_J 14 วันที่ผ่านมา +70

      Can't believe they didn't test the CPUs in 4K

    • @frieza1016
      @frieza1016 14 วันที่ผ่านมา +47

      M8 didn't test in 8k, what a fucking w@nker!

    • @seafafrage6934
      @seafafrage6934 14 วันที่ผ่านมา +7

      this joke is becoming cringy now.

  • @falcdragon
    @falcdragon 14 วันที่ผ่านมา +50

    In general these are pretty minor differences in cache, 33/36 is ~9% difference. The 24 vs 36 is a more substantial shift at 50% but still minor compared to 300% increase that X3D brings. If we go back to your original 5800 vs 5800X3D video there was an average 15% increase in performance for 300% increase in L3 Cache. If we assumed a fixed ratio 15/300 that 0.05% increase in performance per 1% increase in cache. At that level going from 33 to 36 is 9% more cache and assuming the linear relationship you could expect a 0.45% increase in performance. If we go from 24 to 36 that's 50% increase in cache, so again assuming linear relationship you might expect to see a 2.5% increase in performance.
    For this sort of comparison you should be looking at linear regressions between increased cache (either in MB or %) and observed performance change. Also given the predicted performance change it's likely within the run to run variability of your tests, which means you may struggle to be able to detect it.
    @Hardwareunboxed Not saying you are wrong in your conclusion (clock matters more for intel) but your evidence is insufficient to clearly show it, as it's not possible to test a comparable scenario.

    • @falcdragon
      @falcdragon 14 วันที่ผ่านมา +5

      In fact for a number of the games you are seeing a larger than expected performance increase for the additional cache. Which would suggest cache is important and the effect is somewhat non-linear.

    • @GewelReal
      @GewelReal 13 วันที่ผ่านมา +3

      Don't look at the % difference in size. Look at the total size, as instruction size is measured in MB, not %.
      Going from 3MB to 6MB would net the same increase as going from 33MB to 36MB, as you'd fit extra 3MB worth of instructions, despite one being 100% increase and other being 9%

    • @falcdragon
      @falcdragon 13 วันที่ผ่านมา

      @@GewelReal It doesn't make much difference if you use per MB or percentage. It also doesn't really matter if the L3 is mainly cache or data. In both cases, increasing the amount of cache reduces the number of times you have to go out to the main memory and thus saves cycles of the cores waiting for instructions/data.
      If you calculate it out as a percentage performance improvement per MB of additional cache based on the 5800X3D and 5800 (and note that this doesn't account for the non-X3D's higher clock speeds), then what you get is 15/64 = 0.234%/MB. If you assumed the same benefit applies to Intel CPUs (which is reasonable the intel memory subsystem isn't massively better for latency or bandwidth than the AMD one) then you get the following predicted changes:
      24MB > 33MB = +9MB, 9*0.234 = 2.1% estimated improvement in performance.
      24MB > 36MB = +12MB, 12*0.234 = 2.8% improvement in performance.
      33MB > 36MB = +3MB, 3*0.234 = 0.7% improvement in performance.
      Thus, in the tests the guys ran above, assuming linear scaling (which is an assumption), we'd expect to see 0.7-2.8% better performance, which certainly appears to be the case in a number of those games. In some cases, the difference is higher than that. The real question, though, is can the guy's test detect a 2-3% increase in performance, or is that within the margin of error from run-to-run variance?

    • @jackleville546
      @jackleville546 13 วันที่ผ่านมา

      Fully agree with what you say. The increase in cache may not be enough for really cache-sensitive applications to store the data they directly need. And I would rather have liked to see the focus on games that really seem to benefit form more cache: Assetto Corsa, MS Flight Simulator, Borderlands 3...

    • @GewelReal
      @GewelReal 13 วันที่ผ่านมา +1

      @@falcdragon The best CPUs to test cache scaling are X99 Xeons (to a point).
      in fact, I had 7 of them, and scaling is indeed dependant on the raw amount rather than % increase (architectural differences might make it less pronounced).
      Altough, Haswell/Broadwell Xeons somehoe stop scaling past ~30MB (the 50MB chip showed 0 increase in performance in all games over 30MB/35MB ones).
      But the best example for showing how anything below ~30MB is a massive disadvantage, as going from 1650v4 at 15MB to 2650v4 at 30MB INCREASED performance in all games despite 1650v4 running at 3.8GHz all core while 2650v4 only doing 2.5GHz.
      But then going from 2650v4 to 2673v4 at 50MB (2.6GHz all core) didn't actually increase performance (maybe 5-10% at most in 1-2 titles)

  • @CarputingYT
    @CarputingYT 14 วันที่ผ่านมา +261

    Have a nice Saturday Mr. Hardware Unboxed man

    • @AbbasDalal1000
      @AbbasDalal1000 14 วันที่ผ่านมา +25

      Man unboxed

    • @notthedroidsyourelookingfo4026
      @notthedroidsyourelookingfo4026 14 วันที่ผ่านมา +1

      Tf you on about?

    • @cocosloan3748
      @cocosloan3748 14 วันที่ผ่านมา +4

      Soon will be Sunday in Australia but ok...

    • @yueqi7499
      @yueqi7499 14 วันที่ผ่านมา +2

      its actually saturday night but good one

    • @LlywellynOBrien
      @LlywellynOBrien 14 วันที่ผ่านมา +13

      ​@@AbbasDalal1000that is their OF channel. They try not to mention it too much.

  • @EmanuelHoogeveen
    @EmanuelHoogeveen 14 วันที่ผ่านมา +28

    For *some* of these games, it does seem like cache makes a respectable difference. 36MB is about 9% more than 33MB, so when we're seeing a 4-5% performance increase that's still fairly decent scaling. Unfortunately there's only 2 data points with 8 cores so we can't extrapolate to say, 96MB - I'd expect it to plateau to some extent, but maybe nearly tripling it would still net a decent benefit. But I guess they'd have to lower the clock speed significantly to avoid frying the 3D cache so it probably wouldn't help overall for their current architecture, as you said.
    Edit: As others have said, there might also be workloads where you need to hit some magic number to see a large benefit - like 64MB or a little more on the side for smaller things. It's a pity that there's no real way to test for that except by looking at AMD chips (which have entirely different constraints).

    • @WilliamJasonSherwood
      @WilliamJasonSherwood 12 วันที่ผ่านมา

      I'd argue that L3 would likely behave similarly to RAM or V-RAM where it kind of matters at thresholds, 25-36MB all behaves the same because they're all going out to main memory anyway. But if you got to say 50 MB or 80 MB, suddenly you've got enough L3 to be able to skip 80% of those main memory calls.
      You would need something stupid like the w9-3495X to be able to test it, and I don't even know if they'd support fafing about with cores & cache configurations. IMO that would be the way to test it, especially since with w9's potentially have even more L3 than the AMD parts with a maximum of 105 MB.
      BUT another consideration is that having the cache split up on 50 odd different cores while AMD has it all on say 8 cores, means the intercore transfer is pretty significant and perhaps THAT's what you are gaining on, perhaps EVERY core need access to say 25 MB of L3 to see these kinds of performance improvement, and say a shared L3 arrangement COULD perform the same.

    • @conanpiggott9465
      @conanpiggott9465 11 วันที่ผ่านมา

      I'm with you on this, there was clearly a difference in the 8 core samples and a mild difference in the 6 care samples.

  • @MarkHyde
    @MarkHyde 14 วันที่ผ่านมา +136

    Can't get enough of this type of specialised CPU content. Thanks!!!

    • @kevinerbs2778
      @kevinerbs2778 13 วันที่ผ่านมา +2

      Hardware unboxed or someone should try testing this same theory with mutli-GPU setups vs single card setup. There's only few mGPU games to test this on anyways that actually use mGPU. I'm pretty sure clock speed is still better for mGPu than cache for it. As I'm someone who has RTX 2080 ti's in S.L.I & have noticed that my 5600x seemed to give better mulit-gpu performance from higher clocks (4.85ghz) over then 5800x 3D (4.45Ghz) I replaced it with. While the 5800x 3D seems to be happier pushing a single card better.
      Here is the list of mGPU games.
      1. Ashes of the singularity = Natively supported for All cards even different vendors
      2. Chasm the rift = Vulkan
      3. Deus EX mankind Divided (DX12 AFR must have single card set in Nvidia's control panel.)
      4. Rise of the Tomb Raider
      5. Shadow of the Tomb Raider
      6. Quake 2 RTX = Vulkan
      7. Far Cry 5
      8. Red dead Redemption 2 Vulkan
      9. Gears of war 4
      they'd probably only be able test this on Intel's Z690 boards with a I think a 12900K as that's the last boards on the intel side that support S.L.I/ while only about 5 x570 boards supported one being the MSI MEG ACE x570 board that I'm using right now. On side note that I think people don't know but All of AMD's RDNA line up support mGPU. All you do is with two of the same tier cards, find mGPU in the control panel & enable it restart & you're done. It does however disable Rebar in bios when enabled.

  • @imglidinhere
    @imglidinhere 14 วันที่ผ่านมา +81

    I think it's safe to say that Zen 3 was massively bandwidth-starved, with Zen 4 showing much the same result, and the true performance of that architecture can be seen once that problem was alleviated with the added V-cache. It would make sense given how heavily memory speed affected overall performance. I remember seeing as much as a 15% swing from using the slowest to the fastest memory, with Zen 3 at least.
    Intel isn't hobbled in the same way, clearly. The difference between using 4800mhz and 7200mhz memory is measurable, however minor, compared to AMD.

    • @malcaniscsm5184
      @malcaniscsm5184 14 วันที่ผ่านมา +15

      Well I guess that's why AMD made a deal out of unifying the L3 for Zen3, so each core sees 32MB rather than 8MB for the Zen2.

    • @naamadossantossilva4736
      @naamadossantossilva4736 14 วันที่ผ่านมา +1

      You hit the nail in the head.

    • @Angel7black
      @Angel7black 14 วันที่ผ่านมา +7

      Then the next thing i have to say is why do AMD know this about their product, but for a second time in a row now decided to first launch Zen 4 at marked up prices as a mediocre product that at best went nearly even with 12th gen Intel in gaming and way worse in productivity due to core count differences? Dont miss understand the point here im not being a fanboy and only comparing AMD vs Intel in gaming before Vcache to make AMD look Intel look better, im going after AND for willfully knowing they should be putting Vcache on all their chips but launching gimped for a year then putting out the chips the way they should be and jumping prices up again and charging damn near $500 for a 8 core 16 thread. AMD needs to stop milking the consumer base, it really doesnt even help them and Zen 5 needs tp be 3D vcache on ALL cars not this parked cores shit, that of learn how to do hybrid cores or make non 3D cores focus on background tasks while 3D cores focus on gaming similar to how big little cores operate. They still havent fixed that issue fully with the 7900X3D and 7950X3D parking or using the wrong cores fully

    • @nepnep6894
      @nepnep6894 14 วันที่ผ่านมา

      ​@@naamadossantossilva4736Intel does still have great memory scaling, but speed does almost nothing on it's own. You need to tune subtimings to get any meaningful gains. AMD does also but a lot of the scaling is from fclk.

    • @DeadAimKillah
      @DeadAimKillah 14 วันที่ผ่านมา +7

      @@Angel7black Ummm you do know 3dvcache wasn't made for gamers right? 3dvcache is more efficient in there work making them suitable for server processors (look up Zen c cores)
      they didn't gimp there processors or hold back because they were better than intel. they aren't on all chips because they are expensive and they save most of them to go on their server c cores.
      this is a case of trickle down technology where we get a cool addition to our cpus because its something AMD thought we'd like.
      it was the same thing with HBM. but nvidia rendered that useless for gamers when they purchased the best memory chips to overclock them for GDDR5X

  • @Ralipsi
    @Ralipsi 14 วันที่ผ่านมา +141

    Now we do understand clearly why Intel are so focused on clock frequency. Thanks for the impeccable bench test and analysis! I absolutely love HardWare Unboxed. You guys know you stuff and do a great job. ...making it so easier for us to understand our hardware needs and what are the best gears to get.

    • @RobBCactive
      @RobBCactive 14 วันที่ผ่านมา +7

      A treacherous path to follow though.

    • @alejandrocalori6298
      @alejandrocalori6298 14 วันที่ผ่านมา +7

      Wins on single threaded applications and lets people brag about intel just showing single core benchmarks, like some dumba** on twitter that I saw just now.

    • @maynardburger
      @maynardburger 14 วันที่ผ่านมา +4

      @@RobBCactive I mean, it's not like they're designing specifically for clocks and nothing else. They understand the need to push IPC and whatnot as well. Alder Lake was a huge lift in IPC and I have no doubt Arrow Lake will see another decent lift here as well, especially if they have to lower clocks a bit for the new process a bit. Intel's first generation for a new process tend to not clock super well.

    • @RobBCactive
      @RobBCactive 14 วันที่ผ่านมา +3

      @@maynardburger but they've had to push both P & E cores high up the power wall to compete in benchmarks pushing frequency and using E for die area efficiency has caused sacrifices on AVX and we know to stability on the K series which try to compete with Zen4.
      The problem with IPC is games in general simply don't hit anything like the CPU theoretical values that numerical code can reach.
      I would not be surprised if the close difference between i7 & i9 cache is down to small size difference so they haven't found games with highly boosted hit % rate.

    • @maynardburger
      @maynardburger 14 วันที่ผ่านมา

      @@RobBCactive Intel are still hamstrung a bit by being a whole node generation behind AMD at the moment(talking desktop here). If they were on the same or similar node as Zen 4, they could have relaxed things greatly in terms of power draw without sacrificing clockspeed. And whether or not any given game is fully utilizing architectural IPC improvements is kind of pointless to argue, because IPC still does play a significant part of improved performance in games. It was the main reason Alder Lake was as big a leap as it was in gaming performance.

  • @GS-kh5se
    @GS-kh5se 14 วันที่ผ่านมา +135

    I understood how important cache was when my Pentium Pro with its fast L2 cache was still usable years after regular Pentiums had to be retired.

    • @exscape
      @exscape 14 วันที่ผ่านมา +21

      I learned about it when I disabled the cache (only the L2 cache was disabled IIRC) on my Pentium 100. It's truly crazy how slow it became, tasks that used to take seconds could take minutes.

    • @Sebastianino
      @Sebastianino 14 วันที่ผ่านมา +4

      cache is a lot faster then any ram so yes better have more then less

    • @dakoderii4221
      @dakoderii4221 14 วันที่ผ่านมา +7

      I had a bunch of different Intel chips back in 2014 with the same architecture but different cache amounts by disabling cores on high end i7s to match a lesser i7 or i5 or i3 but with more l3 cache available to them. I saw a marked improvement by just adding more l3 cache for each core. I was saying why don't just add a little bit more, at least on some models? I wasn't the only one. I've seen a few people talking about in videos and on forums years before it was announced by AMD.

    • @SpoonHurler
      @SpoonHurler 14 วันที่ผ่านมา +4

      ​@@Sebastianinonot just faster but orders of magnitude less latency... even if its a similar speed if you had something crazy like 8 channel ultra fast ram, the cache still is better for anything that fits in it... which is another important consideration for vcache, it allows more within it

    • @pauldunecat
      @pauldunecat 14 วันที่ผ่านมา +1

      The Pentium Pro was such a stud.

  • @masterchi17
    @masterchi17 14 วันที่ผ่านมา +48

    I think the conclusion drawn is not entirely supported by the experiments. The cache increase is quite small from 24 to 36MB, while for AMD it is 32 to 96MB. So 64MB extra for roughly 15%, and we did saw for many 6-core tests a roughly 10% increase for 12MB increase.
    Also the 4-core results indifference to cache size can be explained that with less cores the game will launch less worker threads. And each thread has its own memory structures, so reducing the number of threads reduces also the pressure on the caches in terms of capacity.
    But apart from that, another excellent video answering the 6 to 8-core debate 😅.

    • @glenndoiron9317
      @glenndoiron9317 14 วันที่ผ่านมา +5

      Yeah, the 33MB -> 36MB cache increase isn't very big ( i9 means almost nothing for most games for 14th gen processors, and you probably should be snagging the i7 for a gaming system (assuming you don't want to deal with AM5 motherboards and/or 7800X3D stuttering issues while web browsing - if you're ok with those, then the 7800X3D is a better pick for gaming.)

    • @aapzehrsteurer9000
      @aapzehrsteurer9000 14 วันที่ผ่านมา +8

      I was thinking the exact same thing. The difference between AMD's 3D cache and this is so huge that it may cross a threshold at which bigger data structures can be loaded into the L3 cache.

    • @chriswright8074
      @chriswright8074 14 วันที่ผ่านมา +4

      Intel architecture isn't memory sensitive zen always has been and eypc has great improvements with 3d cache

    • @M3N04
      @M3N04 14 วันที่ผ่านมา

      ​@@chriswright8074💯

    • @A.Froster
      @A.Froster 14 วันที่ผ่านมา +16

      ​@@glenndoiron9317 First time i have heard of the 7800x3D stuttering issues

  • @heeerrresjonny
    @heeerrresjonny 13 วันที่ผ่านมา +7

    I'm confused by some of the conclusions here for certain games. For games where there does seem to be a difference based on cache, like in Dragon's Dogma II there does seem to be a notable uplift at times. DDII had a 4% uplift in avg FPS with only 3MB more cache. The X3D parts increase cache by much larger amounts, e.g. the 7800X3D is 96MB vs the 7800X at 40MB, a 140% increase. So, a hypothetical 14900K"X3D" might have something like 86MB of cache. At 6 cores, it is even more dramatic with a 14% uplift from 12MB more of cache and again a 4% uplift from 3MB of cache. That's about 1.2fps or 1.3fps PER MB OF CACHE lol. If you scale that up the way it'd be for an X3D part, a 14900KX3D might hit 130fps or higher (in theory). While clock speed seems to be a primary limiter, there is obviously always a limit where you can't increase clock speed further...and then more cache might indeed lead to even more performance. I feel like the conclusions for some games here aren't accounting for the variable amount of cache being added nor accounting for the much larger cache variation we see between X3D and non-X3D parts.

    • @mgk878
      @mgk878 13 วันที่ผ่านมา

      I agree, the FPS boost is actually decent in certain games. I am very curious where the scale ceiling is. It's plausible that a 48+MB 6-core chip would outperform the 14900K at least in DD2, while using slower memory, and running at lower power too. So why hasn't Intel made a chip like that? Maybe it raises cost too much, or they care too much about synthetic benchmarks that don't need much cache...
      Edit: looking at the higher clock results, I guess it wouldn't outperform it, but maybe just match it. Hm.

  • @techpupil-au
    @techpupil-au 14 วันที่ผ่านมา +60

    Really great experiment mate, good stuff

  • @Jason-ol1ty
    @Jason-ol1ty 14 วันที่ผ่านมา +7

    Some years way back, I had an old CPU where the cache on the chip failed, but the rest of the CPU worked. The chip was soooo slow from that, and Windows was not really usable so I went back to DOS until the replacement came. I had to get by on playing Master of Magic, which was the only program I had that worked fast enough lol!

  • @MrAtthedrivein925
    @MrAtthedrivein925 14 วันที่ผ่านมา +13

    Love the testing Steve! Seriously what a great idea to test. Very informative and we appreciate all you do!!

  • @jurgengalke8127
    @jurgengalke8127 14 วันที่ผ่านมา +54

    Why weren't games that are known benefit from Vcache tested i.e. COD, Assetta Corsa, Forza etc?

    • @PowellCat745
      @PowellCat745 14 วันที่ผ่านมา +8

      Good point.

    • @pedroferrr1412
      @pedroferrr1412 14 วันที่ผ่านมา +8

      Even more, like FS, RF2 and in VR! I really do not understand why hardware sites, never benchmark CPU´s in VR... it´s the future, even if most do not think so, just try once, and you will never return to an pancake. I have just 1 eye, lost the other 1, two years ago, duo to cancer. But i still only play SIM games in VR, started many years ago, with 3 monitors, but never returned. I have 3 monitors connected to the PC, but just for work.

    • @LupusAries
      @LupusAries 14 วันที่ผ่านมา +3

      Or DCS, Arma3 and if you want to eliminate the GPU bottlenecking, Falcon BMS 4.37.

    • @HowlingMoai
      @HowlingMoai 14 วันที่ผ่านมา +7

      ​@@LupusAriesthose games aren't used for benchmarking because they are much more variable and not as good for performance comparison because of how they are designed. A benchmark in Arma 3 can average 75 fps on one run but then get 65 fps on the next run (even though they should be identical).

    • @LmgWarThunder
      @LmgWarThunder 14 วันที่ผ่านมา +2

      You have a very good point. I think the biggest reason is that if you look at blanket benchmarks of the 7800 x3d versus the 7800x, you would find that it performs better in almost all games, not just some and these games would be included in that.

  • @HanSolo__
    @HanSolo__ 14 วันที่ผ่านมา +55

    I'm reading comments - people seem to not understand the idea of this test and its results.
    Intel processors do not work the way AMD CPUs do. Those are different types of constructions.

    • @2drealms196
      @2drealms196 14 วันที่ผ่านมา +4

      AMD's 3D cache makes use of a chip stacking/packaging technology (Chip-on-Wafer) that TSMC invented. So Intel would to make similar advances in its chip stacking capabilties before they'd be able to implement a similar stacked cache.

    • @OGPatriot03
      @OGPatriot03 14 วันที่ผ่านมา +3

      The difference in cache is so small that I don't think you can draw those sorts of conclusions, for example 3D cache for AMD isn't a tiny increase in chache it's rather massive which makes a lot more room for potential gains.
      Some games don't use much of the cache and in those cases neither would see much or any benefit, however plenty of games do make use of it and I would imagine where those games gained, AMD would have similar and proportionately larger gains due to the increased cache.

    • @K31TH3R
      @K31TH3R 13 วันที่ผ่านมา +4

      Would probably be helpful to explain why it is the case instead of saying "they just work differently."
      AMD uses a "chiplet" based approach which relies on something called the "Infinity Fabric" (IF) for the communication between the memory controller (IMC), cores (CCDs), and SoC (which primarily handles PCI-E communication). Games are often very sensitive to latency, and the frequency at which the IF runs has a large impact on the overall performance of AMD CPUs since it directly impacts overall latency. But unlike the name would suggest, the IF can't actually run at infinite speeds, and because of this, it will always introduce additional latency between the CCD(s), IMC, and DRAM. The 3D cache significantly reduces how often the cores need to cross the IF to access DRAM, which means it will significantly reduce overall latency as long as the working set mostly fits within the 3D cache.
      Intel CPUs do not use chiplets (for now), they use a monolithic design with an integrated IMC, so accessing the DRAM has an overall much lower latency penalty than AMD's design. That being said, Intel CPUs will also benefit from additional cache, but since the DRAM subsystem already has much lower latency than AMD's design, AMD chips will tend to benefit more.

  • @Nunkuruji
    @Nunkuruji 14 วันที่ผ่านมา +4

    If Intel had a part that dropped e-cores, un-hampering the ring bus, and replaced the die space with cache, they could have probably consistently edged out the X3D parts in gaming. Oh well, things will get real interesting once re-architected parts start being produced with backside power delivery.

    • @OGPatriot03
      @OGPatriot03 14 วันที่ผ่านมา

      Yea but now everyone is running Intel stock baseline, which brings all the numbers down a good bit.

  • @vulcan4d
    @vulcan4d 14 วันที่ผ่านมา +5

    They did, the i7-5770C back in the lga1150 days. It was rare and expensive. I upgraded from a i7-4790k and whoa was it better. Games that stuttered, didn't stutter anymore...it was revolutionary. Unfortunately, they did not continue. Probably because Intel likes to stall back in the days.

    • @twilightravens9798
      @twilightravens9798 13 วันที่ผ่านมา

      It was a bit different though, the L4 on the 5775C wasn't nearly as fast as the L3, and was on a separate die (was slightly faster than my DDR3 2400 but not by much). I used my 5775C for a while until the 4 cores became the bottleneck. Was a cool idea but wasn't really feasible for a more mass production run because 14nm yields at the time were terrible and Broadwell wasn't exactly known for hitting high clocks, with or without the L4.

  • @clifflenoir4323
    @clifflenoir4323 8 ชั่วโมงที่ผ่านมา +1

    Interesting that 8 cores can occasionally give better performance than 6 cores, even with identical clock speeds and cache (Ratchet & Clank data @ 10:56)

  • @sublime2craig
    @sublime2craig 11 วันที่ผ่านมา

    I have asked myself this same question many of times inside my own head, so happy you guys are here to give me an answer!!! Love this kind of testing, please keep em coming!

  • @makobe584
    @makobe584 14 วันที่ผ่านมา +3

    Would've liked to have seen some cache-limited games like Factorio, Assetto Corsa, and flight simulator.

  • @BuzzKiller23
    @BuzzKiller23 12 วันที่ผ่านมา

    Very interesting. Thanks for doing these tests!

  • @Roll_the_Bones
    @Roll_the_Bones 14 วันที่ผ่านมา +1

    From what I understand from the original TechPowerUp tests, the 14600K can be OC'd up to 14700 level on a decent air-cooler, but the 14700 brews up if it's even mildly OC'd? And the 14700/14900's need to have substantial expensive cooling to gain any slight advantage, at least for gaming? So, I'm thinking for the 14th series, the 14600K might be the value/performance sweet-spot for gamers?

  • @nurofenusa1
    @nurofenusa1 12 วันที่ผ่านมา

    @ Steve - I am always impressed by your deep dives, research and project management that must be undertaken in your efforts to produce content that really digs deep into the hardware and related information that no other channel is prepared to undertake. The information you provide graphically is easily digestible by the layman, and gives me insights on bettering the communication to others I may be working with on projects. You and Tim put in the hours where its needed and have a great synergy together on screen. Oh and I'm a Pom so take that priase for what it is :)

  • @Skidzo19
    @Skidzo19 9 วันที่ผ่านมา

    5:37 "Ratchet and Clank like Assassin's Creed mirage" Wow they're so in sync.

  • @PowellCat745
    @PowellCat745 14 วันที่ผ่านมา +64

    3D V Cache helps Zen 4 enormously because Zen 4 is primarily memory starved. The default 2000MHz Infinity Fabric for Zen 4 offers a per-CCD bandwidth equivalent to DDR5-4000. Intel running DDR5-7200 is equivalent to a Zen 4 CPU with a 3600MHz Infinity Fabric.
    This is my theory why L3 cache helps Zen 4 enormously while it barely helps on Raptor Lake.
    Edit:
    People don’t realize how much of a memory bottleneck there is on Zen 4.
    Steve tested the same thing for the 10th gen Intel CPUs, but the memory used was DDR4-3200, equivalent to a 1600MHz Infinity Fabric speed on Zen 4, which is merely 20% slower than an actual Zen 4 CPU. Steve’s Raptor Lake test bench uses DDR5-7200, equivalent to a 3600MHz Infinity Fabric speed. This is 225% faster than the 10th gen and 180% faster than Zen 4. That’s why memory bandwidth is rarely an issue on RPL and more L3 cache didn’t help much.

    • @Hardwareunboxed
      @Hardwareunboxed  14 วันที่ผ่านมา +32

      In the past it has also helped Intel CPUs a lot.

    • @PowellCat745
      @PowellCat745 14 วันที่ผ่านมา +11

      @@HardwareunboxedYes, but the 10th gen only supports DDR4, and your 10900K test setup used DDR4-3200 CL14, which is equivalent to a 1600MHz infinity fabric, much closer to Zen 4’s memory bandwidth than that of Raptor Lake.

    • @PowellCat745
      @PowellCat745 14 วันที่ผ่านมา +24

      @@Hardwareunboxed maybe as a follow up, you can retest with ddr5-4000 😁

    • @Aggrofool
      @Aggrofool 14 วันที่ผ่านมา +5

      Referring to Intel Broadwell-S, L4 looked good because Skylake was stuck with early-gen DDR4.

    • @aberkae
      @aberkae 14 วันที่ผ่านมา +1

      This will happen again with AM5/Zen5 3d where Intel is heading towards 10 ghz ram kits.

  • @conza1989
    @conza1989 14 วันที่ผ่านมา +1

    Always good to see these videos, just another factor to consider in purchasing, do I go for the better part, not because of more cores but more cache. Ryzen 8000/9000 (probably 9 isn't it), might be interesting to see if there's differences... If they sell enough parts for there to be cache differences, aside 3D Vcache of course.

  • @depth386
    @depth386 14 วันที่ผ่านมา

    This is fascinating, in the context of how different it is from that 10th gen. I remember watching that. Thank you so much for making this video Steve.

  • @limeisgaming
    @limeisgaming 14 วันที่ผ่านมา +5

    what would be more interesting is a comparison between the i5 14500 @5Ghz which uses the old 1.25MiB L2 vs the i5 14600@5Ghz which uses 2 MiB L2 per P-Core just like the Server Cpus of 4th and 5th Gen Intel Xeon Scalable.
    i5 14500 6 P-Cores 6x1.25MiB L2 + 24MiB L3
    vs
    i5 14600 6 P-Cores 6x2MiB L2 + 24MiB L3
    and compare that to the relative results between a @5Ghz capped ryzen 7 7700 and a Ryzen 7 7800X3D (its also capped at 5Ghz)
    PS.: there does intel with "3DV-Cache" exist its those high Cache Emerald Rapids SKU's, with Emerald Rapids also using Golden Cove for its Cores it should offer some comparison.
    Intel Xeon Gold 6548N 32 Cores 32x2MiB L2 60MiB L3 (1.875MiB L3/Core)
    Intel Xeon Gold 6530 32 Cores 32x2MiB L2 + 160MiB L3 (5MiB L3/Core)
    Server(Xeon 4th and 5th) Desktop/Mobile(Alderlake, Raptorlake, Raptorlake Refresh) are all based (more or less) on "Golden Cove"
    Server has 1.875MiB L3 as its standard with 5MiB L3/Core for high Cache Variants all of which have 2MiB L2/Core
    Desktop had until "Raptorlake Refresh" (1.25MiB L2 + 3MiB L3)/Core, Raptorlake-Refresh has (2MiB L2 + 3MiB L3)/Core

  • @michaelmcconnell7302
    @michaelmcconnell7302 14 วันที่ผ่านมา +4

    what an extremely useful and informative video. thanks, steve.

  • @themalcore_
    @themalcore_ 14 วันที่ผ่านมา +1

    Something that would be worth mentioning is the higher the core clock speed (relative the memory access latency) the more exaggerated the penalty for instruction window full stalls will be, and therefore the more valuable large caches are.

  • @tremelai
    @tremelai 13 วันที่ผ่านมา +1

    Unlike the 10th generation, (Comet Lake) a refresh++ of Skylake which had a uArch that was initially released and optimized for throughput server workloads, the 12th generation Alder Lake and by extension, 14th Raptor Lake both using the Golden Cove and Raptor Cove P-Core respectively was initially released for desktop workloads and released later for server workloads in Intel's sapphire rapids. Intel has been down scaling CPUs uArchs to desktop/laptop workloads since the 686 Pentium Pro with the 12th generation being the first to go the other direction since the 586 Pentium.
    AMD followed Intel in the practice of designing and releasing uArchs for workstation/server workloads first with the K8 Athlon64 FX/Opteron release, followed on by the downscaling to socket 939 desktop variants. This trend continues today with the Zen uArch.
    Server/Workstation uArchs traditionally have more bandwidth, both IO and memory, than their desktop/laptop counterparts, as such, the integer and FPU/streaming execution units are optimized for the larger number of memory channels afforded to the server platform. An Epyc Zen3 sports 8-channels of DDR4-3200 or 204.8 GB/s or theoretical bandwidth. Compare that to the 57.6 GB/s of theoretical bandwidth that is afforded dual channel DDR4-3600. X3d was developed for Zen3 to afford more bandwidth to high core count Epyc CPUs and is not generally available on Epyc CPUs below 16-cores per socket.
    Intel going the other direction, of scaling up a desktop/laptop optimized P-core to server workloads, took the dual channel DDR5 optimized scheduler/execution unit core and scaled up the memory/PCI bandwidth to meet higher than 8 p-cores per socket. AMD, and Intel with 11 back to Pentium pro, would trim core count to find the upper limits of desktop parts while ramping frequency. With Zen3 and Zen4, going from 8-channel / 12-channel memory channels respectfully, allowed for a limited amount of trimming core count before you impact the ability to feed individual core's integer execution units with actionable work to do.
    Adding X3d to the desktop parts helped close that architectural bottleneck. Sizing core count can close the memory bottleneck only so much when your branch predictor is tuned for 4x more memory bandwidth. Same was seen on the 10th generation Intel where the conditions were much the same. (See Intel's Broadwell i7-5775C with it's 128MB of L4 cache, dubbed eDRAM.)

  • @-Noble
    @-Noble 14 วันที่ผ่านมา +1

    idk much about the architecture, but does disabling cores disable some of the cache? or is the cache just there for any of the cores to use as programmed?

    • @tim3172
      @tim3172 14 วันที่ผ่านมา +1

      Intel's "Smartcache" (eyeroll) is unified cache that isn't disabled when cores are disabled.

    • @-Noble
      @-Noble 14 วันที่ผ่านมา

      @@tim3172 ty for the info

  • @Redmage913
    @Redmage913 13 วันที่ผ่านมา +1

    One problem I see right now is that they’re on such a (relatively) large node that to modify current designs to include more cache could blow out the size of the processor to an unreasonable size and therefore price of each CPU to produce, and then have to sell.
    We can hope Intel’s smaller nodes work out, and perhaps the amount of L3 cache can be solved that way.
    I’d love Intel to be more competitive in the energy efficiency department for their high-end parts. I still buy their products - my dad recently got an Intel i7-1260P (I think) laptop that’s amazing. But its temps are all over the place on a whim and needs a lot of fan to cool in sustained work. Good chips, bad efficiency for work :/

  • @bretthake7713
    @bretthake7713 14 วันที่ผ่านมา +7

    "some dorks will complain about 5ghz so we also ran 5.7" how did HU get so goated

  • @thestrykernet
    @thestrykernet 14 วันที่ผ่านมา +1

    Intel massively boosted the cache available with Emerald Rapids while keeping the same/higher clockspeeds and power consumption as the Sapphire Rapids parts they replaced. That's not a perfect comparison given the 2 vs 4 tile design, but it does at least indicate they wouldn't have to sacrifice peak clocks like AMD does. I think Intel's lack of extra cache at the consumer level has more to do with it being expensive for them to implement versus the volume of sales available to such a CPU. I'd be curious to see a 12100 added to the 4c results since it only has 12MB L3, but this would require one of the motherboards with an external clock multiplier.
    Really appreciate this sort of content as it's always good, unique and interesting.

  • @willfancher9775
    @willfancher9775 12 วันที่ผ่านมา

    Great work. The only issue I have with this testing is that 3D vcache adds a lot more cache than you tested here, and the performance effects might not be linear. It could be that there is some threshold of cache that, once surpassed, begins having more significant results. Of course there's no way to test this, so there's not much use speculating. It would be really cool for someone more knowledgeable in chip design to drill down and explain the "why" of these results, though.

  • @Pedone_Rosso
    @Pedone_Rosso 14 วันที่ผ่านมา +4

    Science videos are the best videos.
    (Except when I'm actually interested in buying hardware shortly after watching the video)
    Thanks for your tests and videos!

  • @drival060590
    @drival060590 13 วันที่ผ่านมา +1

    I would find it super interesting to see how much power the 14700K/14900K(S) consumes with disabled e-cores in games at 5.0 GHz (8c/16t). A Ryzen 7 7800X3D clocks similarly and it would be interesting to see what Intel's efficiency looks like when the processors are undervolted or clocked significantly lower at 5.0 GHz compared to completely over the top clock frequencies.

  • @chrischen6664
    @chrischen6664 14 วันที่ผ่านมา +1

    Nice video Steve, but how about another video on 3D V-cache on the Phenom x6 1055t?

    • @cooldudep
      @cooldudep 14 วันที่ผ่านมา

      What?

  • @xxovereyexx5019
    @xxovereyexx5019 14 วันที่ผ่านมา +18

    In general yes all cpu & gpu can get a benefit from stacked cache, but each cpu & gpu has a different case or problem.
    Stacked cache is one of the end game in gaming, smart move by amd

    • @kosmosyche
      @kosmosyche 14 วันที่ผ่านมา +10

      Common sense tells me it's a great way to decouple CPU performance from memory and memory controller speed and bandwidth restraints. After all this is the main reason cache was invented - to minimize memory addressing as much as possible, and especially when it slows things down quite a bit and holds CPU performance. I wonder if the benefit from bigger L3 cache would be higher if tested with lower clocked RAM or even with DDR4 RAM. Basically my theory is there needs to be more of a memory bottleneck for a bigger L3 cache to really start to shine and make a difference.

    • @jamegumb7298
      @jamegumb7298 14 วันที่ผ่านมา

      @@kosmosyche L2/L3 will always be faster than ram.

    • @PowellCat745
      @PowellCat745 13 วันที่ผ่านมา

      @@jamegumb7298Adding more cache can hurt latency and therefore performance. Bigger doesn’t always mean better, but since Zen 4 is severely bottlenecked by the infinity fabric, it benefits a lot from more L3 cache.

    • @jamegumb7298
      @jamegumb7298 13 วันที่ผ่านมา +1

      @@PowellCat745
      It _can_ add latency but does not have to. On Zen it does because chiplets. If one core needs from the cache on the other chiplet, latency penalty.
      I have a 7900X3D since a few weeks and measured it once using a bench tool. Latency to the big cache is hit but if both chiplets had the cache it would be far less an issue with a good scheduler.
      IF is another factor but I do not think it is _that_ bottlenecked by it, at least not in every scneario. The large X3D cache makes _some_ applications (mostly games) faster but for other stuff I'd have been better off using the 7900X.

    • @PowellCat745
      @PowellCat745 13 วันที่ผ่านมา

      @@jamegumb7298 more cache always results in higher latency. Do you even know how cache line works? AMD tried to stack two 64MB cache blocks on **one** CCD and performance went backwards. It has nothing to do with 2 CCDs.

  • @dangingerich2559
    @dangingerich2559 14 วันที่ผ่านมา

    Great analysis, as usual.

  • @ttcustompcs
    @ttcustompcs 12 วันที่ผ่านมา

    Do you think that when intel goes chiplet they will need Vcache? I would be interested to know if they will have similar issues

  • @naamadossantossilva4736
    @naamadossantossilva4736 14 วันที่ผ่านมา +1

    Not good when Intel is already facing silicon degradation issues with the 250W PL2.

  • @jimmyjiang3413
    @jimmyjiang3413 14 วันที่ผ่านมา

    I mean going forward needs a 3D V-cache variant for workstation CPUs as well cuz maybe bigger L3 cache shall also help pro visualization CAD rendering workloads in addition to using pro GPUs. This is where there maybe a necessity for creating Ryzen Pro variant for both workstation OEMs and some of SKUs available to general public (through retail for select higher end SKUs). There, there shall be need for dual 3D V-cache equipped CCDs for workstation rendering loads, and similar for Ryzen Threadripper Pro with 3D V-cache for even more workstations

  • @ETophales
    @ETophales 13 วันที่ผ่านมา

    There's a major difference between AMD and Intel with regards to the L3 cache in that AMD's L3 cache is linked to all cores, and remains the same size when fewer cores are available, while Intel's L3 cache has 3MB per P core or four E cores, so if a CPU has fewer cores, the L3 is reduced. This leads to quite a few questions, the first of them being: when you disable cores, do the other cores still have access to their cache? This would affect how valid the results here are. I think it would be good to verify this (by using some software which reports memory access performance for different data sizes) just to make sure.
    (Edit: There does seem to be a change in results based on the amount of cache, which might indicate that this isn't a problem, but it doesn't always show. After testing that, it would be interesting to test whether there's higher latency accessing the cache of another core, though that probably won't be trivial to test.)
    (Edit 2: Raptor Lake has 2MB L2 cache per core, as opposed to 1MB for Zen 4 and 512KB for Zen 3, which could be why L3 cache is less important on the Intel side.)

  • @rodrigoferreiramaciel4815
    @rodrigoferreiramaciel4815 12 วันที่ผ่านมา

    I wish you guys made some big comparison between generations of one single family of chips, like taking the last 5 or 10 i5's and looking at how much improvement each gen brings. That would be great for anyone that is looking to upgrade an old pc and is considering buying used pieces of hardware. I'd love to see how it compares between intel and amd

  • @rickh8380
    @rickh8380 14 วันที่ผ่านมา

    Great content as usual. Thanks for the info.

  • @jlgroovetek
    @jlgroovetek 10 วันที่ผ่านมา

    In some of the tests with 8 cores, 5% performance increase from just 36mb vs 33mb cache. If there was another 30mb increase, there could be 50% increase based on that, which becomes very significant, no?

  • @pristine3700
    @pristine3700 14 วันที่ผ่านมา

    Steve, have you been able to see Level1 Tech's video on 'Liqid Fabric PCIe'? The concept seems tailored to someone like you who has to swap out multiple GPU/CPU for testing.

  • @Petch85
    @Petch85 14 วันที่ผ่านมา +2

    Interesting video.
    But what is actually stored in the L3 V-cache when gaming. Course 36 MB to 96MB is quiet a lot more than from 33-36 MB.
    Maybe 96 MB is enough for some games to reduce the number of calls to memory needed for each frame, but 36MB requires several calls. That might not be a linear scaling.
    But with a linear scaling The last of Us part 1 gained 5% by adding 3 MB cache (9%). Thus adding 60 MB (266%) would give you 100% (133% if you used the %).
    This is clearly not true, course L3 cache and performance do not scale linear. And this is only for one game.
    But we already know that X3D chips only benefits in some games. But I think your data shows that Intel could benefit form having more cache.
    If we try to imagine the perfect memory system for a CPU, that would be a memory system that always have the data needed available for computation. Memory faster than that would not provide any benefit. Thus if the RAM was as fast as the L3 cache that would be the same as the L3 cache having the same size as your memory.
    The question is, how many clock cycles is an Intel CPU wasting while gaming. I don't know if that is something any tool could tell, it might only be intel who cooled even hope to know this. But if 50% of all cycles are wasted infinit memory speed could only give os a 100% performance improvement. But I have no idea how many cycles are wasted in a cpu while running a game.
    I tried looking for a Intel Xeon CPU with more cache, but you end up with minimum something like a W5-3435X (16 cores 45 MB Smart Cache) or the w7-3455 (24 cores 67.5 MB Smart Cache) (or w9-3495X, 56 cores 105 MB Smart Cache). They are all expensive or super expensive, they might be ok at video editing if you want to give it a go 😂. But it is going to be tuff, the motherboards and BIOS on this type of systems are often very limited and non of them can run at 5 GHz (unless you find a way to OC them), so you have to rerun the benchmarks at 4.5 GHz or something, to do the testing.... I would defiantly watch the video, but I think it would be hard to get a good return on investment unless you finde something else you also could test. Maybe number of memory channels, gaming with 1TB of memory, (google chrome with 4 TB memory 😂), lov clocked P cores vs E cores (course the W5, 7 and 9 only have performance cores, thus they have to clock down when using 16, 24 or 56 of them) , cooling test (I think they limit them to 270W all core). Maybe others can come up with better idears but those videos sound more like a LTT video than a HU video to me.

  • @ramiad90
    @ramiad90 14 วันที่ผ่านมา +1

    Good review, but note that the difference in cash between the i7 and i9 here is just 3mb
    With a larger cash pool for the cores to use freely things might be different
    So to wrap it up we woulnd never know for the 14th gen if more cash would help in gaming

    • @OGPatriot03
      @OGPatriot03 14 วันที่ผ่านมา

      Obviously more cache is better if it doesn't sacrifice speed/latency.

  • @NenMage
    @NenMage 14 วันที่ผ่านมา

    what are the full bios settings of an 13700k to clock it to 5.5GHz and disabling my e-cores completely, I'm running a NZXT N7 Z690 motherboard, thank you!

  • @Nelthalin
    @Nelthalin 13 วันที่ผ่านมา

    Nice video! I expected more difference with the L3.
    At je start of the video you said the 3D v-cache increases power consumption. While that is correct amd clocks the cores lower and issues lower voltages. Because of that those parts use less power than a 5800X and its faster with games. It offsets the powr consumption of the cache. The performance per watt of these parts is really good. My 5800X3D used like 100w full load the 7800X3D even 5-10 watt less.

  • @REgamesplayer
    @REgamesplayer 13 วันที่ผ่านมา

    This is why I love Intel chips. They focus on what is truly important for a CPU. A single threaded performance. 7800x3D doesn't do that and it runs terribly single thread apps like MSI Afterburner. It also tend to bottleneck at its main core.
    Whatever we like it or not, most apps are not excellently optimized. There will always be a higher dependence on a single core and CPU manufacturers should move towards prime core concept. Making a primary core a lot bigger and powerful than the others. Then they should add performance cores and then efficiency cores which they exclusively use for background applications. Such a design is a lot better suited for high performance gaming.

  • @vladislavkaras491
    @vladislavkaras491 13 วันที่ผ่านมา

    Oh, so games starts to utilize more than 4 cores!
    What a great news for Xeon owners! :D
    Thanks for the benchmarks!

  • @osenoqxd_8369
    @osenoqxd_8369 14 วันที่ผ่านมา +11

    Intel would likely segment these to the higher end i9 CPUs, giving another excuse to buy an i9 for gaming. They never made sense for gaming anyway because you would not use more than 1/3 of the cores and drew way too much power. An i5 such as the 13600K or R5 7600X gets most of the performance of these higher end chips for a fraction of their price

    • @evergreatest7644
      @evergreatest7644 14 วันที่ผ่านมา

      What about I7 and R7?

    • @nimrodery
      @nimrodery 14 วันที่ผ่านมา +1

      @@evergreatest7644 Intel doesn't make the R7 and the i7 (or whatever they're called now) aren't in the same price range, since buying more cores doesn't help for gaming the i9 (or whatever) needs more marketing appeal. AMD's product range shows more scalability or performance with price, the cache isn't just marketing (although it could highlight differences between AMD/Intel memory subsystems).

    • @Aggrofool
      @Aggrofool 14 วันที่ผ่านมา

      Problem with buying i9 as a gamer is you're stuck with a whole bunch of useless e-cores.

  • @joachimsoubari3772
    @joachimsoubari3772 14 วันที่ผ่านมา

    Great content thx !

  • @CakeCh.
    @CakeCh. 14 วันที่ผ่านมา +1

    12:27 So I'm interested to see a test results with some Xeon CPUs.

  • @xsvrrx
    @xsvrrx 13 วันที่ผ่านมา

    so the cache is also usesless for cores when they are turned off?

  • @khachoang1719
    @khachoang1719 12 วันที่ผ่านมา

    Nice, but is it because their drivers are set to run with limited cache in the first place?

  • @J_..._
    @J_..._ 14 วันที่ผ่านมา

    I know Intel and AMD have different methods of evicting lines in the L2/L3 cache. Explaining the differences between them might make for an interesting short~ish video.

  • @WilliamJasonSherwood
    @WilliamJasonSherwood 12 วันที่ผ่านมา

    Not disagreeing with your results, but it is possible that there is a threshold where between the 36 MB L-3 Cache of the Intel Parts and the nearly triple 96 MB of the X3D parts. Cache would likely function similarly to ram/v-ram, where if you either have enough to contain all the things you frequently need, or you don't and going from say 8GB to 10GB won't make a significant difference, but going to 16GB suddenly increases performance by 60% (okay, maybe 30%) or something ridiculous.
    While an absolutely ridiculous use, one potential way to test for this could be using something like the Intel w9-3495X with its 105 MB of cache and disabling the appropriate number of cores, and testing different cache configurations. But they only clock to like 4.8 Ghz.

  • @yeahitsme4427
    @yeahitsme4427 13 วันที่ผ่านมา

    It would be good to join 7800X3D to the charts as a comparison, because its an 8core part with 5GHz boost clock. The same clock you limited in the Intel series.

  • @Frozoken
    @Frozoken 14 วันที่ผ่านมา

    lower core counts probably increases per p core usage meaning the individual l2 cache runs out more and the l3 as a result has to get used more aa a result

  • @B3TR0Z
    @B3TR0Z 14 วันที่ผ่านมา

    But why did you reduse the Ring speed...?

  • @EYESCREAM0
    @EYESCREAM0 13 วันที่ผ่านมา

    Is hyper-threading enabled or disabled?

  • @patrikb1161
    @patrikb1161 14 วันที่ผ่านมา

    Ty!
    Could u do a 14900K pcore only review, pretty plz?

  • @Indyfficient
    @Indyfficient 14 วันที่ผ่านมา

    I’m curious though, are these results due to hardware limitations or due to the manner in which amd vs Intel have their scheduling programmed for instance?

  • @pino_de_vogel
    @pino_de_vogel 13 วันที่ผ่านมา

    i have a feeling the architecture has bin stretched so much the past 4-6 years that there are other limiting factors within the architecture that just cant keep up with the cache or clocks.

  • @pvdgucht
    @pvdgucht 3 วันที่ผ่านมา

    More cache seems to make a huge difference in flight simulators. More than in any other “game”. Idk why, the only reason I can think of is that maybe the aerodynamic calculation fits inside the cache and because the sim calculates that 2 to 4 times per frame that might make a difference idk 🤷🏼‍♂️

  • @stephenkennedy6358
    @stephenkennedy6358 14 วันที่ผ่านมา

    I wonder if we would see a bigger difference between 8 and 6 cores if hyperthreading was disabled

  • @dominicsilvestre9412
    @dominicsilvestre9412 13 วันที่ผ่านมา

    Remember i7-5775C with 128mb L4 cache? good times.

  • @ramonzaions7522
    @ramonzaions7522 14 วันที่ผ่านมา +1

    Whit this Voltage And Tempture 3D Cache will burn!

  • @HDJess
    @HDJess 14 วันที่ผ่านมา

    If they would just slap 3dv-cache on current intel cpus, nothing much would happen. The architecture needs some tuning as well to make use of that extra cache. The software, including games, need to be optimized too to make good use of that cache. It is likely that a game's engine running on an intel CPU with 36MB of L3 cache would not even know what to do with an extra 100MB of L3 cache, if the developers haven't tested and built the game's engine for it.

  • @Bryan-T
    @Bryan-T 14 วันที่ผ่านมา

    Fascinating. I would've thought the cache was a factor.

  • @danielt91
    @danielt91 14 วันที่ผ่านมา +7

    It would be interesting to see the exact same test with the Ryzen 7000 lineup. Again - same clock speed, same core count, just differences in l3 caches (including the x3ds)
    At a later time it could even be used to compare the IPC gains of Ryzen 9000 🙂

  • @chromedonut2441
    @chromedonut2441 14 วันที่ผ่านมา +3

    Maybe Intel should consider having someone else manufacture their chips for them since they are till having trouble shrinking their process node. How would the performance differ for 14th Gen if say it were done on the same process node as Zen 4?

    • @yuvanraj2271
      @yuvanraj2271 14 วันที่ผ่านมา +2

      What are you talking about? They are already in the process of finalizing the 15th gen arrow lake chips. They have intel's smaller node as well as tsmc's small node.

    • @mingyi456
      @mingyi456 14 วันที่ผ่านมา

      It seems the problem with intel delays now is their design teams, not their foundry teams. Their nodes seem to be ready long ago, just that there is nothing to use them. Ironically, intel also seems to have a cost issue with their nodes, because using tsmc now seems to save them money over using their own nodes, even after letting tsmc get their share of profits.

    • @moist_ointment
      @moist_ointment 14 วันที่ผ่านมา

      RPL was DTCO'd too heavily with Intel 7. Design and Fabs were too tightly coupled and only newer gens separated that. Porting RPL to another node would've been too expensive

  • @IPowder4105
    @IPowder4105 14 วันที่ผ่านมา +3

    With all that heat they can't use 3d v-cache

    • @maynardburger
      @maynardburger 14 วันที่ผ่านมา

      Has nothing to do with it. The CPU's do not inherently put out a ton of heat.

  • @tldrinfographics5769
    @tldrinfographics5769 12 วันที่ผ่านมา

    You know what’s really amazing?
    Intel being able to keep its speed crown for this long without TSMC advantage.
    They did it all without chiplets or 3nm…
    Imagine if 14th gen was on 3nm

  • @ZeroHourProductions407
    @ZeroHourProductions407 14 วันที่ผ่านมา

    Very much appreciating the specialized cpu content. So evidently, present intel architecture doesn't give a rat about cache, it's about clocks, and cores to a lesser extent.

  • @user-vy7td1wp2w
    @user-vy7td1wp2w 13 วันที่ผ่านมา

    It would have been better that tested with heavily Cache bound games. Such as COD, Assetta Corsa, Forza, Starcraft and Factorio etc.
    Most games in this review, (except Cyberpunk) usually are not very well to test it because they are mainly more GPU-bound AAA titles.

  • @techluvin7691
    @techluvin7691 14 วันที่ผ่านมา

    The only thing that can change Intels fortunes is a complete new architecture that doesn’t require a 1000 watt psu and an AIO to “maybe” cool the thing.

  • @thejumper9303
    @thejumper9303 13 วันที่ผ่านมา

    You expect that every test is just frequency limited, even though you never tested 2 different cach amounts vs frequency.
    It could also be engine hard frequency caps like many Ubisoft titles/ports as well as scheduling problems, engine only utilizing 4/6 cores,....
    But for you it can only be frequency dependent. Which might be a symptom but most currently not the main cause

  • @felipeavlopes
    @felipeavlopes 14 วันที่ผ่านมา

    I wonder if the performance reduction when dropping to 4 cores isn't caused mosly by the Operating System itself rather than anything else? Modern Windows are resource gluttons that love to use more resources than what would really be necessary for computers to work.

  • @korcommander
    @korcommander 14 วันที่ผ่านมา

    If only Intel can get that heat under control

  • @ShermSpinner
    @ShermSpinner 14 วันที่ผ่านมา +1

    Very interesting results, especially with the 1ßth gen testing you meantioned in mind.
    I'll say it would be interesing to see how or if that scaling behavior changes with slower memory, given that on AMD it really helped with memory bandwidth limits.
    Though of course i understand that you cannot possibly test everything and there is definitely an argument to be made that even *if* running something like base JDEC DDR5 would flip things around and make cache a massive deal on Intel it wouldnt really change the conclusion at all, given that intel can run these higher speeds just fine.

  • @lietuviss1
    @lietuviss1 11 วันที่ผ่านมา

    It would nice to see the architecture differences, different cache coherence protocols between gens and VS AMD maybe? Faster DDR5 ram could reduce the bottleneck of fetching from ram on cache misses.

  • @jporter504
    @jporter504 14 วันที่ผ่านมา

    Good information.

  • @GeneralLee131
    @GeneralLee131 14 วันที่ผ่านมา

    With NVME SSD bandwidth approaching that of RAM memory, and the discovery that cache size is an apparent bottleneck in all systems, i can envision a near future where L2 cache is up to a gigabyte, L3 cache could grow to several gigabytes, RAM is entirely deleted from the system, and the remainder of left up to the swap file on the disk.

  • @cracklingice
    @cracklingice 14 วันที่ผ่านมา

    Not that it matters much, but it'd be interesting to have this data side by side with some ryzen data for the same selection of titles.

  • @hotscott6619
    @hotscott6619 14 วันที่ผ่านมา +1

    know do same test with thread ripper how do we know all cashe used

    • @Hardwareunboxed
      @Hardwareunboxed  14 วันที่ผ่านมา +1

      TR is completely different and the L3 cache is per CCD.

  • @cliffordjohnson943
    @cliffordjohnson943 14 วันที่ผ่านมา

    I’m interested in what their new Ultra desktop CPUs do without the hyper threading.

  • @TheZoenGaming
    @TheZoenGaming 14 วันที่ผ่านมา +1

    My first thought is maybe, but only a little since the V-cache is more heat sensitive and Intel's chips seem to be less thermally efficient than AMD's.

  • @TheGamerUnknown
    @TheGamerUnknown 14 วันที่ผ่านมา

    Why do these at ultra quality if you're measuring the CPU?

  • @Gnarfendorf
    @Gnarfendorf 14 วันที่ผ่านมา

    Given that the 7800x3d also only has 8 cores and doesnt really clock higher than 5ghz iirc, maybe the issue is that even with 36mb l3 its simply not enough to have a noticable impact?

  • @ojassarup258
    @ojassarup258 14 วันที่ผ่านมา

    Wasn't Opera doing some sketchy stuff a while back?

  • @AlexHusTech
    @AlexHusTech 14 วันที่ผ่านมา +1

    *Great video, learning something new every time*

  • @ThePurplePassage
    @ThePurplePassage 14 วันที่ผ่านมา

    And here I was thinking that Adamantine would enable Intel to retake the performance crown

  • @XAB0R
    @XAB0R 14 วันที่ผ่านมา

    Question about 5:40 R&C Rift Apart. It looked like on a previous video that AMD had much more difference in scaling with cores than Intel. Is there a reason for this? The video at 6,30 called "Was The Ryzen 5 5600 a Mistake...."

  • @funkymoine821
    @funkymoine821 5 วันที่ผ่านมา

    SUGGESTION@HardwareUnboxed:
    do exactly the same tests, but with cracked versions of the games. (where drm has been removed).
    the results will tell a completely different story. (not talking about how drm hogs game performance; talking about how they also hog the scalability)