Support our in-depth testing, journalism, and analysis! Grab a Modmat, mouse mat, shirt, or more at store.gamersnexus.net/ Watch our PSU Reviews, Deep-Dives, & Explainers Playlist! th-cam.com/video/yDX_1PWUWdw/w-d-xo.html
THANK you for this video! We've had customers GPU's returned to us because they exhibit that exact behavior, tested (Corsair AX 1600's), found to be working flawlessly sent back and forth a couple more times before we've figured out that their PSU is (although above "spec") not handling the loads some of these monsters pull under spikes.. GPU's should have transient spike recommendations included. Easy as that. Bad marketing for Nvidia/AMD? Ok, we (small retailer) do not give a f*ck, current system is misleading. AND! A lot of customers don't enjoy getting "up sold" on PSU's.
I am a Mechanic (40+yrs), and I have a general rule. Your battery should be 2 CCA (cold cranking amps) per cubic inch of engine displacement. I follow the same rule when selecting power supplies. I make sure the PSU can handle near double what the system can draw. Your PSU will be most efficient at around 50%, it will also run cooler and quieter, and it will last a lot longer.
quick correction: the magnetic field is what creates the charge, not the other way around. Edit: I would like to see what effect undervolting has on these spikes if any at all. I know you can get some interesting behavior out of a GPU by undervolting a little too much.. but doing it just right seems to make it more stable since clocks are not constantly shifting.
Electrical Engineer here, I have to say, what a superb animation of how current clamps function. This piece of journalism is outstanding, you guys rock.
@@GamersNexus Industrial automation electrician here, and I wanted to say the same thing! You might be inspiring the next generation of industrial controls tradesmen and women! lol!
@@GamersNexus Andrew really killed it on that animation, his animations really add a whole new level of explanation to the videos over the last year or so.
The level of quality content your team is putting together is down right impressive. You guys keep this space in check and I truly respect what you all are doing.
I find it kind of interesting that this is just now being discussed for computers. I did car audio installation professionally for over 14 years and this was something we dealt with on a daily basis for huge car stereo installs. In fact it is still something that is dealt with today for multi amp power hungry car audio systems. We had to install large capacity capacitors in the system to "suck up" high transient power demands to keep from potentially hurting a cars alternator. You can visually see what these high capacity caps would do because the headlights wouldn't dim, or not as much, under large bass drops which would draw the most amount of power. Large caps are designed to quickly deliver large amounts of current very quickly while a cars alternator can provide large amounts of current but just over a longer period of time. A computer power supply is roughly the same. It produces a fixed voltage with a variable current output, very similar to a cars alternator, and neither is extremely suited for potentially a massive current draw in relatively short amount of time. For cars it has always been you upgrade to a larger alternator, or more than one if needed, you add several farad caps or both sometimes. It is getting to the point where either manufacturers are going to have to start recommending more accurate power supply requirements to account for this issue or someone is going to have to design a power bank of large caps that go in between the power hungry graphics cards and the power supplies.
@@StephenGillie The best solution is an elegant one. Graphics cards should implement their power limit properly. It seems like really messy design to me that you can run a load of never above 200W, yet for less than a millisecond it might spike to more than double that. This can be avoided with sane/smart design. But of course the GPU industry is the epitome of superficial power craze, so we cannot expect them to sacrifice any potential in order to avoid having to oversize your power supply.
The solution to transient spikes is not just huge capacitors, at least not on the power supply itself. The power supplies (switching DC/DC converters) have a maximum amount of output capacitance that they can deal with before they go out of stability and start oscillating (which will burn out components). A power supply manufacturer can't know how big the spikes will be for every GPU that you put in your system. That's where the board designers of the GPU come in, they should make sure that they have enough filtering on their power supply input so that they don't trip the OCP. They are the ones that should place the large capacitors on their board. But ideally you don't want to have these spikes at all, and you also place inductors in the line to suppress these huge spikes, so just a lot of caps alone could be detrimental to system stability.
@Lurch7861 No, measurement latency is not a college physics thing. Heck, in my CPE labs, which were the same ones the EEs took, we never bothered to correct for it. Quite frankly unless you're looking for something extremely specific, it's not worth the pain. Wall power is atrociously noisy as is, and on the DC side we aren't doing weeks of data analysis. We're fixing the problem! Heck, most of the O-scopes I used were CRT. There was no data export. Even the digital scopes I think the most I ever did was export screenshots. And my senior design project was on power monitoring...
@Lurch7861 : It's not a matter of having the education to know that you need to do something, it's about having the time and money to spend on doing it.
@Lurch7861 My background was hard science and I had to take more than a few college physics classes. Never was taught this… This is an electrical engineering concept.
Now PSU manufacturers or brands can capitalise on this issue by advertising how massive their Caps are in their unit. Nominal Load : 850W Transient Load : 1700W@20us
At a bare minimum, ATX 3.0 + PCIe 5.0 certification of PSUs (with independent verification) will go a long way to educate consumers here. SFX PSUs may perhaps give way to SFX-L..
@@ikjadoon Nah, there's too many people willing to spend hundreds more on a part to save a cm of space. I'm sure Corsair will find a way to make their SF series well over 1000W, if it means they can charge a boatload for them compared to normal ATX psus
Im hoping they also start to put more capacitance on the GPU board. Loading up the caps in the PSU destabilises the 12V rail which will reduce the stability of the whole system, not just the GPU The good news is capacitor technology is advancing relatively quickly, so it will at least be possible to keep up
@@LtdJorge all capacitors charge and discharge at the same speed, somewhere between a millisecond and a nanosecond for a full charge. But the voltage drops linearly as they discharge, so the more you load them up the more the 12 volt rail will drop below 12V.
This channel is much more than a “gamers nexus”. Thank you so much for your scientific, transparent and objective media reports. What you’re doing is invaluable.
Seriously, I honestly feel like his the name has been outgrown by what he is doing now. This is serious pc builder news. Although it all does tie back to gaming in the end I suppose to keep his scope in check.
I would dare to say GN is a bit like what savagegeese is to car reviews. So if you like honest people applying no-bullshit non-sellout scrutiny and are even remotely interested in cars, check him out. (He also has livestreams where he mentions how hard it is to do when you are not being the industry's monkey.)
Please - If you want to up your game on this, go look up the Middlebrook stability criteria and start reviewing power supplies by their impedance and crossover frequencies. You're great at this. I think you're one of the only reviewers that might care enough to actually go do it, educate your viewers on why it matters, all while keeping it entertaining. Middlebrook is the first step basic method for determining stability and impedance matching between power supplies (like the main PSU and one on a board). FYI, I'm a professional in this field (not specifically computers), but I did work on a motherboard power supply for the first Sandy Bridge early in my career. You covered it a bit, but everyone thinks capacitors are holdup. They're more for transient suppression and impedance. Those tiny caps on a 1V 100A rail aren't going to hold up Jack. The big caps on a power supply rail are better, but they'll only respond between roughly 20 Hz and a few kHz at most. Your small form factor supplies are pushing it a bit higher (50 Hz or so). What's more important is that higher wattage supplies is pushing the base impedance down and the crossover frequency doesn't matter as much; so stability is easier to guarantee. The 30x Nvidia series die has multiple full to almost zero load steps with durations of a few microseconds within a single frame. You can literally see a semi consistent pattern at the scene's frame rate if you tap between the on-board supply and the GPU. The on-board supply can take care of most of the transients in the kHz+ range, but those frame-harmonic frequencies (approx frame rate to a few multiplies) are smack dab in the middle of the bandwidth between the cards power supply rail and the main system supply and you have control bandwidth and impedance matching issues between the two. Ironically, power saving efforts for high performance cards in the future will make this worse. You want it to blast out a frame, then sip juice. It's not practical to store enough energy on a board to smooth it out. Edit: Also, you mentioned something about the phase shift between the current sensor and the voltage. That phase shift may be 'real'. I'd suggest calibrating it out on your probes measuring against a linear load. If you put your scope in 50 ohm impedance mode you can probably cal the delay between voltage and current while measuring the current into the voltage probe itself, just using the cal rail on the scope. If your current sensor resolution is too low you'll need a signal generator to do that.
I really hope that Steve and the team find your comment and read it. GN strikes a great balance between in-depth reviews while keeping it accessible for most people, so I'm sure they would be able to incorporate all this into their videos.
Graphics cards need to have enough capacitance to cover the transient spikes they generate and offer a smooth enough power draw that at no point does the card go beyond the publicly stated max power draw. If they have a device that does not or can not do that then it is out of spec and they need to publicly acknowledge this phenomenon and publicly report the max transient power spikes so pc builders can ensure they pair a power supply that can supply the demand. Anything less is not acceptable. This is clearly a design problem on the graphics card manufacturer's side. How many consumers have dead power supplies only because graphics manufacturers incorrectly designed a card and did not report the true max power draw? Repeatedly causing customer power supplies to repeatedly trip OCP and the customer not knowing what it could possibly be given the power budget allotted to the components and the power supply capability.
This was my thinking since these "mystery black-outs" started and still is. The power supply might be the better place for a huge capacitor, though. But it seem like it doesn't actually need to be that huge.
i came here to say something like this and this is perfectly summed up. it is 100% upto the card manufacturers to disclose the max power draw even if it is "transient" because its still the card that is causing the problem. its a massive copout for such a huge company to blame another company "because the psu shuts off" nah my dudes. make your cards within spec or disclose the maximum draw instead of pointing fingers. we have never had problems like this in the past untill nvidia started making these cards that seem to double power usage every generation. as a pc builder it would be super embarrassing to complete a build and have the customer say that there are these kinds of problems. reasons like this is exactly why i prefer radeon cards at the moment. they require much less power for similar performance in most settings.
Just want to also say that the animated depictions were both really helpful and well done. It’s easy to take those for granted, but there’s probably a lot of effort that goes into those, too!
I came to the comments to say exactly this. At 21:05 they didn't have to go through the methodology and explain how it works, but they did. Not only that, they put an animation together to show how the instruments measure the EMF to come up with their readings. That's not a small deal. That takes a lot of extra effort and I said wow, that's really well done, out loud when that came up.
I noticed and was going to say the same thing. Really high quality animations and a fantastic explanation of how the tool worked. Kudos to the GN team.
This is the most well timed video I have ever experienced. I was literally just telling a friend that his crashes might be from these "power spikes," and he was not understanding what I was trying to say. I'm hoping this video with all your graphs and analysis will make more sense to him. Thanks for the breakdown!
What sucks is a 1.2KW or higher PSU would more likely than not solve the problem completely but that is one hell of a cost to pay for system stability that should be a given if the GPU didn't go overboard on the draw randomly. Locking the FPS would also likely make it not happen and in the majority of games that wouldn't even be something that hurts the experience, but in several titles you want the highest framerates possible so you see the most up to date frames at all times to be competitive and in those titles a new PSU seems to be the only option.
Hmm, I do have random system restarts, now that I got a 3060 ti with a 550W PurePower11, but they happened only when not under any load, and changing the PCI-e setting from "auto" to 3.0 seems to have fixed them. Or at least I haven't had any restars since yesterday morning, when I did that. Although they only happened once every 5 if not 10 hours, so can't say for sure yet.
@@bdhale34 This only affects some GPUs and systems I reckon though. Because I'm on an 850w PSU with a card that draws up to 430w constant power and I've never had any sort of shutdown or instability in nearly a year and a half.
I view the issue as sort of like a hypothetical car that locks up when the RPM reaches a certain threshold because it requires more gas than the fuel pump can push out so the fuel pump cuts off for safety. Even if the car OEM picks a fuel pump that can supply enough gas for the average RPM, it won't be good enough for when the driver has to floor it. However, just because you pick a fuel pump with the capacity does not mean it can push enough fuel fast enough, so you could have a car that doesn't shut off but does essentially run on fumes when flooring it.
yeah, if the normal power draw of a 3090 ti is as high as it is without even factoring in the transients, then I can't imagine what the 4090 ti will be like.
From an electronic point of view it makes much more sense to solve the problem on the GPU side where the load characteristics are known and you can even actually actively control the spikes.
@@TheSkace The correct solution is that GPU manufacturers stop lying. If GPU manufacturer says "290 W" but oscilloscope shows 620 W, then GPU manufacturer is lying about power usage. When all manufacturers tell the real power usage, getting the correct PSU is easy. With the current lying trend, you cannot know tomorrow for sure if you need to get a PSU with 2x, 3x, 4x or even more power compared to the number that GPU manufacturer prints on the box. If GPU manufacturer doesn't want to print 620 W in their spec sheet, they need to add enough capacitors to get the power need closer to average.
@@urbansnipe That's absolutely true that building a GPU that has power spikes is cheaper. However, when GPU manufacturers lie about the power requirements of those cheaper GPUs, selecting a PSU is turned into gambling. When a single manufacturer says "290 W" do you really need 750 W, 900 W or 1200 W for the GPU? How much could any specific manufacturer lie? I'm absolutely fine with whatever spikes any GPU has as long as the manufacturer *honestly* declares the needed power levels instead of average power usage which doesn't include actual power required for the spikes.
It’s great to see more people getting in on the whole transient issue. First encountered it when Igor‘s lab made it the focus of their power draw and supply tests but spikes and transients will only get more important to look at the higher power draw we see and the more companies rely on boost clock algorithms and sub millisecond control algorithms for power limits and current draws
Speaking of issues... there will be fires in some older houses or houses with often sub-par repairs like in hurricane zones. It's pretty messed up when you need an electrician to do a safety check before you plug in your new GPU, will Nvidia give 4090 buyers a voucher for inspection and testing?
@@Mr.Morden as far as any business is concerned the quality of the wiring of your house is a you problem, this includes the company that supplies your electricity
@@Mr.Morden Not to forget the 110V Powergrid in the US isnt even that great... For me its always funny when people on reddit tell me that i NEED a UPS and iam just "Mate iam german we have 240V and unless you live somewhere far outside and you are the last person on the line and its not burried in the earth or you do super important work you usually dont use them here"
@@mrn234 quality of powergrid in the US is largely down to the state that you reside in. I've never had issues with power in any location I've lived in, and I've lived in the middle of nowhere as well as in the middle of a city with a population of 9 million people.
It's the way how these companies "cheat" better performance, just let the GPU spike for milliseconds so the average performance is higher. Spikes so minor but also too severe to some PSUs, if not bought mindfully. But how should a normal person know about it, to avoid the problems some 3090s had? That's why i like GN so much and people like Igor but also Hardware Unboxed.
GN is straight up making video essays with all of their expertise about tech put into it. This is so good. Props to the whole team. Keep holding the tech giants accountable.
2017: GPUs are causing massive worldwide power draw. 2018: Countries introduce laws against GPU crypto mining. 2019: GPU prices 5x higher than the systems they're built for. 2021: GPUs affected by worldwide component shortage. 2022: GPUs are frying entire systems due to unprecedented / sudden power consumption. Man, GPUs are like the bad kid in the neighborhood.
i was not expecting the editing to be so good on this video. not to say your regular videos are edited badly, but this is above and beyond with all the animations and stuff. well done, and great information in here
This honestly explains so much about my early experience with my 3090. I saw on nvidias website they recommended 750W for a 3090 with a 10900k so I decided to get a 850W PSU to give myself a little headroom. It didn't matter at all as I was constantly black screening cutting out during gaming which I later learned was OCP triggering. It was super frustrating as 1/3 of the people I talked to said I didn't have enough watts, 1/3 said I was borderline and 1/3 said I had more than enough. Long story short is I had to move up to 1000W to stabilize the build and now I believe the reason why is because of the transient spikes you guys are talking about. I really appreciate your video highlighting these issues as I thought I was crazy for a long time in regards to this.
I'm rocking a Corsair RM850x for a 3090 FE and 12900K non-overclocked. I'm scared to even touch clock settings with these kinds of spikes. Probably going to have to upgrade the PSU at some point. My first gaming setup in 2012 was on a Corsair 600 watt PSU. It's crazy to think how ridiculous things have gotten the past several years.
@@braf7349 my wife's RM850x finally started showing it's age after about 4 years with a 2080Ti. It started black screening in games that were only moderately demanding. Monitoring rail voltages in HWInfo, I could see random 0.3-0.5v drops on the 12v rail, even without any real load. I didn't feel like having to deal with the whole warranty process, so she inherited my old EVGA G2 850w (5950X/3090 setup) and it's still going strong for her. Good luck with your setup! At least a PSU is a pretty easy upgrade if you do have to swap it out.
The 10900k chugs around 200-250W on turbo, so a 350W 3090 with spikes at 700W might indeed trip a touchy psu. (Especially if you have a simultaneous spike on the cpu)
Replaced my 1080 ti with a 3080 ti, had a 750w psu which should have been enough but it wasn't, system ran stable stock when I went with a 1000w psu, likely have some headroom now but I prefer no crashes vs a few more fps
Had a AIO Vega 64 card on a x370 MB and a Rosewill 1350 PSU . Started having this issue occasionally in 2020. Started being a more consistent problem in 2021 where I could reliably shut down my system by loading a stage in Destiny. I was beside myself trying to troubleshoot, crossing my fingers with each AMD driver update. When GPU prices came down from ridiculous to just gut punch levels in 2021 I desperation bought a AIO 6900X card and a B550 board, assuming it was my card struggling to keep up with newer games as I play in 4K. The problem seemed better at first, but I about had a fit of rage when it happened again for the first time. I would play for several more months clenching my chair from the anxiety of knowing my next crash could be at any moment, and I could almost never consistently reproduce the problem. It wasn't until I saw you mention this while you were researching this video that I finally had my "AHAA, this explains everything" moment. I switched PSU to a Seasonic 850, lower power but much higher quality. After the change it's been smooth sailing since. Huge thank you to GN, I had just thought "I definitely have enough power", and had never considered the OCP angle. Will put a lot more effort into choosing power supplies from now on, and you're spot on in how frustrating it is knowing that each of your components, PSU, graphics, and MB can factor in to how this problem manifests. The average consumer could never get enough info to make decisions on parts in regards to this issue and even most independent reviews would struggle to find the time, resources, and motivation to delve into this topic. Great on you for leading the charge, and this level of production is immensely impressive for an independent group. This channel has become a necessity for being the consumers voice of reason to the manufacturers and developers. Unfortunately for me I noticed your work here too late to save my wallet, but now I know better and am a full subscriber and follower.Thanks!
I had this exact same problem. After waiting for hours in line to get into Microcenter when the rtx3070 launched, the card I brought home wouldn’t work on any of my systems, and kept tripping the power supplies internal wattage limit. I tried for over a week straight to troubleshoot the issue but unfortunately had to return the card I worked so hard to get in the first place. So happy someone is finally looking into this, and that I wasn’t just crazy. I hope this means the responsible parties will take a look at this problem and work to address it, as it is just absolutely infuriating to deal with when it comes up.
So, a bit of a tipoff about testing for these transient spikes, as I recently had a problem with them myself that resulted in me just buying on on-sale EVGA 1300W so it can just soak up the problem, because thats far cheaper on time than any other solution. The tipoff here is the BEST way I found of tripping it, happened literally within a few seconds every time, was fallout 4 VR. In general I was able to benchmark the system as I wished and never had even one shutdown during any benchmarks, however I had one shutdown while loading GTA online (weirdly), one shutdown in Phasmophobia VR, and a total of atleast 30 shutdowns in fallout 4 VR before I even got through character creation. I'd highly recommend it as the BEST way to see if a spike can take out your system power. My System contains a 5950x, an MSI 3080ti, (64GB 3600MHz RAM) and was running on a Coolermaster 850W MWE V2, if anyone wants to know
3080ti isnt that far off from an 3090 ... afaik a 3090 can have spikes at almost 600w, worst case for a 5950x would be like 250w and for stuff like ssd, fans, lights etc ill give them ~50 to 75w ... espacially if you look at newer boards ... then i'll take 20% above it and this results in my psu capacity i want to buy for a rig.
I had a similar problem like a decade ago with a 570GTX or whatever it was. Every time I looked at water in World of Warcraft the computer would shut down within seconds. Took a while to isolate the behavior down to that, but once I put realized that was happening I could reproduce it every time. Put a bigger PSU in the system and the problem went away. Transients are fun. Bigger picture though, all PC components are ballooning in power consumption. Dennard scaling was already dying off for years, but it's really hit a wall in the past couple of years. A significant source of extra performance is now coming from simply letting components draw more power instead of getting more performance from a similar amount of power like we saw in years past. This is only going to get worse from all vendors as time goes on.
@@ngc4486diane People don't often understand, and there's a lot of bad advice out there. Pre-builts too often have substandard components, so I'm not surprised at all.
Possibly we need some sort of regulation for power spikes like this , similar to EMC regulations ? A PSU should be expected to be able to handle spikes yes ( due to things like inrush) but not to this level , not in the consumer space.
Most consumer PSU's are generally built to the ATX specifications for power supplies. That spec doesn't really address any specific requirements for transients. However, it does say that DC output voltage must remain within a certain range under ANY load conditions that are within the specified output ratings of the supply. So, if your supply says it can do 40A on a +12v rail, the ATX spec doesn't care if that is a nice steady 40A or if it switches between 40A and 1A every microsecond. It expects your regulation to be within spec no matter how crazy the load cycle is. This is unrealistic, but the spec is silent on transients. Transient requirements for the rails could be added to the ATX spec, but failing that, GPU and PSU manufacturers would have to publish their own tolerances for transients and we would have to compare them ourselves. That or NVidia and AMD are going to have to start their own certification program and publish their own specs so PSU manufacturers can put a "NVidia 4080 Certified" or "AMD 7800 Approved" sticker on the box to help tell consumers what to expect. This is just another indication of just how far out of date ATX has become as a set of system building specs. ATX needs a major overhaul, but Intel isn't in a position to tell everyone what they have to do anymore.
@@kaiwenwu8148 that's not a bad rule of thumb. Maybe 2x instead of 3x, but at least 2x to be safe. Power supplies are typically in their peak efficiency around 50% loading anyway.
We need another episode, a sequel to this one, focusing on daisy-chained cables and multi-rail vs single rail configurations and how they relate to stability under transient loads
I am an electrical engineer, sometimes dealing with transient issues in an industrial environment where we work in kW and MW. I am also a keen audio enthusiast and know for example that a power amplifier is very dependent on its power supply's ability to deliver power at higher frequencies. I thus understand your review and kudos for pointing these issues out to prospective buyers. What the industry needs is a rating system for PC power supplies, in their ability to handle short term overloads. Premium manufacturers like Corsair can then use this rating system to distinguish their power supplies from the vanilla variety.
The issue with a rating system such as the 80+ certification that already exist for PSUs is that it becomes vague and eventually turns into a marketing label overtime. Unless there is a federal standards I think we will just be seeing more of the same.
I would side with @thelol1759 and suggest that the problem needs to be solved on the GPU side. The PSU needs to have solid bandwidth but there's usually a foot plus of cable between the PSU and the GPU. In the world of sudden, extreme current loads, the rail could be collapsing on the GPU side before the PSU has any chance of responding. CPUs have had spike current demands in excess of 100A for a long time but their VRM circuits are spec'd to handle those transients and not pass them on to the PSU so severely. GPU VRM circuits, and particularly their bulk capacitance is just not up to snuff if they can bring down a 12V rail that's basically all theirs.
The upcoming ATX 3.0 specification, which was finalized in April, is specifically supposed to address this issue. So, I'm really surprised there was no mention of it here, at all. I've been eagerly waiting for some kind of news from Gamers Nexus, about when ATX 3.0 power supplies might be expected to start rolling out, or if the RTX 4000 series cards will have the new 12+4 GPU power connector on them, which is part of the ATX 3.0 spec. Please ask one of the power supply manufacturers, about an ETA on ATX 3.0 PSUs. I'm very eager to find out, since I'm planning a Ryzen 7000 system.
They'll surely have the 12+4 connectors on the FE/reference cards, but are unlikely to require it of the AIBs. The big question is what the PSU requirements will be. I'd assume they'd say something like '4090 requires a 1000W PSU, or an 800W ATX 3.0 PSU'. Also, as you noted, its unclear if the ATX 3.0 PSUs will be ready for an October 40 series launch. It would be really dumb to buy at launch if you have to upgrade to an ATX 2.x PSU that's becoming obsolete.
Looks like you are expecting people go out to change their PSU with ATX 3.0 spec. Tell it to those people with no knowledge, or those with 1000W+ Platinum rating. I doubt those with some years old 750W Gold PSU or above will do that so this info is very useful for the years to come.
@@Xirpzy Assuming that everyone should buy into the new standard and that anyone can afford it is actually ignorant. A good PSU stays a good PSU for as long as there are devices it can run, and people either don't see a reason to change their PSU because of a new standard or they don't want to change their expensive high end PSU just because of a new standard. Some people may not even be able to afford a decent PSU at all. Your comment sounds very pretentious as if you're saying "lol just buy it you poor pleb". Very very disrespectful.
Twenty years ago, when I was in high school, car audio had a similar problem drawing down voltage when the bass hit. The "solution" was adding massive 12V capacitors with claimed 1-2 farad capacitance. A lot of them were junk. I wonder if we'll see similar snake oil for sale in the computer market soon.
I was sitting wondering about that. There *were* capacitors you could put in-line for audio cards. Those, well, *worked*, but didn't make a difference because, shocker, the audio cards did it on their own and didn't really have massive amounts of differences in demand anyway. Now I am starting to wonder if those products might work for this sort of thing. Just buffer it juuuuuuustt enough. Given that it didn't produce negative issues to it when tested for the audio cards.. erm. Maybe it would work? Seriously a ducttape dirty hack.
I loved that stuff. I was only 12 but my bro and brother in law had some cool setups. My brothers S10 with huge subs was nuts. Everyone should feel the sensation of that much bass in such a small cabin
This is extremely well done, Steve (and crew). A few of my colleagues have had this issue and it was tough for me to explain it to them what was happening with their systems. Now, I'm just going to refer them to this video, as it easy to understand the problem from a viewers perspective. Thanks for all the work you guys do over at GN!
12:55 That moment when you start wondering if your 1000watt PSU will be enough. I actually got mine because my 2080TI and i9-9900k were giving me these exact OCP-ish shutdowns on my 850 so I figured I'd go overkill. So much for overkill... NVidia really should be building more capacitance into the card design if this is how their going to operate. It's ridiculous to expect PSUs to handle this kind of stuff as a general market feature.
Remember that throwing a bigger PSU at the problem can end up tanking your efficiency numbers. PSUs are MUCH less efficient at lower percentage loads. Going from a 1000W to 1600W PSU on the same system could end up costing you 2-4x the idle wattage. This kind of thing is wasteful, and costly.
Strange, I ran a power OC bios (can't recall the power but it was the max available on any bios) on my 2080ti liquid cooled with a 215watt 8700k for years on a 650watt seasonic PSU with zero issues. Though I did need to upgrade as the 650 didnt handle my 3080
@@kidShibuya Just like Steve mentions, the PSU brand and individual design matters as well. A high quality seasonic 650 could probably handle high transients much better than a poorly designed 850. That's the biggest problem here -- you can't even go off your PSU's wattage rating, there's so many variables, including how well your motherboard regulated your CPU transients (VRM buffer caps could also have helped you there). This is why ultimately these huge massive transients coming out of modern GPUs are such an issue. And it's one they could moderate themselves by providing better buffering on the GPU's power delivery, or not letting their cores push transients that high in the first place. It's the randomness and multi-variable nature of this that makes it so troubling as a trend. Glad GN made this video, needs more general awareness.
Seems odd that the motherboard choice affects whether the problem shows up. Could be that some boards have extra capacitance on the 12V rail that "help out" the PSU when dealing with spikes. Or it could be that in some cases it's the motherboard that causes a shutdown if a 12V sag causes it to switch off the PSON signal.
could it possibly be the pcie slot has more throughput for voltage on some x boards? Ge it can do 110w vs the "normal" 75w. so when it hits a transient spike, the slot compensates for it.
@@Dracossaint Seems like a legit possible theory, the motherboard should never do that though the spec is there for a reason. Above all though the GPU should never need to be drawing 700 watts and if it is going to be doing that the GPU should be the part with the extra capacitance to handle those spikes without requiring other components to carry the load for their poor design.
Both the amount of capacitance, and the amount of energy the inductors can store will affect the behaviour of a motherboard or GPU card during transients.
It’s not only capacitance. With transients this big, your board needs to be designed differently. If the copper carrying power to your PCIe slot is too thin, or it’s ground is not directly underneath, the added inductance can negate capacitance on the bus.
@@bdhale34 It's not the motherboard that should not do that, it's the graphics card's duty to not draw more than in total 75W (66W from 12V rail) from the board. On the other hand many boards can provide more than 75W of power. I think since the RX480 there were no GPUs drawing more than 75W from the PCIe slot.
25 years back when I was the dude that built all my friend's PCs I found out, that the most erratic and hard to reproduce errors could be attributed to the PSU. I had my list what things to go through if ppl were having troubles but it very often came down to cheap psu that were used in pre builts. Once the pc was upgraded the errors started to mount up. This seems to serve as a good explanation of these experiences. If you buy the PSU judging only by the specification numbers you can have a bad awakening. Thanks for you hard work
Besides all that you say and are at the Core of the problem, one should consider also that the majority of issues coinceiding with the PSU was also in parallel to 2 other factors, One was poor Ground for PC chassis and the other one was also poor Grounded House Power plugs which also contributed in instability of the whole system.. that was happening in Older PCs (2012-2013 and prior) and Also the more mainstream PSUs were then of questionable Quality, more often than Today.. Ofcourse PC's weren't that demanding in Power back then .. and also Some brands of Quality existed. I remember picking up a Totally Silent PSU (Passive Cooling, must have been 450W or 500W, dont rem actually ..that was at early 2000 and the price was then 135EU which was expensive at tge time) powering along a Ati 1600X.. its still runnimg today in a Cube setup Powering a Amd FX 55 System.
@@neutechevo poor electrical fuses were also often an issue but it was easy to rule out and just use a wall plug in a room with less load on the fuse. These issues often peaked during lan sessions as too many consumers were on one fuse. I didn't say that all old psus were bad. Just the ones oems cheaped out on in their prebuilts as it drove down their costs.
@@Dodara87 reviews i guess. Its hard to come by proper reviews for psus since dirkvader de shut down however. I try to reasearch as best as I can though.
@@benjaminolry5849 yes ofcourse. I said that more mainstream PSUs back then were of more questionable quality due to lower power demands.. also lots of us have more than one, burnt hdds psus mobos..etc.. and you could not pin point the actual problem.. i vividly rem. a problem for my own system.. and then i figured it out at sbout a month away was due to 1 or 2 screws of the motherboard not tight in the chassis.. these problems back them were more common.
Yeah got that first hand when I upgraded to a 3080ti. 850w should have been plenty for the steady state, but transients were enough to send the PSU into safety shutdown. Since you asked for details Steve: EVGA 3080Ti FTW3 Ultra Gaming, ASUS MAXIMUS X Formula, 8086k Der8auer edition @5G all-core, Seasonic Prime TX 850 from 2018 couldn't hack it after upgrading from a 1080ti. The 1kw model of the same line is doing fine, absurd as that is on a system that draws 450w average while gaming according to my UPS.
@@GamersNexus i have been running a gigabyte 3080 on a rm850 that is quite old for a good while and haven't seen these issues. It's a 320w hard limited model so I've made my own curve for it where it really doesn't do the stupid push past 2040mhz at like 1060mv. Mine mostly run 2025mhz at 1000mv so I wonder how this curve affects transients.. I'm not finished with the video yet obviously but I would be interested to see transients without Nvidia boost algorithm.
This explains an issue I had with a SeaSonic Prime Titanium 1000W about a year and a half ago. Shortly after adding a Gigabyte Waterforce 3080 to a Ryzen x570/5900X system, the PSU blew out. I could not understand how a PSU I had had for 2+ years across multiple platforms would all of a sudden give up the ghost, but to SeaSonic's credit, it only took itself out and every other part was fine, but I literally had no idea it could have been the GPU until now. Yikes!
This video deserves far more recognition. The production level, the overcommitment to accuracy and honest transparency, and the attention to detail shown here are nothing short of superb. The GN team is doing amazing things for the PC industry as a whole, and manufacturers should take note.
Wow, GN - this is getting very professional. As an engineer I am very happy to see this "marriage" of quality knowledge and consumer/user friendliness. All the best to you. Keep it up :-)
20:45 And when we also account for the folding frequency, it might actually be better to build a separate setup that is offset by half a period, then combine the results into one function. That will eliminate any issues coming from the equipment generating errors due to the high resolution data processing. 22:10 - You can use mathworks MatLab for this, easily.
Moral of the story: Everyone needs 1200w PSU's, separate and isolated breakers, tesla solar roofs with power banks, and a dedicated nuclear power generating facility to accommodate next level gaming hardware.
@@LawrenceTimme I almost got one, but I hesitated until they sold out or the sale ended because they put a different plug on them. Does yours have a different plug on the cord from the wall?
I was aware of power spikes in GPUs, but until watching your video, I had no idea they were this severe. I'm getting ready to overhaul my PC and my failing old Corsair PSU needed a replacement. I've just ordered a PSU with more headroom than initially planned to prepare for newer cards and I'm sure it'll save me a lot of trouble in the future. Thank you GN. You're the only mainstream hardware reviewers that are covering topics like this with actual data. For how small this operation is, it's incredible how far ahead your data and methodology is versus much bigger hardware reviewers.
This is excellent content - thank you so much for the work you're doing here. I just had this problem. I upgraded my GPU from a 1080ti to a 2080ti, and in the process, I decided to change the case and water-cool the system, which also meant adding a lot of fans, a pump, and all of the RGB that comes with it. The build that eventually had the issue was with an EVGA Supernova 750W G3, a GigaByte Aorus Elite Z390 MB (with a seated Intel 9700k), and a Gigabyte RTX 2080ti Waterforce GPU. The PSU was about 3 years old, and the system would shut down randomly after gaming for 5/10 minutes. Opening up MSI Afterburner and reducing the power-limit on the card to 90% solved the issue at a cost of performance, so I knew I had to upgrade the PSU. I wasn't in a place where upgrading the MB/CPU/RAM made sense - I'm waiting to see what this next generation of processors look like from AMD and Intel. I went big and bought be quiet! Straight Power 11 Platinum 1200W. Needless to say, the system is stable now, even after overclocking the GPU and maxing its power consumption limits. If you're having this problem, try under-powering the gpu to see if you can stabalize the system. It's not a permanent solution, but it might allow you to game while you figure out your next steps.
I am a CNC Machinist and shop owner. In dealing with old CNC equipment I have dealt with Controller failures from blown capacitors many times. (in the circuit to handle transient power draw) Thank you for this video explaining the problem to your viewers. (something I would just naturally assume when the system crashed) :)
Awesome to see this investigated and laid out in a straightforward way. IMO it seems like most of the blame lies w/ the GPU manufacturers for not controlling the spikes better, but I agree that overbuying on PSU is the easiest way for consumers to address it. I'm likely putting together a new build late this year. I was planning on once again reusing my Corsair HX1000 that's now over14 years old, but that might not be a good idea even though its 1000W... better to keep it in a system that won't have as high spikes and continue getting reliable operation from it and go with a newer unit for my new build. I'd be interested in additional data on GN Extras!
In my experience a seasonic focus gold 1000W PSU is able to withstand a 3090 FE with a 5950x on a custom water loop. Before that one I had a Corsair RM850x that crashed during transients.
@@ahfreebird ... my computer and monitors have to be on separate breakers (cause the monitors are enough to trip a 120V 15A breaker when added to my computer).
Mainly with Nvidia GPU's. AMD is also rising a bit in power consumption, but their cards are literally hundreds 100+ watts behind Nvidias in consumption. Also, both Intel and AMD are getting to kinda rediculous power draw on CPU's as well. Intle is already there, and AMD's AM4 socket maxed at 105w TDP's, the new AM5 Socket is going up to 175!
In the world today efficiency should be a huge concern, I was happy when AMD finally started to take it seriously after there older GPU's and CPU's. It's high time we start expecting more from new generation hardware, bigger and faster doesn't mean anything if it's pulling 2x to 4x the power of previous generations. Sloppy drivers and software play there part but with Gamers Nexus on the case we can finally hope for some truth and justice to be served. Thanks Steve and you're amazing crew, thank you as well for explaining how you're equipment works instead of just saying "it knows".
there is no choice, silicon based chips are extremely close to their limit, so they will get bigger and bigger and consume more and more .... and obviously they will be lot more expensive
efficiency doesn't matter. if you're so concerned about your power draw, you can always undervolt for a marginal penalty. more power usage an emotional, but not logical feels bad. so long as it can be cooled, it doesn't matter how much power a card uses.
@@zannare There is choice, just look to Apple and the M1 chips for a perfect example. As much rightful hate as apple gets for some of their practices, the M1 chips are insanely efficient. My m1 macbook has a battery that lasts two days easy, literally doesn't have fans and still doesn't get hotter than any laptop I've ever used before even under max loads. It's definitely possible. Not easy, but there is a choice. The issue is the time and money that companies like amd and nvidia would have to invest, and there is littl incentive to do so if you can just offload the problem to the consumer by forcing them to buy bigger and bigger PSUs.
This reminds me of when reviewers added "frame pacing" and 99th percentile frame rates to their review tests for GPU display capabilities. The averages weren't good enough to capture the user experience. One thing both types of manufacturers can do to help: add specs for "max transient voltage or current" over a specified duration such as 100 microseconds. GPUs guaranteeing that they won't exceed that value, PSUs guaranteeing that they can sustain that value.
This was a fantastic analysis. Really love this level of information. I do wish there were AMD GPUs mixed in with the Nvidia graphs. Having just worked on a 6900XT build I was wondering how it compares. Thanks so much for all your team does.
Thank God someone is seriously talking about this. I remember when I thought getting the 980ti and 1080ti was a far cry from what was really possible for me. Oh how little I knew. And oh how much worse it can get, even from where we're at now. It sure feels like tripping people's breakers is the deep part of the iceberg but I wouldn't bet on that.
good lord, the music at first had me thinking this wouldn't be the greatest of your pieces, but then you dialed it in and got all graphics-y... you guys produce the best content in the field. keep up the fantastic work!
@@DairyQueen2187 damn, you got that with a 1000W PSU ? Do you have a lot of power draw from other sources ? Because if it's just the normal CPU + PSU + 1-3 SSD+HDD, the 1000W should be able to handle it no problem.
So few people out there actually understand and take the time to learn about hardware, and then really point out flaws. Not just surface-level or consumer obvious ones, but really deep dives. When you guys speak, people listen! Your channel makes a big splash. Every generation the power draw grows, the heat grows, and the heatsinks get bigger and bigger. It's not sustainable.
That last part isn't really accurate. Power consumption (for both GPUs and CPUs) has see-sawed over time, but definitely with the 40XX series we're looking at record-setting levels of power consumption. With this generation, Nvidia is trying to beat AMD by throwing efficiency to the wind and just pouring raw power through GPUs to force faster clock speeds. This video shows one of the challenges of doing that, in that if your PSU draw 450 watts normally, it might need to be able to survive transients of 1500 watts. Both the day-to-day operating costs and the cost of requiring such a robust PSU will make these GPUs very poor value propositions. Also, by taking this irresponsible approach of "MOAR POWAH!!!!11!!" and boasting about their power consumption numbers, they're just asking governments to step in and enact hard and fast regulations limiting maximum power consumption. And that likely will just make everybody unhappy.
It might not be sustainable, but it doesn't have to be. At least for mainstream consumer cards. Consider what the current high end cards can do: 4k, 120FPS, ray tracing. At least two, possibly all three. Whats left on the graphics wish list? 8k? Impractical on account of the shear size and cost. 4k VR? Already have it, just turn off the RT. More FPS? Diminishing returns past 60 already, and your not running competitive FPS like CSGO with 8k textures and ray tracing. Therefore, hello 600FPS. There are just not any technical feature left to add in, and while there are a few niche combinations of games that can max out a 3090 and have sub 60FPS, lets say the 4090 doubles the graphics power for not double the power cost. RT is the final word when it comes to graphics features (I'm saving about 2 pages of explanation with this) and that has a more or less fixed compute cost. So lets say the 4090 has enough graphics power to run anything that you can reasonably run. Someone will have to draw the line in the sand and say" Our card can do 4k VR with RT with 8k textures with 240FPS... what do you not have phones, err 8k VR, err eyes that can break physical limits?. So the 50 cards will be refining the design, dropping power, and dropping cost. Thats not accounting for the artistic/scientific types, but they are going to always be wondering how they can fit xx90 #5 into there build.
@@megachonk9440 considering how relatively small the market for high-end enthusiast graphics cards is, government regulation vis a vis power consumption seems highly unlikely.
@@sharpe3698 Given how the trends are, we might be seeing silly power consuming on low and mid tier graphic card where power consumption vs performance become a cost vs value issue.
@@nickierv13 As long as cards *can* advance, games will continue to push the boundaries of what's possible. Everyone always says that there will be a limit, but it's never been true. There will always be some level of detail, some additional effect that opens the floodgates to more performance being needed. Todays that's raytracing, tomorrow it might be in-game fluid simulations, or per-hair physics or SOMETHING. No one is ever satisfied with stopping and saying "good enough".
1: That music. Nice. 2: Back in the "old days", general wisdom was to get a power supply with a 25% higher capacity than what the system added up to. It was mostly to guard against efficiency drops, but it seems like it might be good to help with the spikes.
This is a video I feel like I've been waiting for! When I did my recent build (12900K & 3080 Ti) I went for a 1000W PSU (EVGA SuperNOVA 1000 P6) even though I could have gone with less in theory. An actually quantifiable set of data demonstrating this need is a nice kind of post-purchase confirmation to get. Thanks for this kind of work!
I went thru 4 RTX 3060ti’s before I realized this issue. I had a 600 watt power supply which is enough if you go by the manufacture standards. I now have a 850 watt and have no issues. So I think it was the transient spikes that was frying the cards
I love the depth and angle on this piece. The writing is great, and Steve is a great story-teller. He could easily do books on tape, if GN ever falters.
Would have been great to see what happens in lower-tier cards, like a 3060 or 3070. Sure, the 3060's TDP is 170W so most PSUs could take a transient as normal, but consider someone on an older 450W power supply, which technically meets Nvidia's minimum spec.
@@Metalwrath2 how would it be throttling when it never goes above 350w leaving 200w head room for spikes . also that 350 is under full load lmao gotta love 65w cpu's
They only want to produce cheap low cost boards at a high sale price, expecting them to put high quality components on is simply asking too much of them.
This was also where CAPs inside the power cables were also a possible solution 5+ years ago and I think will start coming back soon especially with such high spikes becoming more common place.
There are multiple tiers of power delivery. The GPU has local regulators to convert the high-power 12V rails into low-voltage, ultra-high current rails of somewhere around 1V, plus or minus. If the 12V rail dips due to excessive load, it is going to have to fall to less than the GPU operating voltage, plus some drop-out margin, before it causes a problem. The GPU transients are actually _already_ being serviced by the point-of-load bulk capacitance. The transients on the 12V line that remain are from current draw to replenish those caps. The 12V rail's sagging could cause problems elsewhere, but there's not a whole lot left in a PC that actually runs at 12V. It's almost entirely used as a fat current pipe. So the CPU will have its own local point-of-load regulation and the same kind of resilience as the GPU. Larger cap banks in the PSU will absorb transients better, but that actually has consequences as well. For one, they will completely bypass short-circuit protection by having the capability to supply a LARGE quantity of power in a very short time, which isn't always desirable. Also, as transient response improves, regulation actually suffers, since by nature of being able to smooth over changes in load, you're making change in supply less visible to the switching controller. Dealing in KW power is inherently risky. It's just a lot of energy, and handling that amount of energy carries some risk. Nothing in a PSU is going to blow up having to carry 2x load for 100uS. Tripping the OCP in spans of time of 1ms or so is solely to prevent cables from rapid heating, shorted contacts from welding, and things like that. The clips in this video where the fan speed dropped significantly .. that's not a 100uS spike. That's a prolonged draw. _THAT_ is just an insufficient supply, because 1-2 seconds of heavy load is no longer a "transient." It's _a load,_ and you need to size the supply appropriately. More caps aren't going to solve that problem.
@@nickwallette6201 It's going to take a whole ecosystem approach to address the issues with Transients and Transient load. When you think about it, motherboard OEMs beefing up VRMs lately at a rapid pace is starting to make a whole lot of sense imo. Anyway, I think we are going to see a revamp of the 80 plus certification and anything less than "Gold" will slowly start to fade away in one to two GPU generations if there is no revamp of the '80 plus certification' on the lower end.
Ran 3600X + 3080 Ti with 650W Seasonic Focus Plus Gold - no issues. Changed 3600X -> 5900X, 3DMark Time Spy - system shutdown. System runs okay 750W Corsair RM, but 3DMark CPU clock frequency chart shows anomalies when running 5900X PBO + overclocked 3080 Ti. 850W Seasonic Focus Plus - all is well and no issues besides PSU coilwhine when under load.
Hey Steve and GN team, what effects does Undervolting a GPU that exhibits high transients have? Undervolting already has a ton of benefits, would love to see if it can be a way to deal with transients to avoid having to replace a power supply!
@@Mpdarkguy @Cătălin Power Limit wont fix as the max power limit of rtx3090 is ~350-400w depending on the model so it never should be able to even go to 600w+ spikes but lower voltage means there is really less power going into the card and therefore should improve the spike behaviour
I was thinking this as well. The down side is that the silicon lotto plays a role in how much you can undervolt, only allowing a generalized answer at best. It would still be handy to know if that indeed helps or not.
I undervolted my 3080ti FTW3 to 875mv which was quite a bit lower than the ridiculous stock current that EVGA decided to pump into this thing, (Which was absolutely unnecessary because I stay well above this card's boost clock and stable at 875mv) and while it did reduce my peak spikes, I still hit 425+ watts during spikes. (I originally would have them in the 460s) These spikes were just during regular gaming. I can't imagine what I still might manage to hit during some kind of GPU torture test. That said, I've never had a lockup or PSU shut down of any kind. Although, I am running an EVGA 750 Watt gold SuperNova, which are pretty rock solid. So in total, I'm using an 8700k typically running at an almost 5GHz all core overclock, the 3080ti FTW3, two mechanical hard drives, three SSD drives, four 120mm Corsair LL series RGB fans, a single 120mm slim fan, a four port USB 3.0 expansion card, and about 8 USB devices total going into the PC with, typically, half of them drawing some kind of power at all times. I'm probably OK because of the 8700k. I'd imagine if I put in any newer chip with 8+ cores, I might run into some issues.
Honestly just undervolt. It's so beneficial. My FE 3070 has it's power consumption cut by two, i can do 850mv at 1860mhz + 500Mhz vram OC stable.. On forza, i go from 205-220w to 110-135/140 at worst..
I really appreciate this experimental approach, rather than just presenting existing benchmarks. Yours is the first technology channel I have $thanked.
I'm wondering if this these transient draw situations were the kind of thing that led to some of these 3090 GPUs having VRM transistors that failed under load during certain games? We already know that some games and some setting combinations create worse transients than others, so my thinking is that this could have led to a "stack-up effect" that ultimately surpassed the ratings of the VRM transistors for long enough that they just couldn't take it any more.
Buildzoid @ Actually Hardcore Overclocking talked about this one. IIRC it's because of improper phase balancing, that is, some phases are overloaded, while others arent really doing anything. The whole VRM design on the 30 series is just shit, according to him. I can't remember which vid it was, but it was a recent one.. IIRC he was fixing a 3090.
@@patrikjankovics2113 30 series in general, or that specific 30 series card he was working on? I have an Asus 3070 Ti in my computer right now, probably fine but I'm never gonna mess with it other than putting a water block on it. And I know that Asus does go kinda crazy with the VRM design on their motherboards at least, so it wouldn't surprise me if they did something similar with their Strix or Tuf line of 30xx cards (mine's a Strix). But yeah unbalanced phases does sound like a wacky problem to have, is there any particular reason they can't just parallel all the various phases together? I know "ganging transistors together" doesn't generally work like that, but I was more thinking of ganging together individual buck regulator VRM sub-units, so they'd basically all be wired in a "diode-OR" type fashion.
@@44R0Ndin The last solid top end card made by nVidia was the 1080Ti. 20 series are prone to power delivery failures and the 30's are even worse. Certain cards are extra shit (Gigabyte) but the 30's in general are unreliable. Then there are dummies who think flashing their unreliable cards with "1000W bios" is a good idea.
@@TwiztedJugallo Good thing I'm relying on 3rd-party cards from more-or-less reputable brands then? You're not gonna get me to give up my 3070ti for a 1080ti or 1660, and so long as I don't mess with the VBIOS it shouldn't have any issues. And in case you didn't realize, or I forgot to mention it earlier I don't have a 3090 or 3080Ti. I have an ASUS Strix 3070Ti, and it's been doing just fine for me. My previous 2070 did die on me after 6 months, but I traced that back to the Micron-branded GDDR6, and not a VRM failure (it even looked like memory failure on my screen, plus I found out that there was a bad batch of GDDR6 from Micron that got used in a lot of 2070 and 2080 cards which is why they were dying frequently, but that particular problem stopped happening when they switched to Samsung branded GDDR6). Now given all the problems that the 3090 has been causing, I'm glad I didn't have the money to get one, but even if I did I wouldn't have, all I do is play video games (I don't make videos, I don't even use Photoshop or anything like that), so what I have is perfectly enough for the needs I have.
Should also note that certain models of power supply are more vulnerable to transients than others. A 750W EVGA G6 might be able to handle a 3080 just fine, while a 750W EVGA GA likely will not. The Seasonic Prime-based units are also problematic in this regard, just to name a couple of the more popular lines. Either way, GN did a fantastic job of explaining transients here, and I look forward to what's in store. Also, I suggest looking at Aris' Powenetics V2 which he announced recently; could be pretty useful in testing.
@@DimkaTsv There is an awful lot more to transient response than just how big the capacitors are. Capacitance is an important part of it, but you can't know about whether a PSU can handle a 3090 just by looking at the caps.
I’m still using my “1st edition” blue GN mouse mat. Still gotta get a mod mat someday. Would be cool if there was one that was still useful with stuff like the ruler, but had more general diy diagrams and stuff.
I'm someone who has been building my own desktops for over 20 years. To save myself any potential trouble, I bought a Cooler Master's 1000W PSU way before I even got a myself an RTX 3080 with a i9-10900K. It's always good to do your math before building your own PC. And if the buyers intention from the onset is to get a top end quality system, then don't scrimp out any components, not even your motherboard.
You sound like you know what you're doing. And I'm trying to plan ahead better in the event I get a 4xxx series card. I am recent to building my own systems (two years). I have a 3090 and 5900x powered by an NZXT 850w PSU. This is my second PSU, and my MSI Meg Ace x570 had to be RMA'd, and it's working again. So, it's not been all rainbows and butterflies, but I'm learning and the system is stable now. Question: This week I picked up an Antec Signature 1300w PSU just so I had the option of getting a 4xxx series card if it works out. Good purchase or should I return and hold out for an ATX 3.0? Linus indicated PSUs would or could be hard to find in the fall?
If you were trying to use excel to deal with a large .csv file, I highly recommend you use the Pandas Python 3 library. It can handle millions of data points before it gets too bogged down. I use it at work to deal with large .csv files. Also great explanations of the DC amp clamps! I use them at work for 3 phase power analysis and know the pain of trying to keep them calibrated and accurate.
I'm an avid VR user for gaming and until late last year I had issues with games being stable and it was very hard to find out why. After all that trouble I decided to buy the best power supply I could find hoping it would have the best quality components that could handle any (my supposition here at that time) very short power peaks that might disrupt my VR experience. I invested in the Corsair AX1600i 80Plus Titanium Power Supply with a 10 year warranty. At the time my rig was an ASUS ROG Strix 390-F Gaming Motherboard, an Intel i9-9909k CPU, 64 GB RAM, two Samsung NVMe SSD for operating System and my most demanding games (had other SSD's for other games), Nvidia 1080Ti on Windows 10. I could now discount the PSU for any problems I was having. Shortly after, I upgraded to an ASUS ROG Strix GeForce RTX 3080Ti OC and this solved all my issues with VR. I upgraded again for flat screen games to a Viewsonic XG320U 32" monitor running at 144Hz in 4K. I limit in Nvidia's Control Panel all games to a maximum frame rate of 144. With this combination I can even run (all without mods) Fallout New Vegas, Fallout 4 flatscreen and VR. These games have some terrible coding that most people complain about but now I don't have any crashes or stuttering. For those playing Fallout 4, in advanced settings you must not set Ambient Occlusion to HBAO+ (Ultra) so set it to SSAO (High) and set Weapon Debris to Off (best performance). All other setting can be maxed out. I also use in all games Image Scaling On in Nvidia Control Panel and in GeForce Experience Image Scaling set Render Resolution to 85% 3264x1836 and Sharpen at 50. I am now on Windows 11 with no issues. This represents several years of trial and error and I hope someone finds it useful.
GTA V @4k max settings killed an old Gigabyte 3080-ti Vision GPU after just a few days of owning it. Died after about 1hr of gaming, fans just span with no signal. I was using a RM750 PSU, 5950x and Asus Dark Hero mobo. Upgraded to a RM1000x and refunded the GPU for a 3080-ti Hall of Fame (Which was somehow $300 cheaper) and it's been running perfect since. I wonder if I experienced something similar to what's happening in the video.
The video was excellent. The explanation of the topic was fascinating, B-roll kept the flow of the video up & the title really drove the click. Good work GN, more like this with your focused videos.
This was an excellent video! I haven't had a problem with it myself, but I've heard of transients and how problematic it has been for the RTX 30 series. I had no idea why and how it happened though, and this was the most concise explanation I've found! I feel like I finally understand it - so thank you, GN! Awesome work on this video, all the research you did, and the way you presented it. It's really good info to know, and I'm really glad to hear you'll be addressing it more from now on.
You make a very good point around the 25-26 minute mark about over-buying the PSU. I suspect these spikes are what often kill pre-builds PCs that use the cheapest PSU that barely meets, and sometimes don't even meet, the GPU power recommendations. I disagree with that way of thinking though. In my system I always over-buy the PSU. For an extra 30$ you'll get an extra 100+watts. The nice thing about having the extra head room is future proofing for upgrades BUT if you aren't pushing it to its limit the PSU generally runs cooler making it last longer. So do you want a 750 that will last 5 years or an 850 or 950 that will last 10 years and keep running through those spikes? Also they run quieter when they're not being pushed to their limits. Is it wasteful and more expensive? Yes, technically. Is it worth it for all the benefits? I would argue yes.
As someone who worked with UPS and power bank installation and maintenance for a living, I'll let you know that, without getting too technical (at least in the main channel lol), you just raised the awareness on transients by quite a few Farads ;-)
I was concerned that this channel was going to become less entertaining as it became more invested into technical benchmarking equipment and data. From this video I think I see your vision now, and it's brilliant. Important work being done here. 👏
Thank you guys for putting this out. I actually encountered this awhile ago on my RX 5700 XT - random shutdowns on a 650W ATX PSU. One thing to note here is that the PCI-E slot is capable of supplying up to 75W according to spec, and I would imagine that some motherboard manufacturers have incorporated additional capacitors for handling sharp transients. I tested it as extensively as I could (given that I didn’t have an oscilloscope and had to rely on the built-in Radeon monitoring software to track power consumption), and was able to determine that the base load was higher than a single PCI-E power cable could reliably handle (as, according to spec, each line only has to be able to supply 150W of power). One cable plus the power from the board gives you 225W of power according to spec, and I noticed that the board was routinely topping 250-300W of power consumption as a baseline-so it was no wonder that until I started using a separate PCI-E power connector directly to the PSU (rather than daisy chaining the 8+6), I was seeing OCP shutoffs. And even once I corrected that issue and started using two separate PCI-E power cables, I was STILL getting the occasional OCP shutoff. The fact that boards can have such insane spikes is really frustrating as a consumer, though, because it realistically means that PSUs are being taxed far more aggressively by GPUs than had been previously reported, and that will cause them to age more quickly. That same 5700 XT actually ran a 750W SFX PSU from be quiet! into the ground-power supply worked well for a year, and then one day it just started triggering OCP on just about every boot. I shudder to think what it’s doing to the Lian Li I replaced it with, or if I should consider investing in a beefier SFX PSU for my rig (or else abandoning SFX altogether until this trend dies down).
This is interesting. Which 650W PSU and GPU custom do you use? I haven't had single shutdown in the ~2 years use of my Sapphire RX 5700 XT Nitro+ with a 550W be quiet Straight Power 11, but that is also a high quality PSU, not a cheap unit. I'm running it together with an R7 3700X or, for the first year, with a 2700X and both at stock. I wouldn't have thought that this could happen with a GPU of this caliber and a good PSU. I always had both power connectors on the GPU connected and I never used a cable splitter, if that's what you're referring to with daisy-chaining.
@@dennisjungbauer4467 It was an EVGA 650W ATX PSU, I forget the exact model, but it was also more likely an issue with the GPU itself. I purchased the Gigabyte RX 5700 XT Gaming OC pretty early in the product’s lifespan, and had to deal with a lot of growing pains while they refined the firmware and the drivers. This board in particular seemed to be plagued by uncomfortable power spikes, as a few reviews I read and a few forum posts I found at the time suggested that it was a problem. The card is currently *mostly* okay on a Corsair SF600 (order history confirms it was not a Lian Li PSU I purchased), but it nuked the be quiet! BN639 SFX L 600W PSU that I had been using it for in my ITX build after only a year.
@@joshcurtis3752 Ah, right, I almost forgot about all the issues this card had for some people. I guess I was lucky in that regard as I haven't had any hardware issue, only few driver issues switching (well, fully new PC) from an Nvidia card (GTX 970). And while I didn't buy it very late, I bought it at end of 2019, so ~5 months after release.
2 ปีที่แล้ว +6
Awesome, as always! Thank you so much Steve, your team does an amazing work! I can not find the right words to describe how valuable what you, guys, do in this channel! Thank you soooo much!!!
The research you guys put into all this and continuously improving the quality of content is just mind blowing. This is exactly what is called giving back the community what it deserves who helps you grow as a channel.. Hats off
I've said it before and I'll say it again; I think Gamers Nexus is the true heir apparent to the legacy of The Tech Report. This type of thing is exactly the type of work we expected from TR back in the day with Scott, Geoff, Cyril, and co. If GN can figure out a way to make this kind of GPU power capture easy to accomplish I think it will have a similar effect to Scott's work with bringing frame time to the benchmarking game back in the day.
It would be great if GN could put together a list of: - PSUs that handle transient spikes well - motherboards that handle transient spikes well Lots of testing, but, it might put the hot poker up the butts of the manufacturers.
who knows, this might be a thing more in the scope for ltt labs. generating databases for common issues like this is something that i would love to see out of that whole project
A point I'd be interested in knowing: how do overclocking and undervolting play into all this? is it just a reduction of power/an increase of power type effect? or are there potentially knock on effects, making overclocked cards more/less susceptible than expected?
It goes both ways if OC'd, you already are pushing the TDP... (higher Volt., higher consumption) that means the transient will be far bigger... And the other way around...
My takeway from this video is that a transient spike is based on the average draw. Increase average draw, and you'll get higher numbers in your spike. So if you OC your card with a higher power limit (say an EVGA 3080 with a 450W XOC bios) you'll get a much higher spike than if you run the same card undervolted, running around 320W or so (this is about what my 3080 FTW3 pulls undervolted according to msi afterburner). So overclocking definitely increases the strain on your psu as the total wattage number in the spike will be much higher.
By the way, who did the animation showing how the clamp worked? That's awesome stuff, right there. _Such_ a nice touch for the channel to help explain - _with_ a high-quality animation - how the testing solution works in an intuitive way.
Wasn't there an aftermarket overclocking GPU that came with an optional capacitor module you could stick directly into the backside of the GPU die? If memory serves, it was sometime in the DX11 era. I feel like that was trying to solve this very issue for extreme overclocking.
@@lgpa565 I remember now; it was the MSI Lightning line of extreme OC cards, starting with the HD 7970 and continuing with the GTX 780. They called it the "gpu reactor".
Transits are a problem known for a good while. I suspect my old board was tripping because the caps were getting old enough to no longer deal with power fluctuations.
I'd really like to see you test the transient handling performance of the Corsair SF750. That's the go-to PSU for SFF and routinely handles top end components. I'd expect it to pass the tests that the NZXT unit failed.
I would like to see such a test too since i am running my Asus 3080 TUF with an Corsair SF600... so far no problems but perhaps its because of the slight undervolt im running (although i didn't in the beginning) and the games i play aren't all too demanding.
yeah it's strange, i have the spikes issue on my 3080 but only happens on Battlefield 2042, Sea of thieves and Battlefront 2, it didn't happen playing cyberpunk at max settings with rt and all, weird lol
I made myself a borderline on TOTAL system power draw under heavy load, and GPU power draw. It is 500 W total and 250 W on a GPU and I am not interested in anything exceeding that no matter what. It leaves me only the best perf/wat options on the table, and considering my needs that is all I want. Currently sittning on a 5600X and 6600 XT, perfectly happy with 1440p resolution.
Gamers Nexus does their homework! This is what makes them so great! Things like when they showed the correct direction of a magnetic field with respect to the direction of current in a wire, I literally clapped out loud!
Support our in-depth testing, journalism, and analysis! Grab a Modmat, mouse mat, shirt, or more at store.gamersnexus.net/
Watch our PSU Reviews, Deep-Dives, & Explainers Playlist! th-cam.com/video/yDX_1PWUWdw/w-d-xo.html
Love how the intro of the video is edited like a Dateline NBC report.
THANK you for this video! We've had customers GPU's returned to us because they exhibit that exact behavior, tested (Corsair AX 1600's), found to be working flawlessly sent back and forth a couple more times before we've figured out that their PSU is (although above "spec") not handling the loads some of these monsters pull under spikes.. GPU's should have transient spike recommendations included. Easy as that. Bad marketing for Nvidia/AMD? Ok, we (small retailer) do not give a f*ck, current system is misleading. AND! A lot of customers don't enjoy getting "up sold" on PSU's.
Great video Steve and his team!!! Glad you guys made this video!
I am a Mechanic (40+yrs), and I have a general rule. Your battery should be 2 CCA (cold cranking amps) per cubic inch of engine displacement. I follow the same rule when selecting power supplies. I make sure the PSU can handle near double what the system can draw. Your PSU will be most efficient at around 50%, it will also run cooler and quieter, and it will last a lot longer.
quick correction: the magnetic field is what creates the charge, not the other way around.
Edit: I would like to see what effect undervolting has on these spikes if any at all. I know you can get some interesting behavior out of a GPU by undervolting a little too much.. but doing it just right seems to make it more stable since clocks are not constantly shifting.
Electrical Engineer here, I have to say, what a superb animation of how current clamps function. This piece of journalism is outstanding, you guys rock.
Thank you! We'll let Andrew know you liked it!
scrolled down to say the same thing, I've ALWAYS wondered just how those clamps work their magic! sick job GN crew
@@GamersNexus Industrial automation electrician here, and I wanted to say the same thing! You might be inspiring the next generation of industrial controls tradesmen and women! lol!
I'm a computer engineering student, and I wish I had seen that current clamp animation for my lab courses - it really is fantastic!
@@GamersNexus Andrew really killed it on that animation, his animations really add a whole new level of explanation to the videos over the last year or so.
The level of quality content your team is putting together is down right impressive. You guys keep this space in check and I truly respect what you all are doing.
Thank you!
@@GamersNexus no thank you
And completely relevant. A lot of trust in this community
Honestly.. what a great fucking video- truly.
So well said...i absolutely love these types of videos from GN. In-depth and very news / informative type for consumers...
I find it kind of interesting that this is just now being discussed for computers. I did car audio installation professionally for over 14 years and this was something we dealt with on a daily basis for huge car stereo installs. In fact it is still something that is dealt with today for multi amp power hungry car audio systems. We had to install large capacity capacitors in the system to "suck up" high transient power demands to keep from potentially hurting a cars alternator. You can visually see what these high capacity caps would do because the headlights wouldn't dim, or not as much, under large bass drops which would draw the most amount of power. Large caps are designed to quickly deliver large amounts of current very quickly while a cars alternator can provide large amounts of current but just over a longer period of time. A computer power supply is roughly the same. It produces a fixed voltage with a variable current output, very similar to a cars alternator, and neither is extremely suited for potentially a massive current draw in relatively short amount of time. For cars it has always been you upgrade to a larger alternator, or more than one if needed, you add several farad caps or both sometimes. It is getting to the point where either manufacturers are going to have to start recommending more accurate power supply requirements to account for this issue or someone is going to have to design a power bank of large caps that go in between the power hungry graphics cards and the power supplies.
Electric amplifiers have had this issue for ages, and the solution is a big capacitor.
I was just going to make this correlation - RMS vs PEAK and what not but you nailed it.
@@StephenGillie The best solution is an elegant one. Graphics cards should implement their power limit properly. It seems like really messy design to me that you can run a load of never above 200W, yet for less than a millisecond it might spike to more than double that. This can be avoided with sane/smart design. But of course the GPU industry is the epitome of superficial power craze, so we cannot expect them to sacrifice any potential in order to avoid having to oversize your power supply.
The solution to transient spikes is not just huge capacitors, at least not on the power supply itself. The power supplies (switching DC/DC converters) have a maximum amount of output capacitance that they can deal with before they go out of stability and start oscillating (which will burn out components). A power supply manufacturer can't know how big the spikes will be for every GPU that you put in your system. That's where the board designers of the GPU come in, they should make sure that they have enough filtering on their power supply input so that they don't trip the OCP. They are the ones that should place the large capacitors on their board. But ideally you don't want to have these spikes at all, and you also place inductors in the line to suppress these huge spikes, so just a lot of caps alone could be detrimental to system stability.
Yup, back in the 90's I had a cap the size of a Fosters can alongside my 1000W Phoenix Gold amp. Ahhh, the good ol' days...
"We adjusted for the latency in the hall effect sensors........". You guys are truly an amazing group.
@Lurch7861 Lurch7861, My sincere apology for making a comment that upset you so much that you felt compelled to call me names.
@Lurch7861 ok buddy you sure are very smart
@Lurch7861 No, measurement latency is not a college physics thing. Heck, in my CPE labs, which were the same ones the EEs took, we never bothered to correct for it. Quite frankly unless you're looking for something extremely specific, it's not worth the pain.
Wall power is atrociously noisy as is, and on the DC side we aren't doing weeks of data analysis. We're fixing the problem! Heck, most of the O-scopes I used were CRT. There was no data export. Even the digital scopes I think the most I ever did was export screenshots. And my senior design project was on power monitoring...
@Lurch7861 : It's not a matter of having the education to know that you need to do something, it's about having the time and money to spend on doing it.
@Lurch7861 My background was hard science and I had to take more than a few college physics classes. Never was taught this…
This is an electrical engineering concept.
Now PSU manufacturers or brands can capitalise on this issue by advertising how massive their Caps are in their unit.
Nominal Load : 850W
Transient Load : 1700W@20us
At a bare minimum, ATX 3.0 + PCIe 5.0 certification of PSUs (with independent verification) will go a long way to educate consumers here. SFX PSUs may perhaps give way to SFX-L..
@@ikjadoon Nah, there's too many people willing to spend hundreds more on a part to save a cm of space. I'm sure Corsair will find a way to make their SF series well over 1000W, if it means they can charge a boatload for them compared to normal ATX psus
Im hoping they also start to put more capacitance on the GPU board. Loading up the caps in the PSU destabilises the 12V rail which will reduce the stability of the whole system, not just the GPU
The good news is capacitor technology is advancing relatively quickly, so it will at least be possible to keep up
@@toby1248 why would loading up capacitors destabilize anything? Aren't capacitors very slow to charge compared to their discharge speed?
@@LtdJorge all capacitors charge and discharge at the same speed, somewhere between a millisecond and a nanosecond for a full charge. But the voltage drops linearly as they discharge, so the more you load them up the more the 12 volt rail will drop below 12V.
This channel is much more than a “gamers nexus”. Thank you so much for your scientific, transparent and objective media reports. What you’re doing is invaluable.
🎯
Seriously, I honestly feel like his the name has been outgrown by what he is doing now. This is serious pc builder news. Although it all does tie back to gaming in the end I suppose to keep his scope in check.
You have 69 upvotes, and while I'd like to be one of them, I'll take one for the team.
Gay.
I would dare to say GN is a bit like what savagegeese is to car reviews.
So if you like honest people applying no-bullshit non-sellout scrutiny and are even remotely interested in cars, check him out. (He also has livestreams where he mentions how hard it is to do when you are not being the industry's monkey.)
Please - If you want to up your game on this, go look up the Middlebrook stability criteria and start reviewing power supplies by their impedance and crossover frequencies.
You're great at this. I think you're one of the only reviewers that might care enough to actually go do it, educate your viewers on why it matters, all while keeping it entertaining.
Middlebrook is the first step basic method for determining stability and impedance matching between power supplies (like the main PSU and one on a board).
FYI, I'm a professional in this field (not specifically computers), but I did work on a motherboard power supply for the first Sandy Bridge early in my career.
You covered it a bit, but everyone thinks capacitors are holdup. They're more for transient suppression and impedance. Those tiny caps on a 1V 100A rail aren't going to hold up Jack. The big caps on a power supply rail are better, but they'll only respond between roughly 20 Hz and a few kHz at most. Your small form factor supplies are pushing it a bit higher (50 Hz or so). What's more important is that higher wattage supplies is pushing the base impedance down and the crossover frequency doesn't matter as much; so stability is easier to guarantee.
The 30x Nvidia series die has multiple full to almost zero load steps with durations of a few microseconds within a single frame. You can literally see a semi consistent pattern at the scene's frame rate if you tap between the on-board supply and the GPU. The on-board supply can take care of most of the transients in the kHz+ range, but those frame-harmonic frequencies (approx frame rate to a few multiplies) are smack dab in the middle of the bandwidth between the cards power supply rail and the main system supply and you have control bandwidth and impedance matching issues between the two.
Ironically, power saving efforts for high performance cards in the future will make this worse. You want it to blast out a frame, then sip juice. It's not practical to store enough energy on a board to smooth it out.
Edit: Also, you mentioned something about the phase shift between the current sensor and the voltage. That phase shift may be 'real'. I'd suggest calibrating it out on your probes measuring against a linear load. If you put your scope in 50 ohm impedance mode you can probably cal the delay between voltage and current while measuring the current into the voltage probe itself, just using the cal rail on the scope. If your current sensor resolution is too low you'll need a signal generator to do that.
I really hope that Steve and the team find your comment and read it. GN strikes a great balance between in-depth reviews while keeping it accessible for most people, so I'm sure they would be able to incorporate all this into their videos.
upvote this above comment
If TH-cam would have a Reward Comment system, I would reward you. This comment is gold.
Someone get this comment pinned, get Steve and the team on it!
Could you elaborate on the small form factor PSU's possible impact on being able to handle transient spikes in regards to the frequency you mentioned?
Graphics cards need to have enough capacitance to cover the transient spikes they generate and offer a smooth enough power draw that at no point does the card go beyond the publicly stated max power draw. If they have a device that does not or can not do that then it is out of spec and they need to publicly acknowledge this phenomenon and publicly report the max transient power spikes so pc builders can ensure they pair a power supply that can supply the demand. Anything less is not acceptable. This is clearly a design problem on the graphics card manufacturer's side. How many consumers have dead power supplies only because graphics manufacturers incorrectly designed a card and did not report the true max power draw? Repeatedly causing customer power supplies to repeatedly trip OCP and the customer not knowing what it could possibly be given the power budget allotted to the components and the power supply capability.
This is exactly it!
I would think so too.
This was my thinking since these "mystery black-outs" started and still is. The power supply might be the better place for a huge capacitor, though. But it seem like it doesn't actually need to be that huge.
i came here to say something like this and this is perfectly summed up. it is 100% upto the card manufacturers to disclose the max power draw even if it is "transient" because its still the card that is causing the problem. its a massive copout for such a huge company to blame another company "because the psu shuts off" nah my dudes. make your cards within spec or disclose the maximum draw instead of pointing fingers. we have never had problems like this in the past untill nvidia started making these cards that seem to double power usage every generation. as a pc builder it would be super embarrassing to complete a build and have the customer say that there are these kinds of problems. reasons like this is exactly why i prefer radeon cards at the moment. they require much less power for similar performance in most settings.
I think part of this problem is software in that today's GPUs are more aggressive in ramping up and down for performance reasons.
Just want to also say that the animated depictions were both really helpful and well done. It’s easy to take those for granted, but there’s probably a lot of effort that goes into those, too!
I came to the comments to say exactly this. At 21:05 they didn't have to go through the methodology and explain how it works, but they did. Not only that, they put an animation together to show how the instruments measure the EMF to come up with their readings. That's not a small deal. That takes a lot of extra effort and I said wow, that's really well done, out loud when that came up.
I noticed and was going to say the same thing. Really high quality animations and a fantastic explanation of how the tool worked. Kudos to the GN team.
This is the most well timed video I have ever experienced. I was literally just telling a friend that his crashes might be from these "power spikes," and he was not understanding what I was trying to say. I'm hoping this video with all your graphs and analysis will make more sense to him. Thanks for the breakdown!
What sucks is a 1.2KW or higher PSU would more likely than not solve the problem completely but that is one hell of a cost to pay for system stability that should be a given if the GPU didn't go overboard on the draw randomly. Locking the FPS would also likely make it not happen and in the majority of games that wouldn't even be something that hurts the experience, but in several titles you want the highest framerates possible so you see the most up to date frames at all times to be competitive and in those titles a new PSU seems to be the only option.
Hmm, I do have random system restarts, now that I got a 3060 ti with a 550W PurePower11, but they happened only when not under any load, and changing the PCI-e setting from "auto" to 3.0 seems to have fixed them. Or at least I haven't had any restars since yesterday morning, when I did that. Although they only happened once every 5 if not 10 hours, so can't say for sure yet.
@@bdhale34 This only affects some GPUs and systems I reckon though. Because I'm on an 850w PSU with a card that draws up to 430w constant power and I've never had any sort of shutdown or instability in nearly a year and a half.
I view the issue as sort of like a hypothetical car that locks up when the RPM reaches a certain threshold because it requires more gas than the fuel pump can push out so the fuel pump cuts off for safety.
Even if the car OEM picks a fuel pump that can supply enough gas for the average RPM, it won't be good enough for when the driver has to floor it.
However, just because you pick a fuel pump with the capacity does not mean it can push enough fuel fast enough, so you could have a car that doesn't shut off but does essentially run on fumes when flooring it.
@@zeratulrus142 I had to change that setting to avoid issues where my PC wouldn't have video when starting up.
This needs to become a part of GPU reviews from now on.
yeah, if the normal power draw of a 3090 ti is as high as it is without even factoring in the transients, then I can't imagine what the 4090 ti will be like.
@@plasmaoctopus1728 Especially if the 4090 Ti is shunt modded 😅😶...
From an electronic point of view it makes much more sense to solve the problem on the GPU side where the load characteristics are known and you can even actually actively control the spikes.
This can happen at any device inside PC, so its better to be solved at power supply.
@@TheSkace The correct solution is that GPU manufacturers stop lying. If GPU manufacturer says "290 W" but oscilloscope shows 620 W, then GPU manufacturer is lying about power usage.
When all manufacturers tell the real power usage, getting the correct PSU is easy.
With the current lying trend, you cannot know tomorrow for sure if you need to get a PSU with 2x, 3x, 4x or even more power compared to the number that GPU manufacturer prints on the box.
If GPU manufacturer doesn't want to print 620 W in their spec sheet, they need to add enough capacitors to get the power need closer to average.
The gpu power delivery system is relying so much on the PSU nowadays it means the GPU CAN BE MADE CHEAPER because less components
@@urbansnipe That's absolutely true that building a GPU that has power spikes is cheaper. However, when GPU manufacturers lie about the power requirements of those cheaper GPUs, selecting a PSU is turned into gambling. When a single manufacturer says "290 W" do you really need 750 W, 900 W or 1200 W for the GPU? How much could any specific manufacturer lie?
I'm absolutely fine with whatever spikes any GPU has as long as the manufacturer *honestly* declares the needed power levels instead of average power usage which doesn't include actual power required for the spikes.
@@MikkoRantalainen true i feel the same way
It’s great to see more people getting in on the whole transient issue. First encountered it when Igor‘s lab made it the focus of their power draw and supply tests but spikes and transients will only get more important to look at the higher power draw we see and the more companies rely on boost clock algorithms and sub millisecond control algorithms for power limits and current draws
Speaking of issues... there will be fires in some older houses or houses with often sub-par repairs like in hurricane zones. It's pretty messed up when you need an electrician to do a safety check before you plug in your new GPU, will Nvidia give 4090 buyers a voucher for inspection and testing?
@@Mr.Morden as far as any business is concerned the quality of the wiring of your house is a you problem, this includes the company that supplies your electricity
@@Mr.Morden Not to forget the 110V Powergrid in the US isnt even that great... For me its always funny when people on reddit tell me that i NEED a UPS and iam just "Mate iam german we have 240V and unless you live somewhere far outside and you are the last person on the line and its not burried in the earth or you do super important work you usually dont use them here"
@@mrn234 quality of powergrid in the US is largely down to the state that you reside in. I've never had issues with power in any location I've lived in, and I've lived in the middle of nowhere as well as in the middle of a city with a population of 9 million people.
It's the way how these companies "cheat" better performance, just let the GPU spike for milliseconds so the average performance is higher. Spikes so minor but also too severe to some PSUs, if not bought mindfully. But how should a normal person know about it, to avoid the problems some 3090s had? That's why i like GN so much and people like Igor but also Hardware Unboxed.
If transient power spikes are going to be accounted for in future component reviews, I am INCREDIBLY grateful. You folks rock, GN!
GN is straight up making video essays with all of their expertise about tech put into it. This is so good. Props to the whole team. Keep holding the tech giants accountable.
2017: GPUs are causing massive worldwide power draw.
2018: Countries introduce laws against GPU crypto mining.
2019: GPU prices 5x higher than the systems they're built for.
2021: GPUs affected by worldwide component shortage.
2022: GPUs are frying entire systems due to unprecedented / sudden power consumption.
Man, GPUs are like the bad kid in the neighborhood.
But us consumers NEED the overkill for gaming and streaming so we can compare our e-peen.
@@mosando No you don't.
Never had this problem I bought AMD. :p
@@Life_Is_A... have you heard of sarcasm?
@@Life_Is_A... yes we do
i was not expecting the editing to be so good on this video. not to say your regular videos are edited badly, but this is above and beyond with all the animations and stuff. well done, and great information in here
Damn that intro is a different vibe. Good to see discussion on this topic!
the "looming doom documentary" soundtrack really hits hard.
This honestly explains so much about my early experience with my 3090. I saw on nvidias website they recommended 750W for a 3090 with a 10900k so I decided to get a 850W PSU to give myself a little headroom. It didn't matter at all as I was constantly black screening cutting out during gaming which I later learned was OCP triggering. It was super frustrating as 1/3 of the people I talked to said I didn't have enough watts, 1/3 said I was borderline and 1/3 said I had more than enough. Long story short is I had to move up to 1000W to stabilize the build and now I believe the reason why is because of the transient spikes you guys are talking about. I really appreciate your video highlighting these issues as I thought I was crazy for a long time in regards to this.
I'm rocking a Corsair RM850x for a 3090 FE and 12900K non-overclocked. I'm scared to even touch clock settings with these kinds of spikes. Probably going to have to upgrade the PSU at some point. My first gaming setup in 2012 was on a Corsair 600 watt PSU. It's crazy to think how ridiculous things have gotten the past several years.
@@braf7349 my wife's RM850x finally started showing it's age after about 4 years with a 2080Ti. It started black screening in games that were only moderately demanding. Monitoring rail voltages in HWInfo, I could see random 0.3-0.5v drops on the 12v rail, even without any real load. I didn't feel like having to deal with the whole warranty process, so she inherited my old EVGA G2 850w (5950X/3090 setup) and it's still going strong for her.
Good luck with your setup! At least a PSU is a pretty easy upgrade if you do have to swap it out.
The 10900k chugs around 200-250W on turbo, so a 350W 3090 with spikes at 700W might indeed trip a touchy psu. (Especially if you have a simultaneous spike on the cpu)
Replaced my 1080 ti with a 3080 ti, had a 750w psu which should have been enough but it wasn't, system ran stable stock when I went with a 1000w psu, likely have some headroom now but I prefer no crashes vs a few more fps
Poor quality or defective PSU? Also undervolting the card saves so many watts.
Had a AIO Vega 64 card on a x370 MB and a Rosewill 1350 PSU . Started having this issue occasionally in 2020. Started being a more consistent problem in 2021 where I could reliably shut down my system by loading a stage in Destiny. I was beside myself trying to troubleshoot, crossing my fingers with each AMD driver update. When GPU prices came down from ridiculous to just gut punch levels in 2021 I desperation bought a AIO 6900X card and a B550 board, assuming it was my card struggling to keep up with newer games as I play in 4K. The problem seemed better at first, but I about had a fit of rage when it happened again for the first time. I would play for several more months clenching my chair from the anxiety of knowing my next crash could be at any moment, and I could almost never consistently reproduce the problem. It wasn't until I saw you mention this while you were researching this video that I finally had my "AHAA, this explains everything" moment. I switched PSU to a Seasonic 850, lower power but much higher quality. After the change it's been smooth sailing since. Huge thank you to GN, I had just thought "I definitely have enough power", and had never considered the OCP angle. Will put a lot more effort into choosing power supplies from now on, and you're spot on in how frustrating it is knowing that each of your components, PSU, graphics, and MB can factor in to how this problem manifests. The average consumer could never get enough info to make decisions on parts in regards to this issue and even most independent reviews would struggle to find the time, resources, and motivation to delve into this topic. Great on you for leading the charge, and this level of production is immensely impressive for an independent group. This channel has become a necessity for being the consumers voice of reason to the manufacturers and developers. Unfortunately for me I noticed your work here too late to save my wallet, but now I know better and am a full subscriber and follower.Thanks!
Yes! Perfect comment :)
I had this exact same problem. After waiting for hours in line to get into Microcenter when the rtx3070 launched, the card I brought home wouldn’t work on any of my systems, and kept tripping the power supplies internal wattage limit. I tried for over a week straight to troubleshoot the issue but unfortunately had to return the card I worked so hard to get in the first place. So happy someone is finally looking into this, and that I wasn’t just crazy. I hope this means the responsible parties will take a look at this problem and work to address it, as it is just absolutely infuriating to deal with when it comes up.
Whats your PSU?
@@Whatthevortez ^ this
You gotta drop the 550watt PSU guy ,I find 750 minimum requirement
@@badcommentbot8349 it shouldn't need to be the minimum, the 2080 ti (for example) could run in systems with 650w and it was perfectly fine
@@chitorunya yeah but this isn't 2019 anymore.
So, a bit of a tipoff about testing for these transient spikes, as I recently had a problem with them myself that resulted in me just buying on on-sale EVGA 1300W so it can just soak up the problem, because thats far cheaper on time than any other solution.
The tipoff here is the BEST way I found of tripping it, happened literally within a few seconds every time, was fallout 4 VR. In general I was able to benchmark the system as I wished and never had even one shutdown during any benchmarks, however I had one shutdown while loading GTA online (weirdly), one shutdown in Phasmophobia VR, and a total of atleast 30 shutdowns in fallout 4 VR before I even got through character creation. I'd highly recommend it as the BEST way to see if a spike can take out your system power.
My System contains a 5950x, an MSI 3080ti, (64GB 3600MHz RAM) and was running on a Coolermaster 850W MWE V2, if anyone wants to know
3080ti isnt that far off from an 3090 ... afaik a 3090 can have spikes at almost 600w, worst case for a 5950x would be like 250w and for stuff like ssd, fans, lights etc ill give them ~50 to 75w ... espacially if you look at newer boards ... then i'll take 20% above it and this results in my psu capacity i want to buy for a rig.
Can you share the framerate you were hitting in Fallout 4 VR?
I feel like this has happened to me. I can't get my system to reboot without pulling the CMOS battery.
I had a similar problem like a decade ago with a 570GTX or whatever it was. Every time I looked at water in World of Warcraft the computer would shut down within seconds. Took a while to isolate the behavior down to that, but once I put realized that was happening I could reproduce it every time. Put a bigger PSU in the system and the problem went away. Transients are fun.
Bigger picture though, all PC components are ballooning in power consumption. Dennard scaling was already dying off for years, but it's really hit a wall in the past couple of years. A significant source of extra performance is now coming from simply letting components draw more power instead of getting more performance from a similar amount of power like we saw in years past. This is only going to get worse from all vendors as time goes on.
@@ngc4486diane People don't often understand, and there's a lot of bad advice out there. Pre-builts too often have substandard components, so I'm not surprised at all.
Possibly we need some sort of regulation for power spikes like this , similar to EMC regulations ? A PSU should be expected to be able to handle spikes yes ( due to things like inrush) but not to this level , not in the consumer space.
Most consumer PSU's are generally built to the ATX specifications for power supplies. That spec doesn't really address any specific requirements for transients. However, it does say that DC output voltage must remain within a certain range under ANY load conditions that are within the specified output ratings of the supply. So, if your supply says it can do 40A on a +12v rail, the ATX spec doesn't care if that is a nice steady 40A or if it switches between 40A and 1A every microsecond. It expects your regulation to be within spec no matter how crazy the load cycle is. This is unrealistic, but the spec is silent on transients. Transient requirements for the rails could be added to the ATX spec, but failing that, GPU and PSU manufacturers would have to publish their own tolerances for transients and we would have to compare them ourselves. That or NVidia and AMD are going to have to start their own certification program and publish their own specs so PSU manufacturers can put a "NVidia 4080 Certified" or "AMD 7800 Approved" sticker on the box to help tell consumers what to expect.
This is just another indication of just how far out of date ATX has become as a set of system building specs. ATX needs a major overhaul, but Intel isn't in a position to tell everyone what they have to do anymore.
The new ATX 3.0 standard will indeed require supporting 200% rated output for 100us. This should help quite a bit.
Just buy bigger psu , three times the power draw of pc the it will be enough .
@@kaiwenwu8148 that's not a bad rule of thumb. Maybe 2x instead of 3x, but at least 2x to be safe. Power supplies are typically in their peak efficiency around 50% loading anyway.
@@noname-gp6hk if account for 120% to 130% ocp trigger point of normal PSU, 2x of power rating should be ok.
We need another episode, a sequel to this one, focusing on daisy-chained cables and multi-rail vs single rail configurations and how they relate to stability under transient loads
I am an electrical engineer, sometimes dealing with transient issues in an industrial environment where we work in kW and MW. I am also a keen audio enthusiast and know for example that a power amplifier is very dependent on its power supply's ability to deliver power at higher frequencies. I thus understand your review and kudos for pointing these issues out to prospective buyers. What the industry needs is a rating system for PC power supplies, in their ability to handle short term overloads. Premium manufacturers like Corsair can then use this rating system to distinguish their power supplies from the vanilla variety.
The issue with a rating system such as the 80+ certification that already exist for PSUs is that it becomes vague and eventually turns into a marketing label overtime. Unless there is a federal standards I think we will just be seeing more of the same.
I’m more on the side of putting bulk caps directly on the GPUs to help deal with the load personally.
@@joshuabenitez3260 I think ATX 3 is intended to address this.
Isn't it what these multiple agencies like the Underwriters Laboratories (UL) in the U.S. are here for to abide by their ratings?
I would side with @thelol1759 and suggest that the problem needs to be solved on the GPU side. The PSU needs to have solid bandwidth but there's usually a foot plus of cable between the PSU and the GPU. In the world of sudden, extreme current loads, the rail could be collapsing on the GPU side before the PSU has any chance of responding.
CPUs have had spike current demands in excess of 100A for a long time but their VRM circuits are spec'd to handle those transients and not pass them on to the PSU so severely.
GPU VRM circuits, and particularly their bulk capacitance is just not up to snuff if they can bring down a 12V rail that's basically all theirs.
The upcoming ATX 3.0 specification, which was finalized in April, is specifically supposed to address this issue. So, I'm really surprised there was no mention of it here, at all. I've been eagerly waiting for some kind of news from Gamers Nexus, about when ATX 3.0 power supplies might be expected to start rolling out, or if the RTX 4000 series cards will have the new 12+4 GPU power connector on them, which is part of the ATX 3.0 spec. Please ask one of the power supply manufacturers, about an ETA on ATX 3.0 PSUs. I'm very eager to find out, since I'm planning a Ryzen 7000 system.
They'll surely have the 12+4 connectors on the FE/reference cards, but are unlikely to require it of the AIBs.
The big question is what the PSU requirements will be. I'd assume they'd say something like '4090 requires a 1000W PSU, or an 800W ATX 3.0 PSU'. Also, as you noted, its unclear if the ATX 3.0 PSUs will be ready for an October 40 series launch. It would be really dumb to buy at launch if you have to upgrade to an ATX 2.x PSU that's becoming obsolete.
Looks like you are expecting people go out to change their PSU with ATX 3.0 spec. Tell it to those people with no knowledge, or those with 1000W+ Platinum rating.
I doubt those with some years old 750W Gold PSU or above will do that so this info is very useful for the years to come.
This guy asking the smart questions. Where's atx 3
@@JorgeAlarconL its very relevant info... Whether you are too ignorant or cheap to buy into the new standard is not relevant.
@@Xirpzy Assuming that everyone should buy into the new standard and that anyone can afford it is actually ignorant. A good PSU stays a good PSU for as long as there are devices it can run, and people either don't see a reason to change their PSU because of a new standard or they don't want to change their expensive high end PSU just because of a new standard. Some people may not even be able to afford a decent PSU at all. Your comment sounds very pretentious as if you're saying "lol just buy it you poor pleb". Very very disrespectful.
Twenty years ago, when I was in high school, car audio had a similar problem drawing down voltage when the bass hit. The "solution" was adding massive 12V capacitors with claimed 1-2 farad capacitance. A lot of them were junk. I wonder if we'll see similar snake oil for sale in the computer market soon.
I was sitting wondering about that. There *were* capacitors you could put in-line for audio cards. Those, well, *worked*, but didn't make a difference because, shocker, the audio cards did it on their own and didn't really have massive amounts of differences in demand anyway.
Now I am starting to wonder if those products might work for this sort of thing. Just buffer it juuuuuuustt enough. Given that it didn't produce negative issues to it when tested for the audio cards.. erm. Maybe it would work?
Seriously a ducttape dirty hack.
I loved that stuff. I was only 12 but my bro and brother in law had some cool setups. My brothers S10 with huge subs was nuts. Everyone should feel the sensation of that much bass in such a small cabin
This is extremely well done, Steve (and crew). A few of my colleagues have had this issue and it was tough for me to explain it to them what was happening with their systems. Now, I'm just going to refer them to this video, as it easy to understand the problem from a viewers perspective. Thanks for all the work you guys do over at GN!
12:55 That moment when you start wondering if your 1000watt PSU will be enough. I actually got mine because my 2080TI and i9-9900k were giving me these exact OCP-ish shutdowns on my 850 so I figured I'd go overkill. So much for overkill... NVidia really should be building more capacitance into the card design if this is how their going to operate. It's ridiculous to expect PSUs to handle this kind of stuff as a general market feature.
Remember that throwing a bigger PSU at the problem can end up tanking your efficiency numbers.
PSUs are MUCH less efficient at lower percentage loads.
Going from a 1000W to 1600W PSU on the same system could end up costing you 2-4x the idle wattage.
This kind of thing is wasteful, and costly.
Maybe a PSU company could just make a capacitor bank for the GPU to sit in between the PSU and GPU
Strange, I ran a power OC bios (can't recall the power but it was the max available on any bios) on my 2080ti liquid cooled with a 215watt 8700k for years on a 650watt seasonic PSU with zero issues. Though I did need to upgrade as the 650 didnt handle my 3080
That might be what was happening to me with my 6800 and mild OCs.
@@kidShibuya Just like Steve mentions, the PSU brand and individual design matters as well. A high quality seasonic 650 could probably handle high transients much better than a poorly designed 850.
That's the biggest problem here -- you can't even go off your PSU's wattage rating, there's so many variables, including how well your motherboard regulated your CPU transients (VRM buffer caps could also have helped you there).
This is why ultimately these huge massive transients coming out of modern GPUs are such an issue. And it's one they could moderate themselves by providing better buffering on the GPU's power delivery, or not letting their cores push transients that high in the first place.
It's the randomness and multi-variable nature of this that makes it so troubling as a trend. Glad GN made this video, needs more general awareness.
Seems odd that the motherboard choice affects whether the problem shows up. Could be that some boards have extra capacitance on the 12V rail that "help out" the PSU when dealing with spikes. Or it could be that in some cases it's the motherboard that causes a shutdown if a 12V sag causes it to switch off the PSON signal.
could it possibly be the pcie slot has more throughput for voltage on some x boards?
Ge it can do 110w vs the "normal" 75w. so when it hits a transient spike, the slot compensates for it.
@@Dracossaint Seems like a legit possible theory, the motherboard should never do that though the spec is there for a reason. Above all though the GPU should never need to be drawing 700 watts and if it is going to be doing that the GPU should be the part with the extra capacitance to handle those spikes without requiring other components to carry the load for their poor design.
Both the amount of capacitance, and the amount of energy the inductors can store will affect the behaviour of a motherboard or GPU card during transients.
It’s not only capacitance. With transients this big, your board needs to be designed differently. If the copper carrying power to your PCIe slot is too thin, or it’s ground is not directly underneath, the added inductance can negate capacitance on the bus.
@@bdhale34 It's not the motherboard that should not do that, it's the graphics card's duty to not draw more than in total 75W (66W from 12V rail) from the board.
On the other hand many boards can provide more than 75W of power. I think since the RX480 there were no GPUs drawing more than 75W from the PCIe slot.
Best PC detectives on TH-cam, maybe even the internet in general. Appreciate you guys, and all the extra work you put into your testing and videos.
25 years back when I was the dude that built all my friend's PCs I found out, that the most erratic and hard to reproduce errors could be attributed to the PSU. I had my list what things to go through if ppl were having troubles but it very often came down to cheap psu that were used in pre builts. Once the pc was upgraded the errors started to mount up. This seems to serve as a good explanation of these experiences. If you buy the PSU judging only by the specification numbers you can have a bad awakening. Thanks for you hard work
By what criteria should I buy a PSU then? Just on the price?
Besides all that you say and are at the Core of the problem, one should consider also that the majority of issues coinceiding with the PSU was also in parallel to 2 other factors,
One was poor Ground for PC chassis and the other one was also poor Grounded House Power plugs which also contributed in instability of the whole system..
that was happening in Older PCs (2012-2013 and prior) and Also the more mainstream PSUs were then of questionable Quality, more often than Today..
Ofcourse PC's weren't that demanding in Power back then .. and also Some brands of Quality existed.
I remember picking up a Totally Silent PSU (Passive Cooling, must have been 450W or 500W, dont rem actually ..that was at early 2000 and the price was then 135EU which was expensive at tge time) powering along a Ati 1600X.. its still runnimg today in a Cube setup Powering a Amd FX 55 System.
@@neutechevo poor electrical fuses were also often an issue but it was easy to rule out and just use a wall plug in a room with less load on the fuse. These issues often peaked during lan sessions as too many consumers were on one fuse. I didn't say that all old psus were bad. Just the ones oems cheaped out on in their prebuilts as it drove down their costs.
@@Dodara87 reviews i guess. Its hard to come by proper reviews for psus since dirkvader de shut down however. I try to reasearch as best as I can though.
@@benjaminolry5849 yes ofcourse. I said that more mainstream PSUs back then were of more questionable quality due to lower power demands.. also lots of us have more than one, burnt hdds psus mobos..etc.. and you could not pin point the actual problem.. i vividly rem. a problem for my own system.. and then i figured it out at sbout a month away was due to 1 or 2 screws of the motherboard not tight in the chassis.. these problems back them were more common.
This honestly some of the best quality, and best explained tech video I have seen in quite some time! Very educational, thank you!
Yeah got that first hand when I upgraded to a 3080ti. 850w should have been plenty for the steady state, but transients were enough to send the PSU into safety shutdown.
Since you asked for details Steve: EVGA 3080Ti FTW3 Ultra Gaming, ASUS MAXIMUS X Formula, 8086k Der8auer edition @5G all-core, Seasonic Prime TX 850 from 2018 couldn't hack it after upgrading from a 1080ti.
The 1kw model of the same line is doing fine, absurd as that is on a system that draws 450w average while gaming according to my UPS.
Great to see real user experiences like this to help establish that this is a factor to consider now!
@@GamersNexus not just that, even my cable extensions from asiahorse weren't able to handle my 3080, just because of the power not being clean enough
@@GamersNexus i have been running a gigabyte 3080 on a rm850 that is quite old for a good while and haven't seen these issues. It's a 320w hard limited model so I've made my own curve for it where it really doesn't do the stupid push past 2040mhz at like 1060mv. Mine mostly run 2025mhz at 1000mv so I wonder how this curve affects transients.. I'm not finished with the video yet obviously but I would be interested to see transients without Nvidia boost algorithm.
This explains an issue I had with a SeaSonic Prime Titanium 1000W about a year and a half ago. Shortly after adding a Gigabyte Waterforce 3080 to a Ryzen x570/5900X system, the PSU blew out. I could not understand how a PSU I had had for 2+ years across multiple platforms would all of a sudden give up the ghost, but to SeaSonic's credit, it only took itself out and every other part was fine, but I literally had no idea it could have been the GPU until now. Yikes!
Had my 3080ti kill a Seasonic Prime Titanium myself, currently in RMA.
This video deserves far more recognition. The production level, the overcommitment to accuracy and honest transparency, and the attention to detail shown here are nothing short of superb. The GN team is doing amazing things for the PC industry as a whole, and manufacturers should take note.
Wow, GN - this is getting very professional. As an engineer I am very happy to see this "marriage" of quality knowledge and consumer/user friendliness. All the best to you. Keep it up :-)
20:45 And when we also account for the folding frequency, it might actually be better to build a separate setup that is offset by half a period, then combine the results into one function. That will eliminate any issues coming from the equipment generating errors due to the high resolution data processing.
22:10 - You can use mathworks MatLab for this, easily.
Moral of the story: Everyone needs 1200w PSU's, separate and isolated breakers, tesla solar roofs with power banks, and a dedicated nuclear power generating facility to accommodate next level gaming hardware.
You forgot home and contents insurance when the fire takes your home and assets 😉
No, replace the fuel rods with xx90/Titan cards and you can recover something like 80% of the power
Well...I got the 1200w psu so far.
I have a 1600w psu. 😎
@@LawrenceTimme I almost got one, but I hesitated until they sold out or the sale ended because they put a different plug on them. Does yours have a different plug on the cord from the wall?
I was aware of power spikes in GPUs, but until watching your video, I had no idea they were this severe. I'm getting ready to overhaul my PC and my failing old Corsair PSU needed a replacement. I've just ordered a PSU with more headroom than initially planned to prepare for newer cards and I'm sure it'll save me a lot of trouble in the future. Thank you GN. You're the only mainstream hardware reviewers that are covering topics like this with actual data. For how small this operation is, it's incredible how far ahead your data and methodology is versus much bigger hardware reviewers.
This is excellent content - thank you so much for the work you're doing here.
I just had this problem. I upgraded my GPU from a 1080ti to a 2080ti, and in the process, I decided to change the case and water-cool the system, which also meant adding a lot of fans, a pump, and all of the RGB that comes with it.
The build that eventually had the issue was with an EVGA Supernova 750W G3, a GigaByte Aorus Elite Z390 MB (with a seated Intel 9700k), and a Gigabyte RTX 2080ti Waterforce GPU. The PSU was about 3 years old, and the system would shut down randomly after gaming for 5/10 minutes. Opening up MSI Afterburner and reducing the power-limit on the card to 90% solved the issue at a cost of performance, so I knew I had to upgrade the PSU. I wasn't in a place where upgrading the MB/CPU/RAM made sense - I'm waiting to see what this next generation of processors look like from AMD and Intel.
I went big and bought be quiet! Straight Power 11 Platinum 1200W. Needless to say, the system is stable now, even after overclocking the GPU and maxing its power consumption limits. If you're having this problem, try under-powering the gpu to see if you can stabalize the system. It's not a permanent solution, but it might allow you to game while you figure out your next steps.
Thank you so much guys, for all the awesome content - loved this video
The more analytic episodes are always much higher quality, although I like the news format uploads as well.
I am a CNC Machinist and shop owner. In dealing with old CNC equipment I have dealt with Controller failures from blown capacitors many times. (in the circuit to handle transient power draw) Thank you for this video explaining the problem to your viewers. (something I would just naturally assume when the system crashed) :)
Awesome to see this investigated and laid out in a straightforward way. IMO it seems like most of the blame lies w/ the GPU manufacturers for not controlling the spikes better, but I agree that overbuying on PSU is the easiest way for consumers to address it. I'm likely putting together a new build late this year. I was planning on once again reusing my Corsair HX1000 that's now over14 years old, but that might not be a good idea even though its 1000W... better to keep it in a system that won't have as high spikes and continue getting reliable operation from it and go with a newer unit for my new build.
I'd be interested in additional data on GN Extras!
In my experience a seasonic focus gold 1000W PSU is able to withstand a 3090 FE with a 5950x on a custom water loop. Before that one I had a Corsair RM850x that crashed during transients.
The sheer exponential insanity that seems to be going on with GPU power consumption from generation to generation now is scary enough...
Don't worry, North America isn't getting any more than about 1500w on 1 circuit. So at least there is that limit 😅
@@ahfreebird Well your monitor consumes some power as well :D
Might get funny in the future.
@@ahfreebird ... my computer and monitors have to be on separate breakers (cause the monitors are enough to trip a 120V 15A breaker when added to my computer).
@@Ri_Shin_Marco No big deal, just gotta run an extension cord to an outlet on a different circuit.
Mainly with Nvidia GPU's. AMD is also rising a bit in power consumption, but their cards are literally hundreds 100+ watts behind Nvidias in consumption.
Also, both Intel and AMD are getting to kinda rediculous power draw on CPU's as well. Intle is already there, and AMD's AM4 socket maxed at 105w TDP's, the new AM5 Socket is going up to 175!
This content is super informative and interesting, thank you guys for doing this piece.
In the world today efficiency should be a huge concern, I was happy when AMD finally started to take it seriously after there older GPU's and CPU's. It's high time we start expecting more from new generation hardware, bigger and faster doesn't mean anything if it's pulling 2x to 4x the power of previous generations. Sloppy drivers and software play there part but with Gamers Nexus on the case we can finally hope for some truth and justice to be served. Thanks Steve and you're amazing crew, thank you as well for explaining how you're equipment works instead of just saying "it knows".
there is no choice, silicon based chips are extremely close to their limit, so they will get bigger and bigger and consume more and more .... and obviously they will be lot more expensive
efficiency doesn't matter. if you're so concerned about your power draw, you can always undervolt for a marginal penalty. more power usage an emotional, but not logical feels bad. so long as it can be cooled, it doesn't matter how much power a card uses.
@@zannare there is a choice, It’s just the more difficult one
@@zannare There is choice, just look to Apple and the M1 chips for a perfect example. As much rightful hate as apple gets for some of their practices, the M1 chips are insanely efficient. My m1 macbook has a battery that lasts two days easy, literally doesn't have fans and still doesn't get hotter than any laptop I've ever used before even under max loads.
It's definitely possible. Not easy, but there is a choice. The issue is the time and money that companies like amd and nvidia would have to invest, and there is littl incentive to do so if you can just offload the problem to the consumer by forcing them to buy bigger and bigger PSUs.
@@balls2thewall724 lolol efficiency always matters when it comes to lowering power draw. Were you joking?
This reminds me of when reviewers added "frame pacing" and 99th percentile frame rates to their review tests for GPU display capabilities. The averages weren't good enough to capture the user experience. One thing both types of manufacturers can do to help: add specs for "max transient voltage or current" over a specified duration such as 100 microseconds. GPUs guaranteeing that they won't exceed that value, PSUs guaranteeing that they can sustain that value.
This was a fantastic analysis. Really love this level of information. I do wish there were AMD GPUs mixed in with the Nvidia graphs. Having just worked on a 6900XT build I was wondering how it compares. Thanks so much for all your team does.
yea, would be nice to know the general difference between AMD and nVidia in that regard.
Absolutely loved this video!
All the effort that went into this video really paid off. (I especially liked the animation 😅)
Thank God someone is seriously talking about this. I remember when I thought getting the 980ti and 1080ti was a far cry from what was really possible for me. Oh how little I knew. And oh how much worse it can get, even from where we're at now. It sure feels like tripping people's breakers is the deep part of the iceberg but I wouldn't bet on that.
good lord, the music at first had me thinking this wouldn't be the greatest of your pieces, but then you dialed it in and got all graphics-y... you guys produce the best content in the field. keep up the fantastic work!
Music over commentary still hurts the video imo but I'm willing to accept that I might be in the minority in this opinion
Gamers nexus putting in work! So much respect for doing proper reviews and in depth articles. This is the stuff the industry needs! Love from the UK
I’ve had this issue with my Strix 3090 and a 1000W Corsair psu…was very frustrating trying to figure it out. Keep up the great work!
I got a 3090 FE and a corsair 1000W power supply as well, luckily never ran into this issue even with some gpu and cpu overclocking.
@@Luke-yd4xq yeah I wasn’t that lucky unfortunately. I couldn’t figure it out honestly my friend helped my say for certain what it was
@@DairyQueen2187 damn, you got that with a 1000W PSU ? Do you have a lot of power draw from other sources ? Because if it's just the normal CPU + PSU + 1-3 SSD+HDD, the 1000W should be able to handle it no problem.
@@Winnetou17 Just a 5950X that wasn’t overclocked and my nvme drive. Don’t think the psu was faulty faulty but it was a serious problem
What psu did u get to fix it?
Goal achieved GN! Learned a lot while watching this, & the way it was presented was very entertaining as well. Kudos to the crew for this! 🥂
So few people out there actually understand and take the time to learn about hardware, and then really point out flaws. Not just surface-level or consumer obvious ones, but really deep dives. When you guys speak, people listen! Your channel makes a big splash.
Every generation the power draw grows, the heat grows, and the heatsinks get bigger and bigger. It's not sustainable.
That last part isn't really accurate. Power consumption (for both GPUs and CPUs) has see-sawed over time, but definitely with the 40XX series we're looking at record-setting levels of power consumption. With this generation, Nvidia is trying to beat AMD by throwing efficiency to the wind and just pouring raw power through GPUs to force faster clock speeds. This video shows one of the challenges of doing that, in that if your PSU draw 450 watts normally, it might need to be able to survive transients of 1500 watts. Both the day-to-day operating costs and the cost of requiring such a robust PSU will make these GPUs very poor value propositions.
Also, by taking this irresponsible approach of "MOAR POWAH!!!!11!!" and boasting about their power consumption numbers, they're just asking governments to step in and enact hard and fast regulations limiting maximum power consumption. And that likely will just make everybody unhappy.
It might not be sustainable, but it doesn't have to be. At least for mainstream consumer cards. Consider what the current high end cards can do: 4k, 120FPS, ray tracing. At least two, possibly all three. Whats left on the graphics wish list? 8k? Impractical on account of the shear size and cost. 4k VR? Already have it, just turn off the RT. More FPS? Diminishing returns past 60 already, and your not running competitive FPS like CSGO with 8k textures and ray tracing. Therefore, hello 600FPS.
There are just not any technical feature left to add in, and while there are a few niche combinations of games that can max out a 3090 and have sub 60FPS, lets say the 4090 doubles the graphics power for not double the power cost. RT is the final word when it comes to graphics features (I'm saving about 2 pages of explanation with this) and that has a more or less fixed compute cost. So lets say the 4090 has enough graphics power to run anything that you can reasonably run. Someone will have to draw the line in the sand and say" Our card can do 4k VR with RT with 8k textures with 240FPS... what do you not have phones, err 8k VR, err eyes that can break physical limits?.
So the 50 cards will be refining the design, dropping power, and dropping cost.
Thats not accounting for the artistic/scientific types, but they are going to always be wondering how they can fit xx90 #5 into there build.
@@megachonk9440 considering how relatively small the market for high-end enthusiast graphics cards is, government regulation vis a vis power consumption seems highly unlikely.
@@sharpe3698 Given how the trends are, we might be seeing silly power consuming on low and mid tier graphic card where power consumption vs performance become a cost vs value issue.
@@nickierv13 As long as cards *can* advance, games will continue to push the boundaries of what's possible. Everyone always says that there will be a limit, but it's never been true. There will always be some level of detail, some additional effect that opens the floodgates to more performance being needed. Todays that's raytracing, tomorrow it might be in-game fluid simulations, or per-hair physics or SOMETHING. No one is ever satisfied with stopping and saying "good enough".
1: That music. Nice. 2: Back in the "old days", general wisdom was to get a power supply with a 25% higher capacity than what the system added up to. It was mostly to guard against efficiency drops, but it seems like it might be good to help with the spikes.
couldnt u put a bank of capacitors on the 12V rail?
You guys have again demonstrated why you’re one of the highest quality computer content creator on the Tubes. Great job raising the bar.
This is a video I feel like I've been waiting for! When I did my recent build (12900K & 3080 Ti) I went for a 1000W PSU (EVGA SuperNOVA 1000 P6) even though I could have gone with less in theory. An actually quantifiable set of data demonstrating this need is a nice kind of post-purchase confirmation to get. Thanks for this kind of work!
I went thru 4 RTX 3060ti’s before I realized this issue. I had a 600 watt power supply which is enough if you go by the manufacture standards. I now have a 850 watt and have no issues. So I think it was the transient spikes that was frying the cards
I love the depth and angle on this piece. The writing is great, and Steve is a great story-teller. He could easily do books on tape, if GN ever falters.
I also love that my Patreon support helps fund videos like this!
Would have been great to see what happens in lower-tier cards, like a 3060 or 3070. Sure, the 3060's TDP is 170W so most PSUs could take a transient as normal, but consider someone on an older 450W power supply, which technically meets Nvidia's minimum spec.
Im using an rog strix 550w with a 3070ti , not even min specs lmao
@@blindshot709 is it... working?
@@KellonMelon65 yes... 6 months no shut downs
@@blindshot709 it is probably throttling
@@Metalwrath2 how would it be throttling when it never goes above 350w leaving 200w head room for spikes . also that 350 is under full load lmao gotta love 65w cpu's
Wouldn't throwing beefier capacitors on the boards fix this, seeing as it's such short spikes? Or at least smooth it out enough so the psu could cope?
Yes, but that would mean more cost to the gpu vendors, which they dont want to do.
They only want to produce cheap low cost boards at a high sale price, expecting them to put high quality components on is simply asking too much of them.
This was also where CAPs inside the power cables were also a possible solution 5+ years ago and I think will start coming back soon especially with such high spikes becoming more common place.
There are multiple tiers of power delivery. The GPU has local regulators to convert the high-power 12V rails into low-voltage, ultra-high current rails of somewhere around 1V, plus or minus. If the 12V rail dips due to excessive load, it is going to have to fall to less than the GPU operating voltage, plus some drop-out margin, before it causes a problem. The GPU transients are actually _already_ being serviced by the point-of-load bulk capacitance. The transients on the 12V line that remain are from current draw to replenish those caps.
The 12V rail's sagging could cause problems elsewhere, but there's not a whole lot left in a PC that actually runs at 12V. It's almost entirely used as a fat current pipe. So the CPU will have its own local point-of-load regulation and the same kind of resilience as the GPU.
Larger cap banks in the PSU will absorb transients better, but that actually has consequences as well. For one, they will completely bypass short-circuit protection by having the capability to supply a LARGE quantity of power in a very short time, which isn't always desirable. Also, as transient response improves, regulation actually suffers, since by nature of being able to smooth over changes in load, you're making change in supply less visible to the switching controller.
Dealing in KW power is inherently risky. It's just a lot of energy, and handling that amount of energy carries some risk. Nothing in a PSU is going to blow up having to carry 2x load for 100uS. Tripping the OCP in spans of time of 1ms or so is solely to prevent cables from rapid heating, shorted contacts from welding, and things like that.
The clips in this video where the fan speed dropped significantly .. that's not a 100uS spike. That's a prolonged draw. _THAT_ is just an insufficient supply, because 1-2 seconds of heavy load is no longer a "transient." It's _a load,_ and you need to size the supply appropriately. More caps aren't going to solve that problem.
@@nickwallette6201 It's going to take a whole ecosystem approach to address the issues with Transients and Transient load. When you think about it, motherboard OEMs beefing up VRMs lately at a rapid pace is starting to make a whole lot of sense imo. Anyway, I think we are going to see a revamp of the 80 plus certification and anything less than "Gold" will slowly start to fade away in one to two GPU generations if there is no revamp of the '80 plus certification' on the lower end.
Ran 3600X + 3080 Ti with 650W Seasonic Focus Plus Gold - no issues.
Changed 3600X -> 5900X, 3DMark Time Spy - system shutdown.
System runs okay 750W Corsair RM, but 3DMark CPU clock frequency chart shows anomalies when running 5900X PBO + overclocked 3080 Ti.
850W Seasonic Focus Plus - all is well and no issues besides PSU coilwhine when under load.
Hey Steve and GN team, what effects does Undervolting a GPU that exhibits high transients have?
Undervolting already has a ton of benefits, would love to see if it can be a way to deal with transients to avoid having to replace a power supply!
If the GPU itself can handle it I'm sure it will make a difference. You can also power limit the card and that will guarantee reliability
@@Mpdarkguy @Cătălin Power Limit wont fix as the max power limit of rtx3090 is ~350-400w depending on the model so it never should be able to even go to 600w+ spikes but lower voltage means there is really less power going into the card and therefore should improve the spike behaviour
I was thinking this as well. The down side is that the silicon lotto plays a role in how much you can undervolt, only allowing a generalized answer at best. It would still be handy to know if that indeed helps or not.
I undervolted my 3080ti FTW3 to 875mv which was quite a bit lower than the ridiculous stock current that EVGA decided to pump into this thing, (Which was absolutely unnecessary because I stay well above this card's boost clock and stable at 875mv) and while it did reduce my peak spikes, I still hit 425+ watts during spikes. (I originally would have them in the 460s) These spikes were just during regular gaming. I can't imagine what I still might manage to hit during some kind of GPU torture test. That said, I've never had a lockup or PSU shut down of any kind. Although, I am running an EVGA 750 Watt gold SuperNova, which are pretty rock solid.
So in total, I'm using an 8700k typically running at an almost 5GHz all core overclock, the 3080ti FTW3, two mechanical hard drives, three SSD drives, four 120mm Corsair LL series RGB fans, a single 120mm slim fan, a four port USB 3.0 expansion card, and about 8 USB devices total going into the PC with, typically, half of them drawing some kind of power at all times.
I'm probably OK because of the 8700k. I'd imagine if I put in any newer chip with 8+ cores, I might run into some issues.
Honestly just undervolt. It's so beneficial. My FE 3070 has it's power consumption cut by two, i can do 850mv at 1860mhz + 500Mhz vram OC stable.. On forza, i go from 205-220w to 110-135/140 at worst..
Thanks GN team for all you do! This video is insanely well made, props to you guys for setting the standard for tech journalism
Now THIS is tech journalism, I really hope you guys can keep on doing this kind of content for as long as you are around.
Well, except for the first 5 minutes where I kept waiting for Chris Hansen to bust in the door o.0
I really appreciate this experimental approach, rather than just presenting existing benchmarks.
Yours is the first technology channel I have $thanked.
I'm wondering if this these transient draw situations were the kind of thing that led to some of these 3090 GPUs having VRM transistors that failed under load during certain games?
We already know that some games and some setting combinations create worse transients than others, so my thinking is that this could have led to a "stack-up effect" that ultimately surpassed the ratings of the VRM transistors for long enough that they just couldn't take it any more.
Buildzoid @ Actually Hardcore Overclocking talked about this one.
IIRC it's because of improper phase balancing, that is, some phases are overloaded, while others arent really doing anything.
The whole VRM design on the 30 series is just shit, according to him.
I can't remember which vid it was, but it was a recent one.. IIRC he was fixing a 3090.
@@patrikjankovics2113 30 series in general, or that specific 30 series card he was working on? I have an Asus 3070 Ti in my computer right now, probably fine but I'm never gonna mess with it other than putting a water block on it.
And I know that Asus does go kinda crazy with the VRM design on their motherboards at least, so it wouldn't surprise me if they did something similar with their Strix or Tuf line of 30xx cards (mine's a Strix).
But yeah unbalanced phases does sound like a wacky problem to have, is there any particular reason they can't just parallel all the various phases together? I know "ganging transistors together" doesn't generally work like that, but I was more thinking of ganging together individual buck regulator VRM sub-units, so they'd basically all be wired in a "diode-OR" type fashion.
@@44R0Ndin The last solid top end card made by nVidia was the 1080Ti. 20 series are prone to power delivery failures and the 30's are even worse. Certain cards are extra shit (Gigabyte) but the 30's in general are unreliable. Then there are dummies who think flashing their unreliable cards with "1000W bios" is a good idea.
@@TwiztedJugallo Good thing I'm relying on 3rd-party cards from more-or-less reputable brands then? You're not gonna get me to give up my 3070ti for a 1080ti or 1660, and so long as I don't mess with the VBIOS it shouldn't have any issues.
And in case you didn't realize, or I forgot to mention it earlier I don't have a 3090 or 3080Ti.
I have an ASUS Strix 3070Ti, and it's been doing just fine for me.
My previous 2070 did die on me after 6 months, but I traced that back to the Micron-branded GDDR6, and not a VRM failure (it even looked like memory failure on my screen, plus I found out that there was a bad batch of GDDR6 from Micron that got used in a lot of 2070 and 2080 cards which is why they were dying frequently, but that particular problem stopped happening when they switched to Samsung branded GDDR6).
Now given all the problems that the 3090 has been causing, I'm glad I didn't have the money to get one, but even if I did I wouldn't have, all I do is play video games (I don't make videos, I don't even use Photoshop or anything like that), so what I have is perfectly enough for the needs I have.
If the 12v power sags then the VRM has to draw more current to keep it's voltage up. That would drag the 12v power down even further.
Should also note that certain models of power supply are more vulnerable to transients than others. A 750W EVGA G6 might be able to handle a 3080 just fine, while a 750W EVGA GA likely will not. The Seasonic Prime-based units are also problematic in this regard, just to name a couple of the more popular lines. Either way, GN did a fantastic job of explaining transients here, and I look forward to what's in store.
Also, I suggest looking at Aris' Powenetics V2 which he announced recently; could be pretty useful in testing.
It basically depends on what BIG capacitor is used within power supply, so it can withstand sudden spike
@@DimkaTsv There is an awful lot more to transient response than just how big the capacitors are. Capacitance is an important part of it, but you can't know about whether a PSU can handle a 3090 just by looking at the caps.
@@nunyabeeswax3012 Sure, i guess... I am no electrician in any way, so just answered based on own knowledge, which is limited
I’m still using my “1st edition” blue GN mouse mat. Still gotta get a mod mat someday. Would be cool if there was one that was still useful with stuff like the ruler, but had more general diy diagrams and stuff.
I'm someone who has been building my own desktops for over 20 years. To save myself any potential trouble, I bought a Cooler Master's 1000W PSU way before I even got a myself an RTX 3080 with a i9-10900K.
It's always good to do your math before building your own PC. And if the buyers intention from the onset is to get a top end quality system, then don't scrimp out any components, not even your motherboard.
You sound like you know what you're doing. And I'm trying to plan ahead better in the event I get a 4xxx series card. I am recent to building my own systems (two years). I have a 3090 and 5900x powered by an NZXT 850w PSU. This is my second PSU, and my MSI Meg Ace x570 had to be RMA'd, and it's working again. So, it's not been all rainbows and butterflies, but I'm learning and the system is stable now. Question: This week I picked up an Antec Signature 1300w PSU just so I had the option of getting a 4xxx series card if it works out. Good purchase or should I return and hold out for an ATX 3.0? Linus indicated PSUs would or could be hard to find in the fall?
If you were trying to use excel to deal with a large .csv file, I highly recommend you use the Pandas Python 3 library. It can handle millions of data points before it gets too bogged down. I use it at work to deal with large .csv files. Also great explanations of the DC amp clamps! I use them at work for 3 phase power analysis and know the pain of trying to keep them calibrated and accurate.
Strongly agree with this
Agree. You can easily divide a large dataset into chunks with pandas, using the chunksize parameter.
I'm definitely interested in the GN Extras deep dive version of this!
I'm an avid VR user for gaming and until late last year I had issues with games being stable and it was very hard to find out why. After all that trouble I decided to buy the best power supply I could find hoping it would have the best quality components that could handle any (my supposition here at that time) very short power peaks that might disrupt my VR experience. I invested in the Corsair AX1600i 80Plus Titanium Power Supply with a 10 year warranty. At the time my rig was an ASUS ROG Strix 390-F Gaming Motherboard, an Intel i9-9909k CPU, 64 GB RAM, two Samsung NVMe SSD for operating System and my most demanding games (had other SSD's for other games), Nvidia 1080Ti on Windows 10. I could now discount the PSU for any problems I was having. Shortly after, I upgraded to an ASUS ROG Strix GeForce RTX 3080Ti OC and this solved all my issues with VR. I upgraded again for flat screen games to a Viewsonic XG320U 32" monitor running at 144Hz in 4K. I limit in Nvidia's Control Panel all games to a maximum frame rate of 144. With this combination I can even run (all without mods) Fallout New Vegas, Fallout 4 flatscreen and VR. These games have some terrible coding that most people complain about but now I don't have any crashes or stuttering. For those playing Fallout 4, in advanced settings you must not set Ambient Occlusion to HBAO+ (Ultra) so set it to SSAO (High) and set Weapon Debris to Off (best performance). All other setting can be maxed out. I also use in all games Image Scaling On in Nvidia Control Panel and in GeForce Experience Image Scaling set Render Resolution to 85% 3264x1836 and Sharpen at 50. I am now on Windows 11 with no issues. This represents several years of trial and error and I hope someone finds it useful.
GTA V @4k max settings killed an old Gigabyte 3080-ti Vision GPU after just a few days of owning it. Died after about 1hr of gaming, fans just span with no signal. I was using a RM750 PSU, 5950x and Asus Dark Hero mobo. Upgraded to a RM1000x and refunded the GPU for a 3080-ti Hall of Fame (Which was somehow $300 cheaper) and it's been running perfect since. I wonder if I experienced something similar to what's happening in the video.
This is the most impressive, in-depth, detailed and easy to understand to address and explain the transient issue.
Good Job GN Team.
The video was excellent. The explanation of the topic was fascinating, B-roll kept the flow of the video up & the title really drove the click. Good work GN, more like this with your focused videos.
This was an excellent video! I haven't had a problem with it myself, but I've heard of transients and how problematic it has been for the RTX 30 series. I had no idea why and how it happened though, and this was the most concise explanation I've found! I feel like I finally understand it - so thank you, GN! Awesome work on this video, all the research you did, and the way you presented it. It's really good info to know, and I'm really glad to hear you'll be addressing it more from now on.
You make a very good point around the 25-26 minute mark about over-buying the PSU. I suspect these spikes are what often kill pre-builds PCs that use the cheapest PSU that barely meets, and sometimes don't even meet, the GPU power recommendations.
I disagree with that way of thinking though. In my system I always over-buy the PSU. For an extra 30$ you'll get an extra 100+watts. The nice thing about having the extra head room is future proofing for upgrades BUT if you aren't pushing it to its limit the PSU generally runs cooler making it last longer. So do you want a 750 that will last 5 years or an 850 or 950 that will last 10 years and keep running through those spikes? Also they run quieter when they're not being pushed to their limits. Is it wasteful and more expensive? Yes, technically. Is it worth it for all the benefits? I would argue yes.
As someone who worked with UPS and power bank installation and maintenance for a living, I'll let you know that, without getting too technical (at least in the main channel lol), you just raised the awareness on transients by quite a few Farads ;-)
I was concerned that this channel was going to become less entertaining as it became more invested into technical benchmarking equipment and data. From this video I think I see your vision now, and it's brilliant. Important work being done here. 👏
Excellent. As was the recently updated logo, this video is an improvement of already high quality content. Good job crew!
Thank you guys for putting this out. I actually encountered this awhile ago on my RX 5700 XT - random shutdowns on a 650W ATX PSU. One thing to note here is that the PCI-E slot is capable of supplying up to 75W according to spec, and I would imagine that some motherboard manufacturers have incorporated additional capacitors for handling sharp transients. I tested it as extensively as I could (given that I didn’t have an oscilloscope and had to rely on the built-in Radeon monitoring software to track power consumption), and was able to determine that the base load was higher than a single PCI-E power cable could reliably handle (as, according to spec, each line only has to be able to supply 150W of power). One cable plus the power from the board gives you 225W of power according to spec, and I noticed that the board was routinely topping 250-300W of power consumption as a baseline-so it was no wonder that until I started using a separate PCI-E power connector directly to the PSU (rather than daisy chaining the 8+6), I was seeing OCP shutoffs. And even once I corrected that issue and started using two separate PCI-E power cables, I was STILL getting the occasional OCP shutoff.
The fact that boards can have such insane spikes is really frustrating as a consumer, though, because it realistically means that PSUs are being taxed far more aggressively by GPUs than had been previously reported, and that will cause them to age more quickly. That same 5700 XT actually ran a 750W SFX PSU from be quiet! into the ground-power supply worked well for a year, and then one day it just started triggering OCP on just about every boot. I shudder to think what it’s doing to the Lian Li I replaced it with, or if I should consider investing in a beefier SFX PSU for my rig (or else abandoning SFX altogether until this trend dies down).
This is interesting. Which 650W PSU and GPU custom do you use? I haven't had single shutdown in the ~2 years use of my Sapphire RX 5700 XT Nitro+ with a 550W be quiet Straight Power 11, but that is also a high quality PSU, not a cheap unit. I'm running it together with an R7 3700X or, for the first year, with a 2700X and both at stock. I wouldn't have thought that this could happen with a GPU of this caliber and a good PSU. I always had both power connectors on the GPU connected and I never used a cable splitter, if that's what you're referring to with daisy-chaining.
@@dennisjungbauer4467 It was an EVGA 650W ATX PSU, I forget the exact model, but it was also more likely an issue with the GPU itself. I purchased the Gigabyte RX 5700 XT Gaming OC pretty early in the product’s lifespan, and had to deal with a lot of growing pains while they refined the firmware and the drivers. This board in particular seemed to be plagued by uncomfortable power spikes, as a few reviews I read and a few forum posts I found at the time suggested that it was a problem.
The card is currently *mostly* okay on a Corsair SF600 (order history confirms it was not a Lian Li PSU I purchased), but it nuked the be quiet! BN639 SFX L 600W PSU that I had been using it for in my ITX build after only a year.
@@joshcurtis3752 Ah, right, I almost forgot about all the issues this card had for some people. I guess I was lucky in that regard as I haven't had any hardware issue, only few driver issues switching (well, fully new PC) from an Nvidia card (GTX 970). And while I didn't buy it very late, I bought it at end of 2019, so ~5 months after release.
Awesome, as always! Thank you so much Steve, your team does an amazing work! I can not find the right words to describe how valuable what you, guys, do in this channel! Thank you soooo much!!!
The research you guys put into all this and continuously improving the quality of content is just mind blowing. This is exactly what is called giving back the community what it deserves who helps you grow as a channel.. Hats off
I've said it before and I'll say it again; I think Gamers Nexus is the true heir apparent to the legacy of The Tech Report. This type of thing is exactly the type of work we expected from TR back in the day with Scott, Geoff, Cyril, and co. If GN can figure out a way to make this kind of GPU power capture easy to accomplish I think it will have a similar effect to Scott's work with bringing frame time to the benchmarking game back in the day.
It would be great if GN could put together a list of:
- PSUs that handle transient spikes well
- motherboards that handle transient spikes well
Lots of testing, but, it might put the hot poker up the butts of the manufacturers.
But this is mainly an issue with the add in boards not filtering power properly
who knows, this might be a thing more in the scope for ltt labs. generating databases for common issues like this is something that i would love to see out of that whole project
A point I'd be interested in knowing: how do overclocking and undervolting play into all this? is it just a reduction of power/an increase of power type effect? or are there potentially knock on effects, making overclocked cards more/less susceptible than expected?
It goes both ways if OC'd, you already are pushing the TDP... (higher Volt., higher consumption) that means the transient will be far bigger...
And the other way around...
My takeway from this video is that a transient spike is based on the average draw. Increase average draw, and you'll get higher numbers in your spike.
So if you OC your card with a higher power limit (say an EVGA 3080 with a 450W XOC bios) you'll get a much higher spike than if you run the same card undervolted, running around 320W or so (this is about what my 3080 FTW3 pulls undervolted according to msi afterburner).
So overclocking definitely increases the strain on your psu as the total wattage number in the spike will be much higher.
By the way, who did the animation showing how the clamp worked? That's awesome stuff, right there. _Such_ a nice touch for the channel to help explain - _with_ a high-quality animation - how the testing solution works in an intuitive way.
Wasn't there an aftermarket overclocking GPU that came with an optional capacitor module you could stick directly into the backside of the GPU die? If memory serves, it was sometime in the DX11 era.
I feel like that was trying to solve this very issue for extreme overclocking.
I think 780 ti or 980
@@lgpa565 I remember now; it was the MSI Lightning line of extreme OC cards, starting with the HD 7970 and continuing with the GTX 780.
They called it the "gpu reactor".
Transits are a problem known for a good while. I suspect my old board was tripping because the caps were getting old enough to no longer deal with power fluctuations.
I'd really like to see you test the transient handling performance of the Corsair SF750. That's the go-to PSU for SFF and routinely handles top end components. I'd expect it to pass the tests that the NZXT unit failed.
I would like to see such a test too since i am running my Asus 3080 TUF with an Corsair SF600... so far no problems but perhaps its because of the slight undervolt im running (although i didn't in the beginning) and the games i play aren't all too demanding.
Incredibly interesting and actually very capturing investigation, love this style you've gone with
I think I understand why Microcenter has soo many high end cards in the open box area. Thank you.
Fastforward to April 2024 - Ryzen 5800x3d + 4070ti on 600w Bronze PSU - Cyberpunk on max with RT - no spikes, working fine, lmao
yeah it's strange, i have the spikes issue on my 3080 but only happens on Battlefield 2042, Sea of thieves and Battlefront 2, it didn't happen playing cyberpunk at max settings with rt and all, weird lol
This is well described. @Gamers Nexus I respect you. For good reason.
Great journalism, news, analysis, graphics, everything! Setting a great standard for PC enthusiasts!
I made myself a borderline on TOTAL system power draw under heavy load, and GPU power draw. It is 500 W total and 250 W on a GPU and I am not interested in anything exceeding that no matter what. It leaves me only the best perf/wat options on the table, and considering my needs that is all I want. Currently sittning on a 5600X and 6600 XT, perfectly happy with 1440p resolution.
Gamers Nexus does their homework! This is what makes them so great! Things like when they showed the correct direction of a magnetic field with respect to the direction of current in a wire, I literally clapped out loud!