Will RTX 5090 Be Too Fast For Any Current CPU?
ฝัง
- เผยแพร่เมื่อ 20 พ.ย. 2024
- ► Watch the FULL Video: • DF Direct Weekly #159:...
► Support us on Patreon! bit.ly/3jEGjvx
► Digital Foundry TH-cam: / digitalfoundry
► Digital Foundry Merch: store.digitalf...
► Digital Foundry at Eurogamer: eurogamer.net/...
► Follow on X/Twitter: / digitalfoundry
Just replace the CPU with a GPU silly 🥰🥰
Like duh 😂
This is kind of possible but it takes too long to explain. IBM Cell was one failed attempt at it.
The more you buy, the more you save ~
exactly. you have money for 5090, but not for CPU upgrade lol
@@daniil3815
I think you missed the joke.
Anyway, I remember Intel trying to go the other way with Larrabee trying to use stripped-down x86 cores to create a GPU. It was interesting, but predictably was very inefficient for GPU tasks. So I'm curious WHY they put the money into it. I'm not going to just assume there was no purpose whatsoever based on my limited computer experience.
VR needs more power. Always. My 4090 isn't good enough for high resolution headsets in graphic intense games. The new headsets on the horizon with 3800x3800 resolution per eye will easily chew through the power of the next few generations of GPUs.
Real
Who in their right mind would develop a game that NEEDS something faster than a 4090 to be good.
If VR needs that much power, VR games need to go for several thousand dollars each to make a profit with current tech.
One can hope that next gen consoles and NVidia 6000 series mid range cards get to that level
Hopefully, there are tricks like frame generation and higher resolution where you are looking that help.
@@mattzun6779 Not official games. But the UEVR mod allows you to play non VR UE games in VR mode. Imagine running the latest UE5 games at about 6k resolution, aiming for 90FPS stable. My 3080 couldn`t even run simple games like Talos Principle 2 at a playable quality or frame rate. Best I got was a blurry mess equivalent to 720p at low settings at about 80ish FPS. This is why I am waiting for the 5080/90 so I can finally play those UE games in VR.
Yea and tons of games can do higher render scaling, and the 4090 is not strong enough 4k max 8k Render at 144hz+
@@mattzun6779 I personally would hope games will be scalable enough to use all the power from the lower end up to the ultra high end.
What a world we live in where we are concerned that our GPUs have gotten so good we bottleneck the CPU at 4K high refresh rate. I remember when getting 1080p60 was the dream for high end hardware.
its still is for me on a 2080ti although with DLSS i can go up to 1440p.
@@stephenmeinhold5452 Yh Alan Wake 2 can barely run at 1080P on my 4090 with raytracing on lol.
my 3900x lacks in 4k in some areas believe it or not
@@Veganarchy-Zetetic You know that this is due to developer's laziness to optimize the game, right?
@@LordKosmux I would say it has a lot to do with Ray Tracing.
32" 4K/240hz OLED displays would be an obvious target for the 5090.
8k 120hz/144hz
even a 5090 wouldn't be able to push 4k 240fps... unless u only playing cs and LoL lmao so I seriously see no point in looking at those displays yet. 1440p still makes way more sense for high refresh rates.
That's all well and good, but I've just gotten a 4k 240hz monitor, I have a 4090 and 5900X, and even games from 2015 are CPU bottlenecked, 240hz should never be a target. Making 8k gaming doable seems to be what it'll be targeting, we don't need better GPUs for 4k right now.
@@GatsuRage Well the more FPS the better. 4K looks way better than 1440P, and OLED is better than any display in the market.
@GatsuRage I get 180-220 in cod Vanguard with my 4090 at.4k max no dlss. But yea the 5090 won't be able to do cyberpunk 2077 max 4k 240 and other similar games. But easier games for sure.
Diminishing returns at this point. Games are unoptimized that's the only reason why these ridiculously stupid specs are needed.
so true
correct I have a Rtx 4080 i am not upgrading for another 3 years.
It's getting pointless
@@justhomas83 i play to get 50 series since im on a 3070ti, but maybe AMD us an option too.. sad we have to buy top tier gpu's to play games that should run on 3080 and no frame gen
@@saliva1305 I feel you I just can't do it anymore. I have more bills coming in now and my daughter is graduating college.
I understand though you're right the top tier is the only path now. We are chasing rabbits in Alice Wonderland at this point 😔😔😔 damnit
yes exactly and the fact that RT/PT exists - it's there to make people $pend, $pend and $pend some more
amazing graphics are possible using non-RT/PT - just look at RDR2 from what 2018(?)
I've been saying for years now that devs should be using RT/PT as a TOOL to see how scenes should be lit and then recreate them with NON-RT means which will give the players a ton of performance and be much friendlier to their wallets
plus hand-crafting it can look nicer
As a 4K/120 gamer I can promise you that we're still GPU limited with the 4090. I often have to compromise on AAA games by enabling DLSS 3 or lower settings, and sometimes just hit 60 FPS. At the same time, I do think upscaling is changing our requirements and expectations, so I hope the silicon can be optimized around that. We don't necessarily need more pixel shader performance if we assume upscaling. The die space is better spent on other features.
I can think of one title you would be fully GPU limited on that card is Cyberpunk 2077 path tracing mode.. and that game is an outlier and also extremely CPU heavy, I'm CPU bottlenecked on that title in certain areas on a 4070s so I'd get your card looked at because something wrong there.
And to max out high end VR headsets these days a 6090 is probably not enough...
@EmblemParade - What CPU and RAM are you running? CPU and RAM has a lot to do with it as well. - but I also agree with @starkr3ality - 4090 @ 4k 120 should be running fine on all AAA games with a couple of exceptions being Cyberpunk or Microsoft Flight Sim - A 4090 should be maxing out all AAA @4Kx120hz. You have got to be hitting a CPU Bottleneck and that is why your having to lower graphics settings to get your desired results. Again this is mostly assumption at this point as I have no idea what CPU or RAM your running but this is what makes sense in my mind with what you stated in your post.
@@ghostofreality1222 Look at benchmarks from HardwareUnboxed, GamersNexus, and others, and see that you are very far from the mark. I have a 5800X3D and high-end DDR4 RAM. I'm not saying that I'm not having a great time with this setup, but forget about ultra settings AAA at 4K without the help of DLSS. The 4090 is great, but 4K is a lot of pixels.
@@StarkR3ality cyberpunk is no longer an outlier...Alan wake 2 made sure of that. I am guessing helllblade 2 further confirm that.
I'll only consider the 5090 if it puts my house at risk of burning down. If it can't do that then its not powerful enough.
Nvidia Power Plug Engineer: "Hold my half caf soy milk grande carmel macchiato."
I'm an overheating enthusiast myself and would like to see products that carry risk of spontaneous human combustion.
Easy, just put steel wool on it while playing
Just get a 14900ks then your good to go
it will be more efficient than a 4090, which can powercapped at 70% and running at 98% performance.. meaning a 5090 will be able to run at itsnfull potential at lower than 300w. stock power is too much power already
I need a 5090 for My 1080p monitor, like a fish needs a bicycle
Hey hey hey hey now!
What kind of "fish-racist" are you?
Does a fish not have the right to evolve into a bike riding creature?
Incredible you anti fishists!
@@darkdraconis it's a song bruh
@@professorJorge11 it's a joke brah
Let me teach you, professor. There's this thing called DLDSR. You can render games at 1440p and 1620p, mantain high af FPS and still get a much better and crispy clean graphics on your 1080p. Sure, a 4090 user (or even a 5090 user) could do with a 1440p monitor, but your analogy is entirely wrong in this case as there is a use for those GPUs with a 1080p monitor.
@@CeceliPS3 I have a Radeon 7600. There's no DLSS, it's FSR2
I mean, 4k120 has to be on the table for most games. Even 4k90, throw on some DLAA, call it good. Aside from totally broken PC ports I feel like we're still GPU limited 99% of the time, especially if you have something like a 7800x3D.
depends what res like you've said, but a lot of the games alex mentioned and others, I'm bottlenecked on a 5800x3d with a 4070 super at 1440p in a lot of modern titles, which is crazy right?
For me, only use case for 90 class and even 80 class GPU's is if you're going to be using 4K and ray tracing in every title available. When you're getting GPU's like the 4090 that are doubling performance gen on gen, and then you get a what? 20% increase in perf going from a 5800x3d to the 7800x3d.
CPU's cannot keep up and it's really starting to show.
@@WhoIsLost defo not pal can assure you, I still great good performance, and I'm not talking about every title only the recent big ones, Baldurs gate 3, cyberpunk, witcher 3 next gen.
My point was is if I'm CPU bottle necked at times in some titles, what's a 4090 gunna be which I think offer 2.5x the performance? not to even mention the 5090.
@@StarkR3ality4090 is only ~60% stronger than a 4070 super. But yes, CPU performance is an issue... I think 4K will be a problem, and there's just not a market for 8K... The only avenue that's available might be VR
@@WhoIsLost Negative. I have a 5800x3D and 6900XT and am CPU limited most of the time at 1440p.
@@JBrinx18 4090 is +70% faster in video games than the 4070 super. Not sure what you mean by "only" 60%.
50%, 70%, or even double the power of 4090, as long as it's not melting it's connector it's fine by me
I expect %60 faster rasterization performance and 2.5x better raytracing. While pushing the same Power output.
@@ZackSNetwork exactly. efficiency is really underrated
gonna pack 2 of those filthy ass conectors man lol
@@ianism3 is not, is just that those very high tier cards are made to get the maximum amount of power possible, for efficiency you have the 80s and below that.
@@ianism3 4N node how much efficency can they squeeze out of a similar node🤔.
I will just use DLDSR + DLAA on a 4k 120 Hz TV/monitor. CPU won't be the limiting factor, and I'm always getting graphical upgrade. Add path tracing with maxed out settings and the GPU (even RTX 5090 TI Super) would scream for help. I just love DLDSR and how versatile it is for all type of games… Doesn't matter if I'm playing new or older titles.
[It's 2030, goes to buy RTX 9090 with all of my savings.]
Me: Hey, you have latest GPU, May be RTX 9090.
dealer: (goes inside and comes with a 8 feet server wrack.) here's Your GPU sir, just $100k dollars.
rah. it'll be a tiny quantum board... but you need to buy a giant freezer sized cooler...and not like the fridge you have at your house...like a giant walk in-freezer. You will also have to pay $500K to turn it on, wait for 2 days to get the thing to cold enough, and then spend $2.3M to run crysis, $2.345 for doom, and $76B for fortnite... no the game will not look more realistic.
@@akam9919 I might create physical sets for all those games with that kinda money 🤑💰
@@akam9919
Virtual reality ❌
Reality ✔️
This was somehow very hard to read/understand, but I got it at the end of the day.
What if they get smaller instead? A GPU the size of your smartphone. And the price of a house.
There is always a bottleneck in a pc depending on the resolution and game settings. Doesn't matter if it's a $400 pc or a $4000 pc.
Not exactly if your hardware pars together fine the bottleneck can be the software and not the hardware.
thats why I aim for a monitor bottleneck
Such a dumb argument, there is a difference between something BARELY hindering performance of another part vs a HUGE hinderance.
Even Micro-software is a bottleneck before 8/2024 update.
the 4090 is still hugely gpu limited in games like MSFS2020 in VR, and i expect 2024 to be utterly devastating in VR - and i'm all for it. We get closer and closer to Star Trek computer simulations every year.
Visually yeah a single flagship videocard is getting closer and closer when it comes to rasterization. But the holodeck virtual npc's have GTP-5 or higher level AI, the physics simulations are orders of magnitudes more realistic and computationally demanding, true pathtracing without the need for any denoising, ultra realistic animations. Latency is also an issue with these LLM responses. So you'd need an entire futuristic datacenter's worth of power with futuristic networking that provides exponentially quicker responses, so even a single flagship videocard from 2045 wouldn't be enough.
Actually the cpu is definitely the bottle neck when using a 4090 in MSFS VR, in my experience . I have a 7800x3d and a 4090 using a Vive pro 2 on highest settings and no matter what graphical setting I change it makes little difference in frame rate, but if I lower traffic I get a huge boost, which is all CPU based calculations. Very Poor CPU optimization unfortunately.
@@numlock9653 Vive 2 pro Res is very low
@@numlock9653huh, well you must've either clogged up that CPU with poor bios settings, or your Vive 2 Pro just has subpar resolution.
@@numlock9653 get a better VR headset
Until it does triple 4K at 240Hz without DLSS or frame generation, there's always room for more. :) Whether there's enough of a market for that outside of the sim-racing and flight-sim niches, I don' tknow, but I'd buy one.
Thats not happening ever with new features keep coming out like RT and PT
Battlemage will do this.
@@fcukugimmeausername You are clearly out of your mind
8k 144hz will come out soon definitely need more power. Never enough.
64k even
Alex said especially with RT it'll be CPU bound, because of the BVH workload. But what the next generation could offer is moving the BVH maintenance to the GPU alleviating a whole bunch of CPU work.
Npu s will also come to rescure either on gpus or cpus
@@Hi-levels I don't think that has anything to do with RT's BVH workload, the only thing an NPU could do better is AI denoising, but nvidia already has dedicated hardware to do that on RT cards, and I'm pretty sure an NPU would introduce too much latency for realtime denoising as a separate piece of hardware.
RT is so GPU intensive that I don't see how it could CPU limit games.
If I activate RT in games, CPU load tends to lower because GPU is struggling.
I have i5 and 3080
Il ride my 4070 Super until next gen consoles drop. Thats when the jump to ai NPU's for gaming and other new features will push me to upgrade. Honestly its overkill for 1440p in the games i play at the moment.
i got a 4070 super also and its great for 1440p oled gaming. But im going to upgrade to a 5090 and pass the 4070 super to the wife and get rid of her 4060ti rig. 4070 super should last a while with no problem atleast 3 years more before you start seeing it start to struggle a bit
GPUs are getting faster at a pace that CPUs can't quite keep up with. I have an RTX 3090 paired with a Ryzen 7 5800X 3D and in games with heavy ray tracing like Spider Man Remastered, my 3090 is CPU limited even at 1440p and to some degree at 4K. I am itching a CPU upgrade right now way more than a GPU upgrade, my 3090 is still fine. I am looking to upgrade to a 9800X 3D when it comes out.
Stupid methods of rendering that move too much data around, need to rebuilt complex data structures all the time, ...
It sounds like the game engines aren't designed 'properly'... Can't the (RTX) GPU handle most of the processing necessary for high quality ray tracing? I didn't realize that even with RTX that Ray Tracing was most often restricted by CPU performance! Isn't the GPU supposed to be doing all the additional Ray Tracing processing?
@@StreetPreacherr There is a lot of intermediate steps that need a lot of CPU, in case of raytracing it seems the spatial subdivision is one of them for example. Uncompressed assets, assets with excessive detail, poor code parallelism... are other examples of causes of bottlenecks in general.
That’s because the Ampere GPU’s handled data weird. The 4070 Super is way faster than a 3090 in 1080p and faster in 1440p as well.
Spider-Man is a very specific game, it is heavily CPU-limited (but with high fps, not like DD2), but it is an exception. It is hard to find visible CPU-bottlenecks in most games in 4k, with 4090 and top CPU (or even new mid or smth like 5800X3D).
If we're talking about RT I think it heavily differ from game to game. It's very CPU demanding in S-M:R or especially Hogwart's Legacy, but not at all in CP2077.
Ampere didn't have price cuts of any kind, it maintained the price hikes of Turing.
Considering the 5090 will be $1700 at least, it needs to be twice as fast as the 4090 to be worth it.
Rumor is it's 70%.
twice? 🤣🤣🤣🤣
70% raw performance improvement, along with a better version of frame gen to boost frame rates even more than the 40 series version would be worth it.
that's a weird argument.
I don't think so, it likely will be above $2000 due to them using the same node as Ada and Hopper, their data center B100 is already twice as large as the H100 in order to have a decent performance uplift on the same node, so twice as large on the same node should mean twice as expensive, especially when they still aren't using chiplets yet.
I have an RTX 4090 the only way you would see me with a 5090 is if we can get 4K visuals with RT and can have at least 60 FPS rasterized performance no frame gen or DLSS.
You're the 0.1% who upgrade top end GPUs each generation
What games do you play I can do that already with my 4090?
@@ZackSNetworksame but 4080
never going to happen in the next 3 years. RT takes about 100x more computational power than rasterization does & most RT relies on rasterization as a base. Expect Blackwell to only be 20%-30% faster than a RTX 4090 at most. unless Blackwell comes out with a massive 224 R.O.P.'s ore more it isn't going to be that fast.
dlss set to dlaa is where it's at though imo. Better than TA AA native resolution.
The thing I'd be hoping for on a RTX 5090 would be two 8 pin power connectors and a return to more sane power consumption. More/faster VRAM would be nice too as I feel that that has been the RTX 4090's bottleneck, especially at 4K.
Other grab bag of features would be the return of nvLink/SLI support to scale up via multi-GPU and integrating some Thunderbolt 5 controllers. It'd be nice to be able to plug a USB-C monitor directly into the GPU without external cabling and get full USB support on the display and other peripherals connected to it. Similarly with Thunderbolt 5, it'd be clever to include a mode where you could use the GPU externally for a laptop without the need for a Thunderbolt bridge board in an external chassis. Literally just the card, power supply and a power switch to turn it on. The PCIe slot connector would go unused.
damn u called it
Thankfully AMD have opened a door for us with X3D. CPUs finally start becoming useful for gaming with ultra high speed memory (e.g on-die cache).
it's all about VR.. 50x0 series will be the first that's really capable of full immersion hq hi-res VR.
Yep. PCVR is where these overkill cards will shine.
Once you hit certain quality-experience performance thresholds in gaming, it doesn't matter & becomes overkill for the sake of giggles. It only matters to professional bench markers & internet braggarts.
Not if it burns your whole house down.
Cant they design a better connector instead ?
Can you have more strength than a 12yr old virgin and properly seat the connector?
@@ProjectMore69 if you think it's force related, you need to watch/read more on the topic. GamerNexus did a full investigation on these connectors, and clearly demonstrated how they are flawed by design.
@@ProjectMore69 Has this ever been a problem in the past for you? Are you speaking from experience?
@@rambo9199 deflect more virgin boy.. wanna see a video of my 4090 ive had since launch working perfectly fine? I bet you dont 😂
@@ProjectMore69what if you move your pc and it comes Auth a bit hmm ? It is not like you open the case always to see ther is always a cance for it and most of all even if u put it all the way in It can still burn 🔥 ther is never a 100 that it won't 😂
2000 series NEW tech and NEW features
3000 series nothing changed besides getting a little bit faster, no new features
4000 series NEW tech and NEW features
5000 series will be 3000 series all over again
6000 series NEW tech and NEW features
The current 4080 and 4090 are already bottlenecked by current cpu’s, hyper threading and e cores are total bs
"Hi guys exclamation point"
LOL Dafuq! That was hilarious.
U know what's gonna happen? Nvidia be like, so guys can't keep with us huh? We gonna design our own CPU! That's _exactly_ what happened to Intel. Apple be like, get you $hit together or we going our own way. Now we have M1/2/3 and soon 4. Watch Intel/AMD stocks TUMBLE when Nvidia announce their own CPU!!! AMD/Intel better get their $hit together.
The CPU side is interesting cause CPU limitations seem to be universal to all modern CPU. Like pretty much every major release will either run perfectly fine on the now iconic 3600 or if it's struggling on that it's struggling on everything. Havinh a faster CPU with more threads seem to not make much of a difference at this point. Most modern CPU don't look in any inherent danger of being underpowered for the job. It's just game developers struggling to get to grips with the multithreading and that affects all modern CPUs pretty much just as badly.
*It's just game developers struggling to get to grips with the multithreading..."*
No, it's just that ever since the Great Consolization of 2008 all AAA games have been designed and coded for console HW.
Current consoles have only 6 cores available to games. But further, those cores are considerably less powerful than even the cores of an 11-year-old 4770K.
@@bricaaron3978 cap, ps5 and new xbox both have cpu's equivalent to a ryzen 7 3700x. it features 8 cores and 16 threads. idk where you got your info but its wrong my guy
@@bricaaron3978running at 3.8ghz
@@orlandoluckey5978 *"cap, ps5 and new xbox both have cpu's equivalent to a ryzen 7 3700x. it features 8 cores and 16 threads. idk where you got your info but its wrong my guy"*
I repeat: Both the PS5 and the XBox Series X have only 6 cores available to games, just like the PS4 and XBox One.
Each of those cores has a significantly lower FLOPS than a 4770K from 2013.
Just get a G9 57" monitor or any other 4K+ monitor. You'll instantly be GPU bound again.
I love questions like this because they assume on some level that large companies (Nvidia, AMD, Microsoft, Intel etc) wouldn't be thinking about bottlenecks or be actively working on developing new ways to utilize cutting edge hardware.
Honestly just stop it with this CPU nonsense. Unless you are competitive gaming at 120-240hz, a 3700X will be fine for 60fps for another 4-5 years.
Its not nonsense, saying a 3700x would be fine for another 5 years is crazy when it bottlenecks the latest gpus now on high end gpus even at 4K.
Not everyone plays on 1080p/1440p.... and only 60 FPS??? you're kidding me!
At this point, graphics cards are a side-hustle for NVidia's primary business: AI infratructure.
People getting GPU limited in 95% of their games: ...
The same people getting CPU limited in the remaining 5%, of which only 1% actually results in a sub 100fps framerate: OH NO I'M CPU BOTTLENECKED!!
There are ways to spend the GPU extra power, like on monitors of higher resolution and refresh rate. And if you care much about a specific game where CPU bottleneck is a thing, then invest more in CPU and less in GPU. Not hard, giving high tier GPUs cost multiples of high tier CPUs.
Yeah pretty much. CPU rarely bottlenecks unless you're playing an RTS.
Not to mention that 4090 and 5090 are 4k cards, even older cpus rarely bottleneck the 4090 on 4k.
Honestly maybe a slight increase in performance from a 4090 but with much more efficiency would be awesome. Imagine 4090 level performance at only 200 watts for example, it would be crazy!
Dude then just get a 5080 then. The 5090 will have %60 more rasterization performance and 2.5x better raytracing than the 4090. While pushing the same amount of power. It’s called “performance per watt” not “low watt output”.
@@ZackSNetwork im not going to get either, i was commenting on what i think would make a good product.
That might happen in 2 or 3 generations, not in one. The 4060Ti just reached the 2080Ti's performance under 200W but the 2080TI was a 350W card at most, not 450W.
@@lharsay maybe! Im just speculating here, the 4060 has around the same performance as the 3060 but uses 50% less power
Imagine undervolting.. derp. My 4090 runs at 3ghz with an undervolt and in most games im 300 watts or less.. peaks are 350ish.
if the 5090 is bottlenecked by current market cpus like the ryzen 9 79503xd or the i7 14700k / 14900k which cpu would bottleneck the 5090 the least
9800x3d
@andyakarudolfhessiansack7936
True. gpu get 50 to 70% delta gains cpu gets 5 to 10% with each succesion 🤪.
The question "Will RTX 5090 be too fast for any current CPU?" would not be asked by anyone who understands PC gaming.
The question implies --- and perpetuates --- the common misconception that there is such a thing as an _absolute bottleneck._ There is not.
Whether any component constitutes a bottleneck depends entirely upon what the user wants to accomplish. If a person prioritizes maximum framerate at the expense of resolution and image quality, the concept of a CPU bottlenecking a powerful GPU makes sense.
But if a person prioritizes image quality, the 5090 will not be powerful enough, and the concept of the _CPU_ being the bottleneck is ridiculous.
Let's not forget that Deferred Rendering engines do not support _working_ MSAA, so the only true AA available is supersampling (DSR, Frame Scaling, etc.). 4K at 4x SSAA; (i.e. 8K) _alone_ will not guarantee 60 fps; adding in ray tracing only exacerbates the issue.
And then there is VR, of course.
I just want to upgrade for GTA 6 😂
Rich is a gem.
I use an RTX 3090 right now and I would actually be interested in upgrading if they would start making mini versions of GPU's again. The 40XX series (and AMD's 7900 series) are all gigantic and since I build ITX rigs, they're just not interesting to me. But imagine if they used the smaller processor nodes to keep the performance the same and instead used the node advantage to shrink GPU's. One can dream.
Part of the reason why I went with the NR200 for my first itx build. I still remember seeing someone with a 4090 in theirs lol makes my 3080 ftw3 look tiny
My RTX 4090 GPU usage percentage is all over the place in 4K and in the Pimax Crystal (2x4K) with an i9-12900k.
It never stays at 99% or even 98%.
It usually varies from 100% to 80% constantly in 4K.
My RTX 3080 usually stayed at 99% to 98% in 4K.
In GPU-Z I’m now limited by Voltage and the limiting factor line is always blue instead of green. Even when I overclock my P-Cores to 52 and E-Cores to 40.
My next PC upgrade may need to be the 14900ks.
I would honestly not mind if it would be too fast or even more power efficient with the new architecture, because I would mainly use the card like that for rendering purposes.
A Geforce RTX 3060 is right at home with an old i7-4790 especially when playing games in 4K so a 5090 would be right at home with either a 13th/14th generation i9 or Ryzen 9 7950X/7950X3D.
The problem is x86, no matter how low level your API is, you still have to run everything through the CPU due to how the PCI bus works. Basically we would need a bypass of some sort akin to what AGP did.
Stop ignoring people without monitors that use HMDs instead. These elitist 2D pancake POVs annoy me
Frame-gen is not a solution. It is a work-around. Those with a 5090 won't have to use it.
I think you mean shouldnt have to use it
DLSS is magic in a lot of cases. I’ve played many games now that look better under DLSS quality with sharpening at 100%.
Frame gen is specifically designed to boost performance in cpu limited games. Has little to do with lack of graphics performance.
@@numlock9653 it is not little. Asset quality is on the GPU. With frame-gen you can have those higher assets render at a lower frame-rate and have frame-gen fill in the gap.
@@numlock9653 you don’t know what you’re talking about clearly. DLSS is much more than that
I have no idea why they are talking about 1080p, it is a GPU designed and designed for 4k or more and in those resolutions it is very clear that it will not have problems with current generation processors even with its mid-range whether it is 7600 or I5, hi From Venezuela !!
The 9800x3d will be able to handle it. Nothing else will though.
I'm not too sure. It's rumoured to be only 15% IPC uplift from 78003xd. For flight sims in particular, its still gonna be cpu bound.
@@andyakarudolfhessiansack7936 its overCLockable and it has 16 THREADS THAT WILL BE 15% FASTER
If they rush a 2024 launch, I will wonder more about 4090 owners needing to upgrade. Normally, there is some big, melts your PC (no pun) game that calls for the upgrade, but the only one I can see that might gain from a 5090 is Stalker 2. I would have thought a mid-2025 would have been better.
all new games suck so who cares about running a fast computer, 1080p is fine when all new games suck anyway
Implementation of the of yhe most upto date DP standard can help in this utilisation with higher resolutions available at higher frame rates
Cyberpunk 2077 with path tracing in VR using VorpX/Luke Ross with a render resolution of ~10000x5000 at 90+ Hz would probably break the 5090 Ti Egregious Super too. Maybe when 8090 TiTi Super Duper comes around.
Enough now of this, it’s becoming click bait
$2500 for +15% max over 4090.
I think people significantly underestimate how hard it is to do multithreading, which is the bane of existence for CPUs.
Honestly, if developers were even able, I'd really love to see GPUs be used for certain operations, like all the geometry collisions, rather than doing them in CPU. But considering the first point, I'm sure we'll never see a PC developer actually write CUDA (or whatever language AMD/Intel do for their cards) code to do that.
Unreal 5.4 is supposed to fix this. Watch the presentation.
Games have been literally doing collision calculation in gpu since quite a while ago and there is a very big issue in doing so , GPU needs to access large amount of data to be given to it for the draw call to be efficient because that data again needs to be passed back to CPU and doing so for multiple objects seperately will result in poor performance just because CPU is doing nothing but passing around the data , the engine team and level designers in modern studios have become so much of a seperate entities that the communication between them is borked and most of such optimizations are very hard to perform. And yes most level designers have no clue on how compiler actually optimizes the code , or how to write code that is easier for compilers to optimize, or use compiled language at all .
The solution is hyperscalar CPU architecture, which would kill GPU market because if CPU was truly good at multithreading there would be no need for GPGPU anymore, so we will never see it for that reason.
@@doltBmB Never is a strong word. There will be a time when CPU performance will need to grow beyone what's possible now.
@@doltBmB I don't think there is silver bullet solutions in engineering, the architecture you mentioned may seem good at a glance, and I beleive industry will learn to work with anything if it's widely used. The problem arises with the cost , implementing very big isa for every single CPU core in thousands of cores seems very expensive deal to me .
CPUs only bottleneck at low resolution/high refresh rates. Just means that when it's time to upgrade the CPU, I won't have to worry about updating the GPU (future proofing).
Nice conversations, but did they really answer the question? I think the user was asking that how can the most powerful consumer graphics card money can buy really enhance gaming experiences when we also need equally powerful CPU performance to make this happen which often causes choking on the GPU? I mean I can name a plethora of things video games are lacking today that REALLY need improvements to improve experiences compared to what we have had for the past decade or two, with some being physics in real-time(this will always need improvements), real-time water physics(this is EXTREMELY held back and limited because of limitations in CPU calculations), 3-D volumetric special effects(I believe this is GPU focused and CPU focused) as well as robotic NPC's that suck you out of immersion(this is heavily CPU needed as well), so point being we can have these powerful GPU's but if video game experiences don't improve we are stuck with the same mediocre gameplay with shiny polished graphics at 556 FPS. Also VR, to me at least, IS the absolute FUTURE OF GAMING, being able to actually transport yourself into these incredible worlds, and that requires much more powerful hardware both GPU and CPU. My hope though is that solutions are found to fix these concerns and allow developers to have more fun creating games and not be limited constantly by hardware. PS5 was really exciting when it released because of the way they saw SSD's bottlenecking a lot of freedom for developers, and they found a solution that even pushed outwards into the PC SSD market as well since most PC SSD's at that time weren't hitting speeds PS5 was because of I/O throughput issues. My hope is Cerny and his amazing team come up with bottleneck solutions for improving CPU performance, allowing more efficient multithread performance in games that will stretch out into the industry and become standard you know? And I imagine Ai will also assist in numerous ways with improving performance, user experience and helping developers make games more polished and advanced. SO exciting to think about...
I hope the 14900K and 14900KS can drive an RTX 5090!
They self degrade after 3 months of using, I don't expect them to drive anything in the future.
Get a 7800X3D. Similar performance, much cheaper and more energy efficient.
That's why we need an ARM/Risk architecture as soon as possible.
arm architecture exist already? what do you expect? check apple and snapdragon. risk just doesnt have enough muscle.
@@superertert Desktop arm based CPUs.
Those mentioned by you are still mobile architecture.
Unoptimized games. Our CPUs are insanely powerful nowadays. Force game companies to DO BETTER. That's the plain and simple truth, no exceptions, no twist, no nothing. Unoptimized games are 100% the sole issue here.
Doesn't telling windows to make the hardware scheduler the gpu itself get rid of that?
7800x3D is rarely being used above 50-60% in games using the 4090. (at 4k)
I dont think it will bottleneck the 7800x3D when it comes to gaming at all. Will still be an amazing CPU for it.
Imstead of Silicon why not opt for Graphite
There is no problem keeping a gpu bussi with pathtracing. More samples...more bounces...until we reach unbiased territory
games use to be gpu bound, over the past few years, now they are both cpu and gpu, impressive regression in developing games
VR Performance would be one thing looking at the like of devices like Pimax 8K and 12K HMD's sure can burn some GPU Time (and CPU), ofcourse AI - maybe AI api for gaming use and the bethought PhysiX return and ever more RTX (even in VR).
Improvements are probably gonna be modest, both in software and in hardware.
More interpolated frames is not bad, but it's in the marginal gains category.
I see a couple of fronts where they could make progress, though most of this probably won't come with Blackwell.
Generated predicated frames. Instead of interpolating real frames, the next is simply predicated. This get's rid of the latency issue.
A more integrated neural rendering approach.
Lot's deep learning powered NPCs in games.
What would you guys do in this scenario?
I have an old build which is a 8700k with 1080ti which has served me very well. However, I've gone into sim racing which i'm running triple 1440. The build holds up fine with obvious graphics turned down with iRacing, but other title are obviously a big no.
I have the funds to build current gen with a 5090 & 14700k. But with arrow lake and 50 series potentially close, am I better off waiting a bit longer?
I get that if I do it now, it's a massive upgrade. But what do the pros think?
Thanks
I’d be most interested in how much quicker 5090 renders in blender than the 4090. My 3090 still takes forever in highly detailed projects.
directx13 will prob relive cpu bottlenecks
No it’s not, gaming in 4K with Ray Tracing even with DLAA or DLSS is the most beautiful graphics I’ve ever seen and that’s just with my 3090. With a 5090 I can turn off DLSS and go Native and still get spectacular FPS.
No, games are just badly optimized
Exactly
Best current pairing for a 5090 will be the 7800x3d. its only like 300 bucks and the am5 platform is gonna be around for a few more years, we'll get a more compatible one at some point in the future, no use wasting the extra 300 on a 7950. or wasting 700 on having to replace an intel board and cpu.
The optical GPU is just around the corner. So is the optical CPU 10K times faster than we have now.
When you use DLAA plus ray tracing at 4K in games like CP2077 and Alan Wake, I don't think the CPU will be limiting you. Plus there's the future proof aspect, games tend to get heavier in terms of graphics as a generation goes on, CPU requirements usually stagnate.
Dlaa isn’t the best picture quality objectively, For now cyberpunk ( i think ) you can’t use ray reconstruction without Dlss correct me if I’m wrong
It is unlikely that even the upcoming Intel Arrow Lake 15900K or AMD Ryzen 9 will be able to keep up with the RTX5090. Intel, in particular, will need to introduce its Novalake (16P core xxE core) with DDR6 RAM to support the RTX5090 ASAP .
I'd love to see less power draw - not gonna happen, lower price - not going to happen, thinner - not gonna happen so that leaves about 4 times the ray tracing performance - probably not gonna happen.
No way i have a 4090 paired with an Athlon 3000g and the gpu is killing my performance
I'm getting ready to add a new CPU to my 4090 before adding a 6090 to my next CPU.
If no one can afford it, does it really matter?
Can't wait for games to be even less optimized and just throw a 5070 ti at the recommended specs
My guess is that the 5000 series will be all about AI. We've only scratched the surface of what you can do with AI. Even something "simple" as running a local LLM can max out the highest end GPUs.
The problem still remains lack of multi threaded optimization in games. Modern CPUs just cant fix the fundimental nature of games being programmed to rely highly on a single thread . It would be in Nvidias best interest to explore the use of AI to potentially solve this issue. Else there is little reason to scale much beyond a 4090 without a miracle in single threaded cpu performance.
As someone with no interest in resolutions above 1440, and can personally perceive no benefit from frame rates above about 80… none of this bother me… at all.
well the 9800x3d should be releasing soon once the 5090 launches
its not enough. CPU tech is way behind GPU tech
I think there will be more demanding games that could potentially make the 5090 work 100%. Also, when we think of 4K and the new high refresh rates, then a 5090 has a lot more use case scenarios. At least, with a 3090, I can barely do 60fps in modern games while playing in 4k. I think a 5090 would definitely push it to 120fps before my Ryzen 5900x has any trouble. And it's not even a gaming CPU. Some of the new generation x3D ones from AMD are a lot more powerful than it in games. Obviously, not all games require such powerful GPU, but it will not be more powerful than what we can take advantage of.
Literally ANY card is too fast for current CPUs if you lower the resolution far enough. Even the 3090 struggles at 4k maxed especially with RT/PT and without using fake resolutions/fake frames. Native maxed out PT rendering at 8k or 16k DSR on an 8k monitor/TV? Forget about it. 1 FPS maybe, if the PC doesn't explode and take the neighbourhood with it.
' 4K screens coming online now... ' rofl as if they haven't already been released by a year or two.
CPUs are lightyears ahead against GPUs on 4k res and above (VR).
If its way ahead of the cpus thats good cause it'll push cpu makers to catch up.
AAA titles at 4k with rt on will be the sweet target for 5090
now the 4090 IS a beast, no doubt. But its not like it does 4K 400FPS every game out there. And while you can argue that there realistically arent even displays for that type of fidelity, I'd also argue that the 4090 isnt always CPU bottlenecked. The same should apply to the 5090, once new CPUs come out as well...we just need that kind of power in the mid range and in some gens, the low end.
I don't think bottlenecking will be a problem because by the time the 5090 is out, Intel's Arrow Lake CPU's will hit the shelves more or less at the same time. And of course AMD will release their next gen CPU's in response.
I don't even think anybody even has to ask this question, the answer is 100% yes. The 4090 won't get utilized with a 14900KS or 7800X3D in some games at 4k, even the games where it is at 99% the CPU doesn't have much headroom left, it'll maybe push a 5080, but a 5080Ti/5090/5090Ti or anything faster has zero chance of being useful for 4k until devs start optimizing their games.
No joke, just cashed in about 1,000,000 Marriott points for over $3,000 in Best Buy gift cards 😂 getting ready for the 5090
How muc is a game gonna bottlekneck on a 3950x? Can a game really bottleneck with 32 threads?
Frame gen is like margarine gaming.
I like real butter.
Oh, I'm sure I can make a 5090 choke with PCVR.