I appreciate how you take time to help your audience understand these issues, and how to properly understand reporting data. Another great video and I hope your channel continues to grow as you deserve it 👏
Interesting topic. The same issue happens with Ratchet & Clank. When you change resolution/some settings (texture quality, for instance) or enable/disable upscaling, the framerate may drop significantly for no obvious reason. I think the way these two games manage VRAM is not correct. They just fill VRAM with new data and do not free it from the older one. And both of these games are certainly very hungry in terms of how much video memory they need. Funny thing - these games, as well as Horizon Forbidden West, need 9-12GB of VRAM at Very High preset (depending on the resolution), but they use only 5-7GB of RAM. While usually games use at least as much RAM as they need VRAM, if not more. So it looks like Nixxes latest PC ports heavily rely on VRAM capacity and do not take advantage of RAM.
hey there, game suffers from extreme VRam allocation problems!!! game wont erase textures and etc. when fast traveling resulting in crash after multiple fast travels and climbing up VRAM usage! goes from 9-10GB in 4k DLAA to 24GB after 10-12 fast travels!!! it seems like the game wont erase the allocated textures when you leave an area!
In rdr2 my 6800xt uses around 15gb with 4k ultra and with some texture mods☠️ so for me 16gb vram is minimum I don't want weird texture pop-ind that looks really bad earlier I had a 3050 4gb laptop in that pc I use to get texture pop in problem or texture not loading main problem I was getting was in horizon zero dawn at 1080p but this time I got a hell 4 times more vram that my previous GPU
It can depend on the game. Like in Forspoken for example if you dont have enough vram you dont really run into any performance issues but the texture level of detail drops to a much lower resolution. RDR2 can be vram hungry from what I remember though.
What mods you are using its insane even on 4k ultra in RDR2 to use over 8-9 vram. And yes for HZD for City you needed more than 4vram even on low 1080p. As Terra said it depend on the game RE4 remake could cause even bsod with not enough vram on higher settings only game i experienced that. Before most often problem with lack of vram would be texture loading, but now with newer games its perfomance more often problem. Far Cry 6 would drop fps after some time when you dont have enough of vram on higher settings, but no texture problems would happen
@@RonniSchmidt-mi7pd Some areas can be a bit hungry on vram and on AMD it uses a bit more vram. I did a side by side comparison of the built in benchmark here it's what I am remembering. th-cam.com/video/nqt4YXT3NQY/w-d-xo.htmlsi=xZmxXx7ISS_snop9
I'm also experiencing these vram issues on a 3080 10gb at 1440p when changing certain settings. The vram does not unload but keeps piling up. The frame rate does eventually go back to normal once i change the shadow setting to low. It then manages to unload the vram.
Right, I've heard from a couple others with 10Gb 3080's of running into vram issues. My recommendation would be to dial your settings and not change them. Kinda crazy that 10Gb isnt enough for 1440P in this game.
9:24 CLEARLY skill issue, it's meant to be played as a slow motion ultra cinematic experience obviously . Great video man, these VRAM deep dives really help me personally now that I'm gonna be in the market again for a new gpu. Really curious how the new GDDR7 memory will fair relatively to the amount of each gpu and the requirements of next gen games... Also 3080ti in general seems to fair extremely well with new titles etc.
Yeah the 3080Ti has been great. On of the reasons why I wanted to keep it is because of that 12GB to see how it'd age. And I'm kind of a hoarder when it comes to PC parts lol
for your next game please also add 1440p dlss quality but not maxed out settings. maxed out often cost too much performance compared to high even though the visuals often are almost the same. so 1440p dlss quality high settings without framegen would be a gread addition as this would be to me my personal real usecase and i believe so to many
it seems like a game bug. Also small correction about performance when reaching vram insufficiency. It really depends per game and the severity of it. Some games will just perform slightly lower average fps, some will start stuttering when looking around and when its severe your fps goes to like teens, basically what we see here, hopefully nixxies will fix this bug.
Yeah true, sometimes you just get really bad 1% lows, inconsistent frametimes but it's usually a noticeable difference once you go over that threshold in many cases. Than you have some games like Forspoken that will automatically lower texture resolution to a degree.
Some games also fails to actually load textures or load lower quality ones. Examples of games that do this are Forspoken, Hogwarts Legacy and Halo Infinite.
Very well made video, explains why I have seen some benchmarkers completely close the game and change settings via the launcher. I believe most of the recent Nixes games have similar issues at launch but they get patched quickly. I hope the crashing and vram issues are both fixed soon. I also feel like the VRAM requirements in general are quite high for a game that started its life on the PS4 and that's what most of the assets would be made for. From my personal experience it doesn't matter if you have an AMD card with these issues, if the coding is not done properly and the game doesn't flush its ram/vram cache, then you get these issues. Playing Jedi Survivor now on the 7900XTX and I see that game starting with around 20gb vram but as I play for an hour or so while traveling to different places, the VRAM pretty much reaches 24gb. The system ram also takes up around 20gb I think so overall it's using 40gb+ of system ram and VRAM, it just makes no sense. You have to wonder what kind of magic optimization they did to make these games run on 6GB on PS4 or 12-14GB on PS5, but using double or triple that on PC, even if the settings are higher.
I think the crashes have something to do with the latest AMD driver. I installed this game and played it without crashes. Then came the option to update to latest AMD driver. I installed it and since then the game would randomly crash. I went back to an earlier AMD driver and the crashes didn't happen again. So it was something to do with the lastest driver, imo.
I would like to see a video in which all games so far that vram in 3080ti was not enough Ratchet and clank, avatar And even setting dlss/fsr quality at 4k wouldn't save us
Makes sense AMD runs this game way better than Nvidia this happens in more than one game when maxing out settings other games the fps stays the same but then you have issues with textures loading and pop ins DLSS doesn't fix this issue at all. Can't max out textures in games on Nvidia cards.
Have you tried fast travelling multiple times ? And also, doing that with DSR on ? Because for me the "memory leak" comes from that, i dont touch the settings but after some fast travels it starts to drop to 10fps like that
@@TerraWare I mean, I am not 100% sure but I can't see any other explanation for that kind of behavior. The games starts with like 8G allocated and then after some time and some fast travel it's like 11G+. And it doesn't come down for me until I restart the game. The thing is I don't know if they acknowledge anything or if they plan to fix it.
On a 4090 maxed at 4k I never see more than 14gb vram usage and it’s usually in the high 13s. I keep seeing reports of a leak but I never see VRAM creep up higher and higher until a crash or stutters.
I've heard some talk about memory leak but I didn't really see that. Played on the 3080Ti for close to 2 hours at 4K. I seen memory go up a bit but than also going down. It was very smooth. The issues was when disabling and enabling dlss and FG like I showed here.
With how often VRAM tends to be discussed these days online I feel the example I've provided in this video (starting at 6:45) I feel is a pretty good one to maybe understanding it better and what happens when you run out out and spill over to system ram. In most cases it can be avoided, in some cases its due to lack of code optimizations. Either way it's interesting to take a look at and discuss.
Raster performance can vary game to game. Also AMD works intimately with Sony---via designing their CPU and GPU for their consoles---so I'm sure the game design simply favors AMD each time a port is made.
I literally showed you the 6800XT's performance in this video. And no the base 3080 is not better than 6800XT. Neither is the 3080Ti, not across the board. There are always games where one architecture is favored over the other. I always show evidence of what I talk about in my videos.
@@TerraWare Of course a 3080/Ti won't be much better than the 6800 XT accross the board when there are games that love AMD cards like Forza or COD for example. That disrupts metrics and overall percentages. The 3080 Ti is usually faster than a 6900 XT in raster and performs better at 4k despite having less vram. In raytracing the Nvidia cards are way above the AMD counterparts.
Dont have a 7800XT but have a 6800XT which is basically the same thing far as performance. At a glance the 3080Ti and 6800XT are more or less evenly matched here, should be the same for 7800XT and 4070 Super.
@@TerraWare but yeah you're right tho. From what i've Heard cause i had neither of these 2 gpus when they got release, but from what i've Heard 3080 was competing with 6800xt at first. But now as we see 6800xt aged better competing with 3080ti even in certain games. This îs why i think driver maturity plays a role with AMD it's a known fact u get extra % from drivers in time. It happened again question îs will 7900xt fightn4080s, 7900gre 4070tiS, 7800xt 4070ti
I think XeGTAO looks better. At least the grass has more realistic occlusion, as far as performance I've cycled between the two the FPS didnt change. Tru it next to the gras.
XeGTAO costs 6% at 4k, vs 4% for HBAO+ You are correct. It looks better and can cost more in scenes with tons of grass and shading. As far as "maxing" out goes, XeGTAO is the right call for visuals as well
Yeah that's architectural and node advantages for ya. I remember when I got my 4090 almost 2 years ago and it was twice as fast in ray tracing and like 70% on raster while drawing simlar power sometimes less.
Do you have the 10GB? If so I can see that happening looking at my own numbers. I don't think there's a memory leak or anything as I played on the 3080Ti for 2 hours but there were points where memory would go a hit above 10gb actual use. Are you getting drops to the teens like I showed it this video?
Yup I have the 10gb. I forgot to say I play at 1440p. The 10gb at 1440p doesn't seem to be a problem first glance for me. I play at max settings with dlss quality and fsr frame gen both on. And I get over 180fps when I just boot the game, it stays very smooth for a while. But over a few hours, (I also tab out sometimes maybe this could affect) the frames just starts to drop steadily, until it reaches 60fps and starts to stutter. Then I reboot the game and everything goes to normal again. I also checked temps and my 3080 never went above 70celcius even after 4 hours. I also tried turning off dlss and frame gen like in the video and I get about 80fps on boot like what you got. Hmmm, yeah not sure. I am not the best at troubleshooting. Wish could be of more help
@rohank2432 Frame gen tends to use more vram generally. I would just dial in your settings at 1440P with DLSS to quality and don't use frame generation. Try playing that way without frame gen I think it's the best way to minimize running out of vram. That sucks.
I appreciate how you take time to help your audience understand these issues, and how to properly understand reporting data.
Another great video and I hope your channel continues to grow as you deserve it 👏
I have VRAm related crashes with my 3080ti playing 4k ultra
Once again, we can see that FSR can be really good if implemented properly. I can't wait for version 3.1.
Why bother fsr when dlss is an option for rtx cards
@@gejamugamlatsoomanam7716 did I talk about RTX cards?
Interesting topic.
The same issue happens with Ratchet & Clank. When you change resolution/some settings (texture quality, for instance) or enable/disable upscaling, the framerate may drop significantly for no obvious reason. I think the way these two games manage VRAM is not correct. They just fill VRAM with new data and do not free it from the older one. And both of these games are certainly very hungry in terms of how much video memory they need.
Funny thing - these games, as well as Horizon Forbidden West, need 9-12GB of VRAM at Very High preset (depending on the resolution), but they use only 5-7GB of RAM. While usually games use at least as much RAM as they need VRAM, if not more.
So it looks like Nixxes latest PC ports heavily rely on VRAM capacity and do not take advantage of RAM.
Yeah Ratchet and Clank was the same way. That was the first game I noticed I ran into VRAM issues with on the 3080Ti at 4K with RT enabled.
The programmer literally says "fuvck poor gamer, we can't do memory allocation and deallocation correctly, too much work"
hey there, game suffers from extreme VRam allocation problems!!! game wont erase textures and etc. when fast traveling resulting in crash after multiple fast travels and climbing up VRAM usage! goes from 9-10GB in 4k DLAA to 24GB after 10-12 fast travels!!! it seems like the game wont erase the allocated textures when you leave an area!
Yeah that's what is happening to me. Since you mentioned 24gb I suppose you have a 4090. Are you using Dynamic Resolution Scaling ?
In rdr2 my 6800xt uses around 15gb with 4k ultra and with some texture mods☠️ so for me 16gb vram is minimum I don't want weird texture pop-ind that looks really bad earlier I had a 3050 4gb laptop in that pc I use to get texture pop in problem or texture not loading main problem I was getting was in horizon zero dawn at 1080p but this time I got a hell 4 times more vram that my previous GPU
It can depend on the game. Like in Forspoken for example if you dont have enough vram you dont really run into any performance issues but the texture level of detail drops to a much lower resolution. RDR2 can be vram hungry from what I remember though.
@@TerraWare no.
i play rd2 maxed with 3080 and dont even get vlose tpp vram
What mods you are using its insane even on 4k ultra in RDR2 to use over 8-9 vram. And yes for HZD for City you needed more than 4vram even on low 1080p. As Terra said it depend on the game RE4 remake could cause even bsod with not enough vram on higher settings only game i experienced that. Before most often problem with lack of vram would be texture loading, but now with newer games its perfomance more often problem. Far Cry 6 would drop fps after some time when you dont have enough of vram on higher settings, but no texture problems would happen
@@RonniSchmidt-mi7pd Some areas can be a bit hungry on vram and on AMD it uses a bit more vram.
I did a side by side comparison of the built in benchmark here it's what I am remembering.
th-cam.com/video/nqt4YXT3NQY/w-d-xo.htmlsi=xZmxXx7ISS_snop9
In order to reset the GPU on your system, you simply have to press Windows, Ctrl, Shift and B keys together at a time on your keyboard.
I'm also experiencing these vram issues on a 3080 10gb at 1440p when changing certain settings. The vram does not unload but keeps piling up. The frame rate does eventually go back to normal once i change the shadow setting to low. It then manages to unload the vram.
Right, I've heard from a couple others with 10Gb 3080's of running into vram issues. My recommendation would be to dial your settings and not change them.
Kinda crazy that 10Gb isnt enough for 1440P in this game.
9:24 CLEARLY skill issue, it's meant to be played as a slow motion ultra cinematic experience obviously .
Great video man, these VRAM deep dives really help me personally now that I'm gonna be in the market again for a new gpu. Really curious how the new GDDR7 memory will fair relatively to the amount of each gpu and the requirements of next gen games...
Also 3080ti in general seems to fair extremely well with new titles etc.
Yeah the 3080Ti has been great. On of the reasons why I wanted to keep it is because of that 12GB to see how it'd age. And I'm kind of a hoarder when it comes to PC parts lol
for your next game please also add 1440p dlss quality but not maxed out settings. maxed out often cost too much performance compared to high even though the visuals often are almost the same. so 1440p dlss quality high settings without framegen would be a gread addition as this would be to me my personal real usecase and i believe so to many
no one believes it but its the game{newer games)..not the card..my humble opinion..Vram Ghost of Crashshima...lol
it seems like a game bug. Also small correction about performance when reaching vram insufficiency. It really depends per game and the severity of it. Some games will just perform slightly lower average fps, some will start stuttering when looking around and when its severe your fps goes to like teens, basically what we see here, hopefully nixxies will fix this bug.
Yeah true, sometimes you just get really bad 1% lows, inconsistent frametimes but it's usually a noticeable difference once you go over that threshold in many cases. Than you have some games like Forspoken that will automatically lower texture resolution to a degree.
Some games also fails to actually load textures or load lower quality ones. Examples of games that do this are Forspoken, Hogwarts Legacy and Halo Infinite.
@@MrMeanh yeah that also👍
Unplayable, i stop touching the game and wait for the fix later. Maybe next month
Very well made video, explains why I have seen some benchmarkers completely close the game and change settings via the launcher. I believe most of the recent Nixes games have similar issues at launch but they get patched quickly. I hope the crashing and vram issues are both fixed soon. I also feel like the VRAM requirements in general are quite high for a game that started its life on the PS4 and that's what most of the assets would be made for.
From my personal experience it doesn't matter if you have an AMD card with these issues, if the coding is not done properly and the game doesn't flush its ram/vram cache, then you get these issues. Playing Jedi Survivor now on the 7900XTX and I see that game starting with around 20gb vram but as I play for an hour or so while traveling to different places, the VRAM pretty much reaches 24gb. The system ram also takes up around 20gb I think so overall it's using 40gb+ of system ram and VRAM, it just makes no sense. You have to wonder what kind of magic optimization they did to make these games run on 6GB on PS4 or 12-14GB on PS5, but using double or triple that on PC, even if the settings are higher.
PS4 has more VRAM like 16GB (although its different than PC VRAM). That could be a reason.
I think the crashes have something to do with the latest AMD driver. I installed this game and played it without crashes. Then came the option to update to latest AMD driver. I installed it and since then the game would randomly crash. I went back to an earlier AMD driver and the crashes didn't happen again. So it was something to do with the lastest driver, imo.
I get comparable FPS with a 6800xt + 5600x cpu FSR on/off. Yours seems to be 10% higher sometimes. Really happy I got a deal on a 6800xt for $350.
I would like to see a video in which all games so far that vram in 3080ti was not enough
Ratchet and clank, avatar
And even setting dlss/fsr quality at 4k wouldn't save us
maximum possible settings frame gen on/no dlss scaling, 120-150 fps @1440p. i9-12900ks/4070 super, about 10-11 vram. seems reasonable.
Makes sense AMD runs this game way better than Nvidia this happens in more than one game when maxing out settings other games the fps stays the same but then you have issues with textures loading and pop ins DLSS doesn't fix this issue at all. Can't max out textures in games on Nvidia cards.
Have you tried fast travelling multiple times ? And also, doing that with DSR on ?
Because for me the "memory leak" comes from that, i dont touch the settings but after some fast travels it starts to drop to 10fps like that
I haven't tried fast traveling but have got a few comments like yours so it seems to be a memory leak issue happening.
@@TerraWare I mean, I am not 100% sure but I can't see any other explanation for that kind of behavior. The games starts with like 8G allocated and then after some time and some fast travel it's like 11G+. And it doesn't come down for me until I restart the game.
The thing is I don't know if they acknowledge anything or if they plan to fix it.
@@GalaxyXYZ888 Many of these Nixxes ports seem to be quite vram hungry. Not sure why that is.
On a 4090 maxed at 4k I never see more than 14gb vram usage and it’s usually in the high 13s. I keep seeing reports of a leak but I never see VRAM creep up higher and higher until a crash or stutters.
I've heard some talk about memory leak but I didn't really see that. Played on the 3080Ti for close to 2 hours at 4K. I seen memory go up a bit but than also going down. It was very smooth.
The issues was when disabling and enabling dlss and FG like I showed here.
Try fast traveling
Every Nixxes port on PC have vram issue
if it crashes try turning Anti-lag off.
With how often VRAM tends to be discussed these days online I feel the example I've provided in this video (starting at 6:45) I feel is a pretty good one to maybe understanding it better and what happens when you run out out and spill over to system ram. In most cases it can be avoided, in some cases its due to lack of code optimizations. Either way it's interesting to take a look at and discuss.
base 3080 is better than 6800 xt
you might aswell stop making videos, you sound clueless
Raster performance can vary game to game. Also AMD works intimately with Sony---via designing their CPU and GPU for their consoles---so I'm sure the game design simply favors AMD each time a port is made.
I literally showed you the 6800XT's performance in this video. And no the base 3080 is not better than 6800XT. Neither is the 3080Ti, not across the board.
There are always games where one architecture is favored over the other. I always show evidence of what I talk about in my videos.
@@TerraWare Of course a 3080/Ti won't be much better than the 6800 XT accross the board when there are games that love AMD cards like Forza or COD for example. That disrupts metrics and overall percentages. The 3080 Ti is usually faster than a 6900 XT in raster and performs better at 4k despite having less vram.
In raytracing the Nvidia cards are way above the AMD counterparts.
Curious 7800xt wirh latest drivers vs 3080ti. + A few other rt games:)
Dont have a 7800XT but have a 6800XT which is basically the same thing far as performance. At a glance the 3080Ti and 6800XT are more or less evenly matched here, should be the same for 7800XT and 4070 Super.
@@TerraWare i think 7800xt improved a bit from the first drivers.
@@TerraWare but yeah you're right tho. From what i've Heard cause i had neither of these 2 gpus when they got release, but from what i've Heard 3080 was competing with 6800xt at first. But now as we see 6800xt aged better competing with 3080ti even in certain games.
This îs why i think driver maturity plays a role with AMD it's a known fact u get extra % from drivers in time. It happened again question îs will 7900xt fightn4080s, 7900gre 4070tiS, 7800xt 4070ti
Your ambient occlusion isnt at max its on the intel one. Hbao looks better and is more demanding
I think XeGTAO looks better. At least the grass has more realistic occlusion, as far as performance I've cycled between the two the FPS didnt change.
Tru it next to the gras.
XeGTAO costs 6% at 4k, vs 4% for HBAO+
You are correct. It looks better and can cost more in scenes with tons of grass and shading. As far as "maxing" out goes, XeGTAO is the right call for visuals as well
i like how 3080ti draws 400 watts of power and shows 35% less fps then my 4080 on 260-280 watt consuption ._.
Yeah that's architectural and node advantages for ya. I remember when I got my 4090 almost 2 years ago and it was twice as fast in ray tracing and like 70% on raster while drawing simlar power sometimes less.
And on my Rx 6950xt I have 25% higher performance than RTX 3080Ti, with energy consumption at the level of 240-250W
@@charcan 6800 xt is the best for its money for sure. but 4080/4080 super doesn't 1200 anymore, just 1000-1050
I run it perfectly fine at 120hz 4k settings with reshade aswell. I have a 3080ti ftw3 @TerraWare
@@MaximusDecimus-nb3gk rx 6000 series doens't show memory power draw only the GPU die
I have rtx 3080, and I have same issue. The longer I play the fps drops lower and lower.
Do you have the 10GB? If so I can see that happening looking at my own numbers. I don't think there's a memory leak or anything as I played on the 3080Ti for 2 hours but there were points where memory would go a hit above 10gb actual use.
Are you getting drops to the teens like I showed it this video?
Yup I have the 10gb. I forgot to say I play at 1440p. The 10gb at 1440p doesn't seem to be a problem first glance for me. I play at max settings with dlss quality and fsr frame gen both on. And I get over 180fps when I just boot the game, it stays very smooth for a while. But over a few hours, (I also tab out sometimes maybe this could affect) the frames just starts to drop steadily, until it reaches 60fps and starts to stutter. Then I reboot the game and everything goes to normal again. I also checked temps and my 3080 never went above 70celcius even after 4 hours. I also tried turning off dlss and frame gen like in the video and I get about 80fps on boot like what you got. Hmmm, yeah not sure. I am not the best at troubleshooting. Wish could be of more help
@rohank2432 Frame gen tends to use more vram generally. I would just dial in your settings at 1440P with DLSS to quality and don't use frame generation. Try playing that way without frame gen I think it's the best way to minimize running out of vram.
That sucks.
I got a 7900xt with 20gigs of vram so that i don't have to worry about such things.
My guess for high vram usage is probably Nixxes way to solve direct storage problem that console use.
Yeah it could be. Maybe they are loading more assets in to minimize reliance on storage since it can vary from PC to PC. It's very possible.
Umm, this is a port of a ps4 game - a console with 8gb memory total and a 5200rpm hdd.
don't ever buy gpu that have 8 gb lower vram this days you will regret it
Nice video 👊