What is the best GPU, A6000 or the RTX 3090.The truth for rendering, Workstation GPU vs. Gaming GPU

แชร์
ฝัง

ความคิดเห็น • 1K

  • @nickcifarelli8887
    @nickcifarelli8887 3 ปีที่แล้ว +245

    I'm a 3Ds Max Arch Viz artist with V-ray and was watching this video, praying that you would not overlook the HUGE importance of VRAM capacity in poly-heavy complex scenes. Thank you for your review. Yes the A6000 smashed that render by half compared to the 3090, due, in a large part, to the 48GBs of VRAM. If the GPU doesn't have the VRAM necessary to load the scene with textures, poly counts, particle simulations, etc, it has to page file the textures on the program disk. This accounts for a far slower workflow. Also, with NVLink, you can essentially double your VRAM to 96GB (assuming of course you can afford two A6000s!) and that would allow you to load entire cityscapes, forests etc. HAVING SAID THAT, your point is very valid, for the large majority of content creators out there, performance per dollar, you can get 2 or 3 3090s (MSRP) for the price of a single A6000. So realistically, you have made a very valid point. I personally am still using a Ttitan xP with 11GBs and it cannot load heavy scenes with carpets and Vray fur. I would personally gladly upgrade to 2x 3090s then shell out for a A6000. You can set one as the dedicated interface and the other one for rendering. Nvidea caters to high, high-end post houses (ILM, Weta Digital, etc) with the Quadro series and hence the hefty price tag. Yes the drivers are ISV certified and the RAM is ECC, but for the average guy out there, or even for the more professional prosumer like myself, a A6000 is a pipe dream. Id be happy with dual 3090s tbh. Thank you for the video, and you got yourself a new subscriber.

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +35

      thanks Nick. I totally agree with you. As I have worked at ILM, I know that these houses use the workstations cards, but the sad fact is that even at the A-listed studios, not all artist need the pricy workstations GPUs, the bulk of CG work is not that complex. Set extensions, some simple 3D add and FX. But for these complex Avatar type scenes, they need the Vram in the A6000. and even then all the shots are comp at the end in Nuke. so they use layers.

    • @nickcifarelli8887
      @nickcifarelli8887 3 ปีที่แล้ว +9

      @@MediamanStudioServices Nuke is one of my favourite programs. Still, exporting all the render passes for Nuke to compose still requires a hefty GPU. But I am glad we are on the same page. Well said sir.

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +13

      Hi Nick, thanks for your comments. I agree that the A6000 is only for a smaller sub-set of users, but for 70% of all content creators a RTX3xxx card is all they need. As you stated, you are using a 11GB card and and the 3090 24gb would be a great upgrade for you.
      thanks for watching.

    • @blackgamingstudio5104
      @blackgamingstudio5104 3 ปีที่แล้ว +3

      Nvlink does not double vram that many people said are you sure nvlink double vram

    • @blackgamingstudio5104
      @blackgamingstudio5104 3 ปีที่แล้ว +3

      Is 96gb vram is enough for 4k character creation and 4k interior exterior designing in 3ds max and substance painter and red shift

  • @JetCooper3D
    @JetCooper3D 2 ปีที่แล้ว +22

    I work at Pinewood Studios UK and work for Marvel, Disney Lucas etc. We switch to Geforce card back on Star Wars ep7 and have never looked back. The Geforce cards are stable and good to go. Saved money can be routed to other hardware. Great advice to all and great video - subscribed. Thank you. (We use the new 4090 RTX in all of our workstations now / 3090's before).

    • @MediamanStudioServices
      @MediamanStudioServices  2 ปีที่แล้ว +3

      thanks for sharing your experiences with the channel

    • @The0zx
      @The0zx 2 ปีที่แล้ว +3

      Hi, Bro! Do you create 3d models for Marvel movies and Disney Lucas movies? Can you tell me about the computer specs you and your team are currently using? I dream of working on 3D models like you. But right now I don't know how complex the 3D model that I will make is. I need information about RAM usage capacity, processor, etc.

    • @goldenheartOh
      @goldenheartOh 2 ปีที่แล้ว +1

      ​@@The0zxis it still true Blender 3D is so optimized it can run on a potato? I used to have a similar dream 20yrs ago & Blender 3D was awesome. & I did have a potato for a pc.
      My point is I strongly suggest you get a feel for it on Blender before building a pc for it.

    • @dazrelixs
      @dazrelixs 11 หลายเดือนก่อน

      but you guys render locally or on the farm?

  • @wonderwonder9027
    @wonderwonder9027 3 ปีที่แล้ว +4

    I'm a civil engineer. First thing comes first and I'd like to congratulation you on the way you put the video together...... really straight to the point and full of just the important information instead of wasting time talking so much about things an average viewer won't understand......
    Second thing is that there is no decent reviewer out there that do tests -the same professional level that you do - on computer parts whither its GPUs or CPUs on engineering tasks..... I mean yes there are a lot of artists out there that would like to know how fast their art work will be rendered...... but there are -as far as I know - many professional work loads are not being tested..... I don't want to be very technical but how do these cards handle:
    -Matlap AI workloads
    -BIM applications like Revit and Robot and maybe Sap2000, Etap and Safe for structural analysis
    -GIS analysis and clustering
    -Analysis of aerial photo for environmental, hydraulic and structural purposes
    I mean yes there are some workloads for 3DMax and Maya and some video production that can put artists on the limits of what this hardware can and can't do, but the way I see it that they are very rare and in those rare cases they do have big studios and production companies: ie. can directly address Nvidia about technical stuff. But we the engineers are reaching those limits on a daily basis and we are the ones who really need to make educated decisions about the hardware we use.
    thank you for your time reading this.

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +2

      hi wonder wonder, thanks for the kind words.
      I worked on a project for Warner Bros World theme park in Abu Dhabi and was amazed on how well a lower end GPU could handle such complex geo in Revit. Wish we had this kind of viewport performance in Maya or Max.
      As for how the RTX3090 or A6000 will handle civil engineer type work loads. well I could not say. I have zero experience in these type of project and apps.
      you can check out this channel. maybe he can help
      th-cam.com/users/Tech3DWorld
      thanks for watching my channel

  • @graphguy
    @graphguy 8 หลายเดือนก่อน +9

    You said exactly what I wanted to hear.
    I play zero games, but I do alot of amateur work with Blender 3D and have been perplexed on going with a new RTX or a studio ready graphic card.
    thanks!

    • @rahulkamath6984
      @rahulkamath6984 8 หลายเดือนก่อน

      so what did you actually go with?

    • @graphguy
      @graphguy 8 หลายเดือนก่อน +2

      @@rahulkamath6984 haha decided to go to Italy for 2 weeks, then decide!

    • @rahulkamath6984
      @rahulkamath6984 8 หลายเดือนก่อน +1

      @@graphguy hahaha you don’t need any benchmarking to decide that I guess 😅

    • @noth606
      @noth606 7 หลายเดือนก่อน +1

      "You said exactly what I wanted to hear." - Eh, you don't seem to realize, but that is a very bad thing. It means the video isn't only useless to you, it does you a disservice, when you're evaluating options and have a preference, the input you need is the opposite side of yours. If your criteria survives unscathed, you had and still have the right idea, if not - reconsider. I you instead watch things that confirm your preference, you're invalidating your previous preference to a degree because you're just reinforcing it which is worse than doing nothing.

  • @rupasree8055
    @rupasree8055 2 ปีที่แล้ว +3

    Thx for doing this video , we really appreciate it, as there are only few videos regarding workstation GPUS

  • @CrimsonKing666
    @CrimsonKing666 2 ปีที่แล้ว +8

    Something important about the price is that GeForce cards are more unstable. I used to work with Deep Learning with a GeForce 3090 and was pretty common to see my computer crashing or stopping the training. I'm using an RTX a5000 and I never had that issue anymore.

  • @oscarcampbellhobson
    @oscarcampbellhobson ปีที่แล้ว +4

    Thank you for being blunt and to the point, not babbling about everything nobody cares about

  • @yushkovyaroslav
    @yushkovyaroslav 2 ปีที่แล้ว +3

    Very good video really shows what matters.
    Honestly underrated channel a lot more relevant content than some of the "bigger' channels out there.

    • @MediamanStudioServices
      @MediamanStudioServices  2 ปีที่แล้ว +1

      thnaks Y Y. I am looking to do some new videos soome. Just need to get the equipment. That is the hard part

  • @yubawang7652
    @yubawang7652 3 ปีที่แล้ว +3

    Thank you sir! Finally see someone that knows what he's talking about and showing actual production scene

  • @yayandeleon
    @yayandeleon 2 ปีที่แล้ว +4

    finally a more sensible benchmark on actual workflow usage for these cards. sick and tire of those gaming benchmarks who thinks the only actual usage for GPU's are for gaming only and complain for the high price tag

  • @steve55619
    @steve55619 ปีที่แล้ว +12

    Don't forget about AI and ML work. Larger LLM's benefit from more VRAM. Also note how much heat you produce and power consumption with 2x 3090 in NVLink vs RTX A6000

  • @fanimations2363
    @fanimations2363 3 ปีที่แล้ว +1

    Something i've been searching a lot on TH-cam, great comparison , loved it, thanks!

  • @mistrrhappy
    @mistrrhappy 2 ปีที่แล้ว +2

    I'll second the request for the A6000 vs 4090 comparison! Interesting to see the results!

  • @kentharris7427
    @kentharris7427 2 ปีที่แล้ว +7

    You can rent the cards for $1.50 an hour for one card or $6.00 an hour for 4 RTX6000 cards or $1,000 per month per card, cloud based. I personally have the 3090 card in my PC which is good for most applications. If I need raw speed for any given time I will rent the cards.

  • @andrewfischer247
    @andrewfischer247 3 ปีที่แล้ว

    This was really well done and I appreciate how you compared several scenarios. Subscribed!

  • @TrueMathSquare
    @TrueMathSquare ปีที่แล้ว +4

    I just found your channel and I love it.

  • @concinnity1240
    @concinnity1240 2 ปีที่แล้ว

    This video helped me out so much and answered all the questions I've had while I'm trying to build a workstation for CAD. Thank you so much! Excellent video.

  • @craigvfx
    @craigvfx ปีที่แล้ว +17

    Would you do a comparison with the new 4090 vs A6000 ADA Lovelace cards

  • @faceless_ghost
    @faceless_ghost 3 ปีที่แล้ว +2

    you're replying to every comment,really great! you're putting a lot of time in it!
    thank you so much💗

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +1

      yes Imran Mohd Abdul, I reply to all comments.
      Thanks for watching

  • @essa07
    @essa07 2 ปีที่แล้ว +4

    professional comparison …just what I need

  • @armalik11
    @armalik11 6 หลายเดือนก่อน +2

    Excellent video. Gave some vital information about the cards and how memory has been useful for different situations

    • @user-xj1ll7qu8l
      @user-xj1ll7qu8l 6 หลายเดือนก่อน +1

      the rtx 4090 beats the a6000 haha

    • @FirephoenixX02
      @FirephoenixX02 2 หลายเดือนก่อน

      ​@@user-xj1ll7qu8lStill half of the VRAM tho, if you need complex scenes in a production environment better compute power won't be of much use if the scene doesn't fit fully in the VRAM.

  • @jeremiahMndy
    @jeremiahMndy ปีที่แล้ว +3

    Keep making these please I'm a pro 3D artist and your videos have really helped.

    • @MediamanStudioServices
      @MediamanStudioServices  ปีที่แล้ว +1

      thanks for watching, I hope to make new videos soon. Sorry for the long delay in making content

  • @E_Clip
    @E_Clip 3 ปีที่แล้ว +1

    The memory pooling since the 2080Ti's have been great for production and I really don't see myself buying a quadro (or A as they are called now) ever again. The pooled Vram from 2 x 3090's is more than enough for my workloads (mostly ArchVis).
    Great content mate, glad I found you! Keep it up :)

  • @wagnerdesouza6512
    @wagnerdesouza6512 2 ปีที่แล้ว +3

    Very interesting channel, testing hardware with professional softwares, and not with games.
    Would be nice to see gpu tests with Substance Painter.

  • @pedrorivera1892
    @pedrorivera1892 3 ปีที่แล้ว +1

    Thank you for the video. When doing 3d renderings the biggest difference I found between A SERIES Vs GEFORCE is temperature.

  • @qkayaman
    @qkayaman 2 ปีที่แล้ว +7

    It depends on what you need them for, but for me I need A6000/A5000, not RTX 3090. Why? Multi-GPU setup, where I need peer-to-peer (P2P) access between all GPUs; memory transfer between GPUs through host memory is a no go for me. P2P is part of Nvidia GPUDirect, and only possible with 3090 over NVLink (i.e. only possible between pairs), so if you want a 4 GPU setup and need P2P, forget about it. With A6000/A5000 P2P is possible over PCIe, which means running 4 of them is possible. Also dual slot profile makes it easier to stack (I know blower type 3090 is available, but hard to find). May also be interested to know, in Windows in order to enable P2P over NVlink, SLI mode needs to be enabled in NVidia Control Panel. Fun fact: NVidia disabled this in latest Windows drivers, so if you want it to work, need to roll back to a pre-Jan 2021 driver! Linux drivers don't have this issue though.

  • @surroundrive
    @surroundrive ปีที่แล้ว +3

    Excellent production design: your set, lighting, audio, video, dialogue...liked and sub'd.

  • @oscaroscar9941
    @oscaroscar9941 3 ปีที่แล้ว +1

    Just the things that I want to see. Well done!

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว

      thanks Oscar
      Please check out the rest of the videos on my channel

  • @thewizardsofthezoo5376
    @thewizardsofthezoo5376 ปีที่แล้ว +5

    One thing is the power consumption and the lack of VRam, then if it takes a couple of minutes or hours more to run is less critical, for LLM fine tuning, those consumer cards are useless because of lack of RAM, because it's the size of the VRam that determines what you can load in for training.

  • @joaoalexdias
    @joaoalexdias 3 ปีที่แล้ว +2

    Hi thank you for your review! I’m a 3D character animator using mostly Maya, I worked in the majors studios using workstations with both cards you mentioned, and my thought is that a Quadro card is more efficient in the viewport and computing processes than a GeForce. Even in my personal rigs with lower end cards I’ve noticed that, for example I had a workstation with a Quadro K2000D with 2G of VRAM and a laptop 5 years newer with a GTX1060 and I had a better viewport speeds with the workstation, the workstation CPU was a bit better but wouldn’t justify the differences in performance. Your totally right on that comparison with the A4000, I would definitely choose that one over an RTX 3070 or even 3080. I guess it all comes down on the production usage, if you’re going for the renders or even 3D generalist a Gforce might be a better choice budget wise, but if you’re doing stuff like character animation, vfx and simulations I would definitely choose a middle end quadro.

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว

      hi Joao Dias, I would not compare a K2000 and a laptop 1060, this is not a good comparison at all. you do know that all laptop GPU are rated very much differently than desktop GPU. A RTX3080 Laptop is not the same as a desktop 3080. the Laptop version of a 3080 is more like a 3060 desktop GPU. Laptops just can not deliver the power required to drive these GPUs.
      But thanks for sharing your comments and watching the channel.

  • @CreativeAudience
    @CreativeAudience 2 ปีที่แล้ว +3

    Thank you for your test, I agree with you. I'm a 3D motion graphic designer and animation. I have been working with C4D and Octane render for many years. From my experience, I'm working on Quadro and Geforce. The 3D preview frame rate performance and rendering of Quadro and Geforce are not different. Quadro is just only of their marketing or product positioning but the price is too cruel. Quadro only has more RAM but it's a lot more expensive than Geforce 5 times. For me, It's a huge cost. I try to argue with other people over the years about Quadro and Geforce but no one believed. Especially the computer sellers and people who are not graphic designer.

  • @NimaCn
    @NimaCn 3 ปีที่แล้ว +1

    Thanks a lot for the comprehensive video. Subbed for the future videos!

  • @Andbar93
    @Andbar93 ปีที่แล้ว +3

    Thanks for the video, I hardly found comparisons between the Quadro and the rtx in a professional environment, I wish you could make comparisons in AI tasks such as generated images.

    • @tanguero2k7
      @tanguero2k7 ปีที่แล้ว

      Hi there! Let me save you some time (TLDR):
      The results both in rendering and AI, given the same prompt, parameters AND SEED, are the same ( on both a 6GB RTX 3060 (mobile) and a much faster 24GB Quadro RTX 5000).
      The long version:
      I bought an RTX 3060 based laptop because I would never give more than 300€-400€ ($300 aprox) for an 8GB card for both rendering (blender) and AI (local implementations of stability-ai and BERT related workloads) work. When I get to where I want, I move everything over to an RTX 5000 at my workplace.
      Other than the size of the generated images, I can only say the 5000 (naturally) returns faster: my 6GB 3060 often crashes when attempting to render 4K (blender) or simply refuses to generate textures/images (dreamtextures on blender / stability-ai on the shell ) above 512x512.
      This, however, might change in a near future due to a recent paper by nvidia themselves where the model used for raytracing was told to be bellow 1 MB (yes, you're reading 1 Megabyte). Have a look at 2-minute-papers channel and have a look for yourself.
      Oh! Btw, if you'd like to test some workload before buying, let me know. (Edited to add that I also do some occasional photogrammetry work with the free and open source meshroom)

  • @webdesign6776
    @webdesign6776 3 ปีที่แล้ว

    I always enjoy your videos ,in this one I especially liked knowing that the studio ready drivers had the same bug fixes as the "Quadro" drivers

  • @rashdanml
    @rashdanml 2 ปีที่แล้ว +3

    I think the key point here is that Nvidia only recently started releasing Studio ready drivers for Geforce cards, as of the 3000 series. It USED to be true that Geforce wasn't suited for workstation usage because of the lack of driver optimizations for the Geforce line.
    The underlying hardware has pretty much always been the same with differences in numbers. Weaker workstation GPUs (i.e. fewer CUDA cores than Geforce) were still preferred for workstation use because the Studio drivers were better optimized to use that hardware.

  • @rajis92
    @rajis92 3 ปีที่แล้ว +1

    Finally. Someone doing relevant hardware reviews and comparisons for 3D/VFX purposes. That deserves a sub :)
    I've always said this. There's a reason why all the render engines and 3D programs that utilise CUDA use Geforce cards for demos and put them first in their compatibility lists first before showing a Quadro card. When it comes to rendering the software developers make it so the software utilises CUDA/RT/Tensor cores. It doesn't care whether it comes from a Quadro or Geforce. It only cares for "how many cores can you give me to computer this task?". A lot of studios that do GPU rendering will use Geforce cards because cost to performance makes far more sense. GPU shortage/price hike aside you could buy like 2x 3090 for the price of one A6000.
    Quadros are mostly appreciated for the purposes of:
    1. Rendering crazy amount of polygons and loads of 4k+ textures with loads of UDIMs that are not mipmapped (non tex or tx files). Even then mipmapping and proxies are your friend when it comes to memory on the GPU
    2. CAD like programs utilise Quadro features far more than your standard 3D DCC/Painting apps (Maya, Houdini, C4D, Max, Substance, Mari e.t.c)

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว

      Hi Jermaine, thanks for sharing your experience with the channel.

  • @SteveGrin
    @SteveGrin ปีที่แล้ว +5

    Speaking from experience, my A4000 out performs my 3080 in AutoCAD and Revit in two ways. First the 3080 lags during certain operations - for example "override graphics in view" in REVIT or "layers" in cad. The second thing is the artifacts that you get with the G-force card when rotating a model are annoying. Every time I get a new WS, I try the current top of the line g-force card and every time I end up back to the WS card.

    • @animhaxx
      @animhaxx ปีที่แล้ว +3

      So you saying is A4000 is better at viewport handling?

    • @SteveGrin
      @SteveGrin ปีที่แล้ว

      I guess you could say that.

  • @ChrisDallasDualped
    @ChrisDallasDualped 3 ปีที่แล้ว +1

    Another amazing video thx for posting.

  • @Oldyellowbrick
    @Oldyellowbrick 2 ปีที่แล้ว +4

    I think the cost difference is pretty insane BUT I use Octane render and I am always maxing out VRAM and having to reduce scenes...Not only does it slow you down considerably when you reach the 'ceiling' but you tend to get alot of issues when you reach the 90% mark of the VRAM capacity with systems crashes. 48gb would be very welcome in my workflow but I will wait to see what the 40 series will offer.

  • @tyrannicpuppy
    @tyrannicpuppy 2 ปีที่แล้ว +1

    Very nice. As someone starting to dip their toes into 3D content creation for fun, but only currently has the 4GB 1650 Super I could afford when putting the tower together midway through last year, it's nice to see a video addressing the content usability of the 30 series cards. LTT and that ilk make great videos, but they barely give the render benchmarks a mention and they certainly don't go into this level of detail. I know the new fancier cards are on the horizon, but this has helped convince me to grab a 30 series now and enjoy stable and yet powerful rendering compared to what I'm getting now. I can always splurge again in a few years if the newer ones are really so much better, but by then I might be doing far more complex stuff with it thanks to a few years of practice and need the extra horses.

  • @bravestbullfighter
    @bravestbullfighter ปีที่แล้ว +5

    Interesting! How about follow-up with A6000 vs 4090?

  • @miavsm
    @miavsm 2 ปีที่แล้ว +2

    I like master like you, now people counting views without knowing that technology advances or disadvantages.
    Thank you for sharing your knowledge 🙏🏻

  • @sideffect8
    @sideffect8 3 ปีที่แล้ว +4

    You should benchmark the 2x nvlinked 3090s vs the A6000. Graph the memory usage and bandwidth during the pre render. Curious to know how the complex scenes saturate the extra memory

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว

      hi Sideffect8, I totally would but I do not have two 3090 anymore.
      Thanks for watching

    • @siminc7905
      @siminc7905 3 ปีที่แล้ว

      Nvlink chopsoff about 10% performace. so do the math

  • @loganpenciu7317
    @loganpenciu7317 3 ปีที่แล้ว

    You sir are speaking my language! Been looking for a channel that talking about computers in a 3d production studio setting. Subscribed! :)

  • @ronniecoleby
    @ronniecoleby 2 ปีที่แล้ว +3

    Just what I was looking for thanks so much. I think this solved my dilemma I'll go for the 3090 and maybe add a second card in the future - with prices tumbling down now that seems more sensible! I'm looking to purchase a workstation for a personal project which uses Metahumans in Unreal Engine. Would be great to see how the two hold up in the viewport in Unreal - in a filmmaking (24fps) context. I know this is a more niche use case though! :-)

    • @mikebrown9826
      @mikebrown9826 2 ปีที่แล้ว

      You may want to research more. I believe Unreal can only use one GPU. And I am not sure if the nvlink will work for Unreal. But you could render on one GPU while using the second to continue to work in the program.

  • @yushkovyaroslav
    @yushkovyaroslav 2 ปีที่แล้ว +1

    Thanks for being a guy who actually came out and put the only useful thing out there for comparison: actual rendering numbers.

    • @MediamanStudioServices
      @MediamanStudioServices  2 ปีที่แล้ว

      thanks for watching Y Y. I try to do a good job. Share with your friends to help grow the channel.

  • @scottsturmTWM
    @scottsturmTWM 2 ปีที่แล้ว +7

    Hoping for a 4090 vs A6000 video ; - )

    • @eliahr11
      @eliahr11 2 ปีที่แล้ว

      duuude that's not even a comparison those gpus are generationally different the Axxxx is ampere base where the 4xxx series is Ada based. The 4090 will beat the crap out of the quadro card immediately

    • @scottsturmTWM
      @scottsturmTWM 2 ปีที่แล้ว

      @@eliahr11 you do realize a new ADA based A6000 card is coming out soon, yes?

    • @eliahr11
      @eliahr11 2 ปีที่แล้ว

      @@scottsturmTWM didn't know that, ty

    • @isaacvl95
      @isaacvl95 2 ปีที่แล้ว

      @@scottsturmTWM what would you recommend for fcd simulations cad and some 3d rendering some used rtx Quadro 4000, any rtx from the 3000 series starting from at least a 3080 or the current A4000

  • @Royameadow
    @Royameadow 2 ปีที่แล้ว +1

    We don't get a lot of testing with the Quadro cards as is, so it truly is welcoming whenever we get to see somebody take the time to compare these products to their GeForce counterparts, the added software compatibility and considerably higher Video Memory truly does make a difference and today the subject of Clearance Factor has become ever more important in a time where the RTX 4090 barely has any Dual Slot options when compared to the Quadro L6000 (RTX 6000: Ada Generation) that many people are beginning to flat~out ditch GeForce because of the lack of smaller options to fit in a smaller case (I've been using Micro ATX since I999; never have I used a larger form factor).
    In this era, the core audience that will benefit heavily from products such as the Quadro L6000 is definitely in the AI and Deep Learning space: Now, I don't know what your history is with certain Memory Intensive AI software such as OpenAI Jukebox, but it would be incredibly welcoming to see how much faster Music Continuations with it render on the L6000 when compared to the 4090, even the Tesla cards witness a major handicap at the 0I6 and 024 GB threshold in this particular workload that jumping to a card with a capacity of 048 or better truly does help the process run smoother without either becoming sluggish or crashing due to an Out of Memory error; Jukebox is something that I truly hope we will see more people in the Techtuber scene showcase as part of their Benchmark Suite, it doesn't get a lot of attention outside of a select few and thus having concrete numbers on how fast it works under plentiful conditions would show the Quadros' true worth over GeForce, we're still a few generations away from being able to render a Sixty Second Sample in the same time as its length or shorter and it'll be nice to see which cards get the most out of it until that time ultimately comes. (:

  • @sameerkadam4956
    @sameerkadam4956 ปีที่แล้ว +6

    Nobody is talking about TDP comparison between workstation and GeForce cards. cost of running system and power bills in commercial setup or running workstation or server 24 x 7 RTX3090 TDP is 350Watt whereas a4000 TDP is only 140Watt.

    • @MediamanStudioServices
      @MediamanStudioServices  ปีที่แล้ว +2

      Hi sameerkadam4956, I agree that power usage is a big topic. I will make a video on this subject. However, using slower GPUs for a render does not necessarily reduce the overall power consumption. It just takes longer to render the frame. However, looking closely at power utilization is a big factor that is overlooked in purchasing GPUs for projects. Thanks for watching and the video topic idea.

  • @kitewinds663
    @kitewinds663 2 ปีที่แล้ว +2

    Thanks for video, very helpful! A comparison of Solidworks assembly and drawings performance between the A6000 and the W6800 AMD-card would be interesting. Also the A5500 is of interest. Thanks again.

  • @NarekAvetisyan
    @NarekAvetisyan 3 ปีที่แล้ว +3

    Great review just what I was looking for!
    Q: Can you test the memory pooling of 2 RTX 3090's with NVLink in Blender? I'd really like to see that.

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +1

      Hi Narek Avetisyan, I would love to but I do not have the equipment anymore.
      I only have for a short time to make the videos.
      Thanks for watching

  • @Fyrock
    @Fyrock 3 ปีที่แล้ว

    I’m looking for a new laptop and I have the choice between RTX 3070 or RTX A5000. I’m using Unreal Engine and Blender so I think the A5000 would be much better especially because it has 24gb vram while the 3070 has only 8gb. I’ve also heard Unreal Engine doesn’t support dual gpu, so the a5000 really seems like the better choice. What are your thoughts?

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +1

      Hi Fyrock, I agree with you that the RTX A5000 is the better choice for your needs. with the 16GB of Vram, you will see a big difference in performance.
      Thanks for watching

  • @andreasfjellborg1810
    @andreasfjellborg1810 3 ปีที่แล้ว +4

    Here(Sweden) you can get 3x 3090 for one A6000, going 3x 3090 would be a quite a lot faster than a single 3090...

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว

      hi Andreas, i would love to have 3 RTX3090, but its so hard to find them in the market right now at a good price
      Thanks for watching

  • @nigeldawson5960
    @nigeldawson5960 3 ปีที่แล้ว

    Thanks for the info. I’ve watched many of your vids and they helped me build the best machine for me. Much appreciated.

  • @oliverleemans6363
    @oliverleemans6363 ปีที่แล้ว +5

    Can You test the A5000 against the RTX 3090 or the RTX 4070 ?

  • @awnina7627
    @awnina7627 2 ปีที่แล้ว +1

    I am an architect and ur videos are amazing , please we need more videos regarding the motherboards types we need

  • @wonderwonder9027
    @wonderwonder9027 2 ปีที่แล้ว +3

    Can you please do the following tests on the A6000:
    - AutoDesk Revit architecture render
    -MatLap heat exchange simulation
    -AutoDesk Advance Steel stress and displacement calculation
    -AutoDesk Robot wind load simulation and seismic load calculation
    I know they are out of the scope of this channel but me being a civil engineer have no idea what to expect from this kind of investment if I'm going to make it....... And no other channel are nice enough to read through the comments section let alone give an answer.......
    Thanks for your time

  • @emanggitulah4319
    @emanggitulah4319 3 ปีที่แล้ว

    Great to see this content. As you said a lot of other channels say that you have to have a quaddro.

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว

      Hi Emang, I have worked in so many studios that do not us "Quadros" GPUs.
      Thanks for watching

  • @checkmate8015
    @checkmate8015 2 ปีที่แล้ว +5

    Which one is better for me as a game dev and 3D artist

    • @faradaysinfinity
      @faradaysinfinity 2 ปีที่แล้ว

      I too am asking this. Intuition says a6000. But still. I mainly use Unreal 5

    • @mikebrown9826
      @mikebrown9826 2 ปีที่แล้ว

      Get the rtx3070 as this is the middle tier gpu. So your dev needs to play on this card. And it is still powerful

    • @mikebrown9826
      @mikebrown9826 2 ปีที่แล้ว +1

      @@faradaysinfinity if your doing unreal then get the 3090. Or a6000

  • @rapatouille
    @rapatouille ปีที่แล้ว

    great comparison presentation!
    love your kitchen studio! very nice concept

  • @EdinGacic
    @EdinGacic 2 ปีที่แล้ว +4

    do you have any tests with dual 3090 with NVLink on big scenes like the one you showed where double VRAM made a huge difference? I am debating if adding another RTX 3090 FE to my workstation is better at 750-800 EUR used or selling the RTX 3090 and buying the new RTX 4090. I am leaning more towards two 3090s if NVLink actually works and scales good. It would be cool if you can do test like this :)

    • @thomasrichter1219
      @thomasrichter1219 2 ปีที่แล้ว

      I have exactly the same thoughts. Have you already made a decision?

  • @Gettutorials123
    @Gettutorials123 3 ปีที่แล้ว

    Thank you again for such informative video!

  • @Livingston3d
    @Livingston3d 2 ปีที่แล้ว +5

    Great sir.! Please help me. Will this GPUs (Radeon RX 6800 XT, Radeon RX 6600 XT, Radeon RX 7000 XT ) are good for Maya, 3ds max, Zbrush, painter, blender, etc. Is it good for our modeling and hard file viewport navigation and rendering purposes?

    • @Amarthir
      @Amarthir 2 ปีที่แล้ว

      They will ;')

    • @Pixel_FX
      @Pixel_FX 2 ปีที่แล้ว +1

      Radeon GPUs Render slower in Blender cycles compared to nvidia because of Optix. For every other program they are fine, its rendering that Radeon cards are slower. Dont know about upcoming 7000 series performance. I have a 5700XT and 3080s. my 5700XT was faster than 2070 until optix introduced. after optix in blender, RTX cards became way faster.

    • @The0zx
      @The0zx 2 ปีที่แล้ว

      @@Pixel_FX How about Intel ARC A770 for the same work?

  • @originor4751
    @originor4751 ปีที่แล้ว

    Very useful. Thank you for putting this out!

  • @kuhan333
    @kuhan333 2 ปีที่แล้ว +3

    Hi, Great Video! I have a couple of questions,
    1: What are you thoughts on GPU for Unreal Engine (Content creation/ Virtual Camera/ Green screen maybe but No LED wall ) 3090 rtx vs A5000.
    2: There are many makers for 3090 card, Which one would you recommend, founder's edition vs other makers.(I was looking in 3090 FE vs 3090 ASUS ROG-STRIX but if you have other recommendation please do share)
    Thanks in advance,

    • @MediamanStudioServices
      @MediamanStudioServices  2 ปีที่แล้ว +4

      for UE4 i would use the RTX3090 as UE required a lot of processing power, and the 3090 has more than the A5000
      As for brands, sorry i have not done a comparesion of the different brands. I have a Gigabyte Tubro and it had been great for me.
      I also have a Strix 3060ti and that us also a good GPU model for me. so you will have to find what is best in your market and also available.
      Thanks for watching

  • @Sitrec
    @Sitrec 2 ปีที่แล้ว

    Just wanted to say that I really appreciate you and your content. There is so much misinformation when it comes to hardware in the creative space and content like this has been really missing.

  • @technicallyme
    @technicallyme ปีที่แล้ว +4

    I got a a4000 for 400 (what a change a year makes ) but it solved my problem with the 3070. Not enough memory

  • @Betoromero22
    @Betoromero22 ปีที่แล้ว +2

    Por fin alguien serio que se dedica a hacer videos para creadores!!! Gracias por compartir

  • @dittofarmers9007
    @dittofarmers9007 2 ปีที่แล้ว +4

    Hi would you please the same test but not for rendering, but for computation. Try software such as Ansys mechanical or Abaqus. These software has the option to utilize GPU in the computation. We really would like to know if Geforce can work as good as Quadro for computing in double precision mode. Thanks. That will help a lot of engineers out there!

  • @dustinjenkins8215
    @dustinjenkins8215 3 ปีที่แล้ว

    Excellent content! Definitely subscribing.

  • @Mr_i_o
    @Mr_i_o 3 ปีที่แล้ว +4

    Perhaps you would consider doing performance per dollar with dual NVLink 3090 vs dual a6000 vs single a6000?

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +4

      I would love to but getting the GPU to make the video is very hard. I am still looking for the GPU but once I get the cards, you bet I will make a video.
      Thanks for watching

  • @albertsitoe7340
    @albertsitoe7340 2 ปีที่แล้ว

    This is the best video about the topic! I can finally safely direct people who ask to this awesome Masterclass!

  • @capezonmyback
    @capezonmyback 2 ปีที่แล้ว +4

    I can't find a new NVlink 2x 3090 Benchmark. Would be really helpfull!

    • @mohamedsakka2338
      @mohamedsakka2338 2 ปีที่แล้ว +1

      th-cam.com/video/jw_mnwo9Nag/w-d-xo.html you can skip the building part it might be boring

  • @mikemora6410
    @mikemora6410 2 ปีที่แล้ว

    Thank you for all your content, I truly appreciate it. This video was very helpful.

  • @arjayjalmaani
    @arjayjalmaani 3 ปีที่แล้ว +3

    Would you say for real-time video rendering using Unreal Engine, would I still need an A6000 or would two 3090’s be as well, if not slightly better?

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +3

      I would go with the one A6000 as Unreal does not support multi GPU rendering

    • @arjayjalmaani
      @arjayjalmaani 3 ปีที่แล้ว +2

      @@MediamanStudioServices Good to know. Thanks for the reply!

  • @zaydraco
    @zaydraco ปีที่แล้ว +2

    For AI workloads the memory and tensor cores are important due to the amount in the data sets. If you have bigger data sets it becomes more important to have bigger memory or have a better algorithm to partition the exchange of RAM and GPU RAM... If I remember from the Cuda API you move memory from one to the other, the thing with normal memory in a program is that you can page into swap memory or virtual pages, but it is not the same with GPUs. At least not automatically handled by the OS as far as I know.

  • @epsilonplus3514
    @epsilonplus3514 3 ปีที่แล้ว +3

    can I use mixed multiple gpu on blender?
    like gtx 1070 and 1050ti.

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +2

      yes you can but the renders will be limited to the lowest amount of Vram in any one single GPU. This is the limitations for multi GPU rendering. The package is bundled up with Vram limitations of a single card. So say you have a card with 6GB and one with 12GB. the render packages will only be set for 6GB and the second 12GB card will not use 6 of its GB of Vram. I hope you understand my response.

    • @epsilonplus3514
      @epsilonplus3514 3 ปีที่แล้ว +3

      @@MediamanStudioServices thank you for answering.

  • @patrick-resendiz
    @patrick-resendiz 2 ปีที่แล้ว +1

    Great video! Trying to decide between 3090, 3080 and 3070ti for use with D5 Render. 3090 would probably save me a good amount of time but looking for a sanity check on price to value. I do mostly static image and panorama generation with the occasional video. Scenes are residential architecture with about 30 2k textures and standard D5 objects. So glad I ran across your channel. I’m using a laptop with a 3070 right now. Looking to save time.

    • @MediamanStudioServices
      @MediamanStudioServices  2 ปีที่แล้ว

      well the 3090 is coming down in price but still expensive at around $2000, but it is fast. So hard for me to say what is best for you. Good luck making your choice.
      thanks for watching

    • @patrick-resendiz
      @patrick-resendiz 2 ปีที่แล้ว +1

      @@MediamanStudioServices I think I’ve decided on a prebuilt Omen 45 with either a 3080ti or 3090. My budget is about $3,400 and can either do the 3080 ti with a faster Ryzen or a 3090 with a less powerful Ryzen for same cost. Great option either way. Your channel is amazing!

    • @mikebrown9826
      @mikebrown9826 2 ปีที่แล้ว +2

      @@patrick-resendiz well good luck whatever way you go. You can always get a new cpu later and swap it out. Thanks for watching the channel

  • @maxximus8904
    @maxximus8904 3 ปีที่แล้ว +10

    Why doesn't Nvidia make a card that has a 100% focus on RT cores
    Like 300 RT cores instead of a couple dozen
    Professionals really wouldn't mind giving up their Fortnite FPS for faster render times

    • @DamageNando
      @DamageNando 3 ปีที่แล้ว +1

      Preach to em 👏🏼

    • @iamstartower
      @iamstartower 3 ปีที่แล้ว

      Is the way it is so you can play fornite on the unused cuda while rt cores do their job...

    • @mcan-piano4718
      @mcan-piano4718 2 ปีที่แล้ว

      I agree.They can do everything even realtime viewport, but I dont know why they dont use all of their potential...

  • @sainsay
    @sainsay 3 ปีที่แล้ว

    I have had the opportunity to work with servers running Quadros for deep learning and machine learning. the one thing I learned is that having ECC memory is a blessing if you have long workloads like a week or several weeks. yes, you make backups while processing the tasks but not having or seriously reducing crashes is better than having to restart from a backup. also having 48GB is nice, pooling the memory was/is not always efficient for the workloads I work with. having up to 4 GPUs work independently greatly improves the quality of my workload.
    but for workstations Quadros are becoming more and more obsolete for a lot of people and it is great to see this comparison. also showing that if you scale up the workload it eventually becomes more than worth it.

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +1

      thanks for sharing your experiences with the channel. As you have pointed out, some workloads require the workstation GPUs.
      thanks for watching.

  • @itsaman96
    @itsaman96 2 ปีที่แล้ว +4

    I always install studio drivers on my 1660 super 6gb

    • @duh4293
      @duh4293 2 ปีที่แล้ว +3

      Another 1660 super user in the wild. Heck yeah.

  • @subhaprakashbeura5120
    @subhaprakashbeura5120 2 ปีที่แล้ว

    You did a great job 💯
    Thanks for suggestions

  • @ChosenMan37
    @ChosenMan37 2 ปีที่แล้ว +4

    What about a6000 Vs 4090 comparison test?

  • @nortonf2008
    @nortonf2008 4 หลายเดือนก่อน +2

    very well, friend! thak you for this content.

  • @Nekko_X
    @Nekko_X 2 ปีที่แล้ว +3

    Very interesting, but now I have a question, I plan to upgrade the graphics cards of all my computers but I could only afford to upgrade to an A4000 and not an A6000, but I could upgrade all of them to a 3090, all the computers are used only for 3D modeling and animation, not for rendering or anything else. But usually heavy scenes are created, generally many polygons, so my question is, would the A4000 or the 3090 be better for that type of work? The only thing that interests me is that the scene is seen in real time when modeling, or when animating, and that the viewer of the work area does not get stuck and is seen in real time either when loading a preview in it viewer or simply while animating. It would be very useful if someone helps me with this question, because I can't find videos or information about this, only tests in video games, which is of no use to me at all. Which would be better?
    I'm not looking for the best on the planet either, but I do want to be sure that I make the best decision...
    I've been thinking for 3 weeks and I still don't know. The worst thing is that I have to make the decision before the following week.
    Damn.... :/

    • @khayelihlemngoma2393
      @khayelihlemngoma2393 2 ปีที่แล้ว +1

      Thank you for this comment. I can relate to it. I am building a workstation to run ArchiCAD and I am undecided on the graphics card. By initial assessment it seems the 3090 has more memory than the A4000 and both have ray tracing. I hope he answers your question.

    • @mikebrown9826
      @mikebrown9826 2 ปีที่แล้ว +4

      The 3090 is a way better choice. It has way more Cuda and RT cores than the a4000 and the increase in vRAM is better for the complex/large scene files. Get the 3090 for sure

    • @Nekko_X
      @Nekko_X 2 ปีที่แล้ว +2

      @@mikebrown9826 Yes, I know that the 3090 has more vram and cuda cores, but my fear is that quadro cards are specially made to work with many polygons without problem, which makes rendering them look smooth in the editor, and the preview of the animation in the work area also looks smooth. Well, they were made for that. And I have seen that the "gaming" cards, which in this case is the 3090, will have a lot of power but they are not so good when a high amount of polygons is already used, because the work area no longer begins to look smooth nor the previews in real time. That is my biggest concern, I don't want to buy the 3090 and have the work area not work well, or buy the A4000 and not have enough power or vram. All my life I have used quadro cards in my computers, it is supposed that with each generation their power improves, but now the doubt comes to me between the power of the nvidia 3000 series and the A series quadro. I have also heard and read that the "gaming" series of nvidia tend to crash in programs where polygons are used, while with quadro cards it almost never happens (and I know). That's another one of my fears... I think I will have to take a chance and choosing to purchase the 3090 cards, I sincerely hope that I do not regret my choice. Thanks for the help!

    • @mikebrown9826
      @mikebrown9826 2 ปีที่แล้ว +4

      @@Nekko_X I think you will find the 3090 does just a good of a job and the workstation. GPUs. Good luck on your purchase

    • @goldenheartOh
      @goldenheartOh 2 ปีที่แล้ว

      Was your deadline to put the expense on 2022 for tax filing?
      Without the deadline, worst case scenario would be to buy one 3090 and try it out 1st.

  • @valdisgerasymiak1403
    @valdisgerasymiak1403 3 ปีที่แล้ว +2

    A lot of people use GPUs for machine learning tasks. When I went from 3070 to 3090 I got x3.2 speed increasing mainly because of the RAM (I increased batch size while training) So it's worth going from 3090 -> A6000 if the x2 memory will give x2 speed with lower power usage.

  • @supercolor222
    @supercolor222 3 ปีที่แล้ว +2

    both are good but for 3D industry the more Vram it got the stable the sence and faster it can handle like Unreal Engine, Houdini. for me I will choose to get two 3090 instead of single A6000

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว

      Hi Super, two RTX3090 is a good choice, But Unreal and really only use on of these cards. I am not sure how the NvLink will work with Unreal. Let me know how it Performs. But for Houdini, I would choose a RTX A6000
      Thanks for watching

    • @marekkovac7058
      @marekkovac7058 3 ปีที่แล้ว

      If you only need GPU for houdini openGL stuff, It might be a better choice to go for A6000 because Houdini cannot use more than 1 gpu for a single task. Yes you can use one for openCL and another for openGL but for example 2 gpus both doing openCL wont work. A6000 48GB can also give you a hi-res quality minimal pyro solver results. NVlink wont help in houdini, TCC mode might work(havent tested it yet) but that function is only available in Quadro cards and you also need a third GPU to run the display.

    • @MediamanStudioServices
      @MediamanStudioServices  3 ปีที่แล้ว +1

      @@marekkovac7058 thanks for your input, I agree with you if someone is doing a single workflow like Houdini or Unreal, I would choose the A6000. The extra Vram is required for these types of workloads.

  • @billywilliam7747
    @billywilliam7747 ปีที่แล้ว

    Very good insights I learn from your channel - THANK YOU

  • @renanmonteirobarbosa8129
    @renanmonteirobarbosa8129 2 ปีที่แล้ว +6

    2 things, VRAM and NVLINK. Thats the main difference.

  • @ukaszgaluba8981
    @ukaszgaluba8981 2 ปีที่แล้ว

    Very good content. Keep it going, mister!

  • @maxsimes
    @maxsimes ปีที่แล้ว +6

    Hate to be the pedantic\petty one 😅, but at 3:50 you said the 3080 has 50% more cuda cores than the rtx 8000. But actually the 3080 has 100% more cuda cores than the 8000, whereas the rtx 8000 has 50% _of_ the cuda cores of the 3080

  • @LordShockwave9
    @LordShockwave9 3 ปีที่แล้ว

    Dude, thanks for this! I just managed to get a RTX 3060 for my workstation at home and was thinking about swapping it for an A2000 card but that card performs worse. You saved me some cash! Thank you!

  • @sHuRuLuNi
    @sHuRuLuNi 2 ปีที่แล้ว +6

    Yes, but can it run Crysis?

    • @infamoustimes2927
      @infamoustimes2927 2 ปีที่แล้ว

      🤣

    • @MediamanStudioServices
      @MediamanStudioServices  2 ปีที่แล้ว +4

      yes it can, one of the nice things of the newer Workstation GPUs is that they now support DriectX while in the past they did not

    • @darkairlord
      @darkairlord 2 ปีที่แล้ว +2

      I think the joke went over his head

    • @sHuRuLuNi
      @sHuRuLuNi 2 ปีที่แล้ว +1

      @@darkairlord 😁

  • @arkenavetisyan7308
    @arkenavetisyan7308 2 ปีที่แล้ว

    Most excellent review my dude!

  • @Eneeki
    @Eneeki ปีที่แล้ว +2

    I currently run a RTX 4090 on my Blender workstation and have one simple question. Is it better to build a 2nd PC with another RTX 4090 to do network renders, or can i put X2 4090's in the same system, or lastly, should I spend big money and get dual A6000's in the same workstation to speed up rendering time for more complex renders? A duplicate system to what i currently have is going to cost about 4k less than the cost of the dual A6000's, and adding a 2nd RTX 4090 is a fraction of the cost but i do not know if it is possible without SLI. A work associate of mine has dual A5000's and renders really fast but I do not know if it would be better, or even possibly faster to have 2 computers with 4090's vs 1 computer with dual A6000's or even have dual RTX 4090s? If anyone has a good answer or other questions I do not know to ask, please let me know.

    • @mikebrown9826
      @mikebrown9826 ปีที่แล้ว

      I would get a second system if your 4090 can fit all of the projects into the 24GB of Vram, also, you do not need a super powerful system to put the 4090 GPU for rendering. A 6-8 core system with 32GB of RAM is good enough for a render system. Adding a second GPU to your current system is also an option but having a second system, dedicated to rendering is a good thing. You can also use Deadline, which is free to set up network rendering.

    • @Eneeki
      @Eneeki ปีที่แล้ว +1

      @@mikebrown9826 Thank you. The ram is defiantly needed. I currently run 128 gig of DDR4 3200 and can fill the scene enough that i am close to 90% ram usage during rendering often. My current bottleneck is defiantly V-ram.
      To give you an idea of my current system, CPU is a 5950X, MB is an Asus B550 plus, 128 gig DDR4 3200 ram, Asus Tuf gaming 4090, and all drives are solid state. I will be upgrading the CPU soon and either get a 2nd box or a 2nd video card in the system. I just desperately need to cut down render times to the point i am considered spending 10k on dual A5000's or A6000's.
      Thank you for your input. I will look further into network rendering options.

    • @Eneeki
      @Eneeki 11 หลายเดือนก่อน

      @@mikebrown9826 Well the v ram is not enough currently thus why i am looking for a 2nd GPU. I currently have a 5950X and was going to upgrade to the 7950X3D to prevent another upgrade a year or 2 down the line. A lower end Thread ripper is also an option but I know they run much hotter than the socket 5 CPU's.
      Network rendering is something to consider but I am worried i will trip the breakers constantly with 2 systems in the office. There are a lot of things in that room that require a lot of power like the 3D printer. I would probably do well adding another circuit to the room for the 3D printer, that would solve all of those issues. I will look further into network rendering. thank you.

    • @rahulkamath6984
      @rahulkamath6984 8 หลายเดือนก่อน

      sorry but i am new to the term network rendering. does that mean that you render stuff via network? and how is the GPU being used in that case

    • @Eneeki
      @Eneeki 8 หลายเดือนก่อน +1

      @@rahulkamath6984 Yes. I have a network of 2 computers using Daz 3D's bridge that allows both computers to dedicate the video cards resources for rendering. it is faster for rendering, but know that it needs a really fast network. 10G router and network cards at the minimum for it to really make a difference. this is a very layman way of saying it, but it allows both computers video cards to work together in speeding up a 3D render.

  • @dnashj33
    @dnashj33 หลายเดือนก่อน +2

    Very useful information. I would like to see someone show Smoke, Fire, and Fluid simulations, like a 3D artist might run in most major 3D applications. I am curious if 16GB of VRAM on a 4080 is enough or if they need significantly more.

  • @crckdns
    @crckdns 2 ปีที่แล้ว +3

    the A6000 would be great to train Stable Diffusion models ..so the "AI" can finally render realistic hands!

  • @H2ydrogen
    @H2ydrogen 2 ปีที่แล้ว

    thanks man, well explained and showcased

  • @c.pop.echo.28
    @c.pop.echo.28 2 ปีที่แล้ว +7

    the only reason why I would buy an A6000 gpu is because of Vram capacity. For instance in Chaos Vantage Software I can only use 2 GPUS for rendering, and in Arcviz i do need lots of vram. But, instead of 2x24 3090, I could go with 2x48 A6000. It's just double the capacity, but the price is like 4x bigger. Nvidia, you greedy greedy company.

  • @Birdie_1991
    @Birdie_1991 3 ปีที่แล้ว +1

    3 mins in and you brought it down to way i can understand. "Suck" and "B***S****" caught me off guard as this video screams professional. makes me feel more comfortable as someone whos into tech, music gear, and gaming to know that you don't need big words to get the point across. More smart people should feel ok not having to go to the extremes of showing their level of knowledge by having to use Big words. random thoughts but this soothes my soul. (not satire)