TheDataDaddi
TheDataDaddi
  • 62
  • 213 950
GPU Performance Benchmarking for Deep Learning - P40 vs P100 vs RTX 3090
In this video, I benchmark the performance of three of my favorite GPUs for deep learning (DL): the P40, P100, and RTX 3090. Using my custom benchmarking suite, BenchDaddi, I assess the performance of these GPUs across three major DL architectures: CNN, RNN, and Transformers. Whether you're a data scientist, a machine learning engineer, or just an AI enthusiast, this comparison will provide valuable insights into the capabilities of these GPUs.
In this video, you'll discover:
Benchmark Tests: Detailed performance benchmarks across various AI/ML/DL workloads.
Analysis & Insights: In-depth analysis of the results, highlighting strengths and weaknesses.
Use Case Suitability: Recommendations on which GPU is best suited for different types of AI/ML/DL tasks.
#DeepLearning #MachineLearning #AI #GPUBenchmark #NVIDIAGPU #RTX3090 #P40 #P100 #TechReview #DataScience #MLPerformance #DLPerformance #GPUShootout #TechComparison #ArtificialIntelligence #NeuralNetworks #CNN #RNN #Transformers #TechBenchmark #AIEnthusiast
🎥 Other Related Videos:
AI/ML/DL GPU Buying Guide 2024: Get the Most AI Power for Your Budget
th-cam.com/video/YiX9p8A7LqE/w-d-xo.html
An Open Source GPU Benchmarking Project: BenchDaddi
th-cam.com/video/aCRgkRWY4gw/w-d-xo.html
8 GPU Server Setup for AI/ML/DL: Supermicro SuperServer 4028GR-TRT
th-cam.com/video/JroTsX41e1c/w-d-xo.html
📚 Video Resources:
Link To Looker Report (Data Visualization From Video)
lookerstudio.google.com/reporting/abaf0c09-da02-4a06-9578-65a322b48e3c
Link to Raw GPU Benchmarking Data in Google Sheets
docs.google.com/spreadsheets/d/1IS58k3spPXTYX29HN2bUlzE_j5ujS0S1bBexywkiEaI/edit?usp=sharing
GitHub Repo For BenchDaddi Benchmarking Suite
github.com/thedatadaddi/BenchDaddi
Looker Studio - Free Data Visualization Platform
lookerstudio.google.com/overview
*** PLEASE COLLABORATE ***
I cannot possible buy and test every GPU and hardware configuration, but with your help together we can build a library of benchmark data for the betterment of the AI/ML/DL community as a whole. All you need to do to help is pull down the benchmarking suite and run it on your own machines to test the GPUs you have. Then enter the data into google sheets link below under the "new_results" tab. The og_results tab is the original data I used to make this video.
Collaborative Results Google Sheet
docs.google.com/spreadsheets/d/1IS58k3spPXTYX29HN2bUlzE_j5ujS0S1bBexywkiEaI/edit?usp=sharing
HOW TO GET IN CONTACT WITH ME
🐦 X (Formerly Twitter): @TheDataDaddi
📧 Email: skingutube22@gmail.com
💬 Discord: discord.gg/RyRHEn3yMx
Feel free to connect with me on X (Formerly Twitter) or shoot me an email for any inquiries, questions, collaborations, or just to say hello! 👋
HOW TO SUPPORT MY CHANNEL
If you found this content useful, please consider buying me a coffee at the link below. This goes a long way in helping me through grad school and allows me to continue making the best content possible.
Buy Me a Coffee
www.buymeacoffee.com/TheDataDaddi
As a cryptocurrency enthusiast, I warmly welcome donations in crypto. If you're inclined to support my work this way, please feel free to use the following addresses:
Bitcoin (BTC) Address: bc1q3hh904l4uttmge6p58kjhrw4v9clnc6ec0jns7
Ethereum (ETH) Address: 0x733471ED0A46a317A10bf5ea71b399151A4bd6BE
Should you prefer to donate in a cryptocurrency other than Bitcoin or Ethereum, please don't hesitate to reach out, and I'll provide you with the appropriate wallet address.
Thanks for your support!
มุมมอง: 414

วีดีโอ

An Open Source GPU Benchmarking Project: BenchDaddi
มุมมอง 29514 วันที่ผ่านมา
In this video, I'm excited to introduce BenchDaddi, an innovative open-source GPU benchmarking suite that I've developed. BenchDaddi empowers users to analyze GPU performance across various AI tasks with precision and ease. Throughout this tutorial, I'll walk you through the suite's repository structure, setup, and installation process, showcasing how to leverage its benchmarking scripts tailor...
10G Networking Demystified: Tips, Cost, and Real-World Use Cases
มุมมอง 4452 หลายเดือนก่อน
Are you considering upgrading your network to 10G but unsure where to start? Look no further! In this comprehensive guide, I'll demystify the world of 10G networking, providing you with practical tips, cost considerations, and real-world use cases to help you make an informed decision. Specific Topics Covered: - 10G Networking Basics - SFP vs SFP vs QSFP - SFP Module Types and Considerations - ...
Setting Up External Server GPUs for AI/ML/DL - RTX 3090
มุมมอง 2.6K2 หลายเดือนก่อน
🚀 Join me on a deep dive into setting up external GPUs for high-performance AI computations! In this comprehensive guide, I'll walk you through connecting powerhouse GPUs like the NVIDIA RTX 3090 with NVLink to servers such as the Super Micro 4028GR-TRT and Dell Power Edge R720. We'll cover everything from hardware requirements to configuration steps. ⚙️ What you'll learn: - The advantages of u...
8 GPU Server Setup for AI/ML/DL: Supermicro SuperServer 4028GR-TRT
มุมมอง 4.4K2 หลายเดือนก่อน
In Part 2 of our series, we're diving into the nuts and bolts of setting up the Supermicro SuperServer SYS-4028GR-TRT for optimal AI/ML/DL performance. Discover the step-by-step process to configure this powerhouse server, from installing the GPUs to optimizing software and hardware settings. Learn how to harness the full potential of up to 8 dual-slot GPUs for unparalleled computational power ...
Your New 8 GPU AI Daily Driver Rig: Supermicro SuperServer 4028GR-TRT
มุมมอง 1.7K2 หลายเดือนก่อน
Welcome to Part 1 of our exciting series, where we embark on a deep dive into the transformative world of artificial intelligence, machine learning, and deep learning with the Supermicro SuperServer SYS-4028GR-TRT. In this episode, we're set to explore the myriad benefits of making this powerhouse server your go-to daily driver for all your AI/ML/DL endeavors. Uncover the incredible capabilitie...
Navigating the AWS Applied Science Internship Interview: My Experience
มุมมอง 1203 หลายเดือนก่อน
Description: In this video, I share my journey through the AWS Applied Science interview process, providing insights and tips for aspiring candidates. From the initial application to the final decision, I'll take you step-by-step through each stage, highlighting the challenges and learning experiences along the way. Whether you're an aspiring applied scientist or just curious about the tech int...
Boomers, Millennials, Gen Z: Who Faced Toughest Home Buying Hurdles? - A Data-Driven Exploration
มุมมอง 8435 หลายเดือนก่อน
Welcome to my deep dive into the evolving landscape of home buying! In this video, I take a unique, data-driven approach to explore one of today's most pressing questions: Is it harder to buy a home now compared to previous years (1967-2022)? Join me as I dissect trends in rising home prices and stagnant wage growth, topics that have been widely discussed on social media and beyond. My journey ...
Throttle No More: My Strategy for GPU Cooling in Dell PowerEdge
มุมมอง 1.5K5 หลายเดือนก่อน
Join me on a deep dive into managing and optimizing GPU cooling in the Dell PowerEdge server series. In this video, I share my personal experiences and strategies for preventing thermal throttling to maintain peak server performance. I'll show you how I interact with the server fans by SSHing directly into the server's iDRAC, and how to access the fans using the local command line with Dell's i...
Setting Up a Full Bitcoin Core Node with Docker and Ubuntu 22.04
มุมมอง 2.7K5 หลายเดือนก่อน
🚀 Join us as we dive into the world of cryptocurrency and blockchain technology! In this comprehensive tutorial, we're setting up a full Bitcoin Core node inside a Docker container on Ubuntu 22.04. Whether you're a crypto enthusiast, a developer, or just curious about how Bitcoin nodes work, this step-by-step guide is tailored for you! 🔍 What You'll Learn: Introduction to Bitcoin Core: Understa...
AI/ML/DL with the Dell PowerEdge R720 Server - Energy, Heat, and Noise Considerations
มุมมอง 2K5 หลายเดือนก่อน
Dive deep into the world of high-performance computing with our thorough examination of the Dell PowerEdge R720 Server, a powerhouse for AI, ML, and DL applications. This video is not just a guide; it's an insightful exploration into the server's operational dynamics, focusing on power consumption, heat production, and acoustic management. What You'll Uncover in This Video: Power Consumption In...
Best AI/ML/DL Rig For 2024 - Most Compute For Your Money!
มุมมอง 13K5 หลายเดือนก่อน
🚀 Dive into the future of AI/ML/DL computing with our in-depth analysis of the best rigs for 2024! In this video, we take a close look at the powerhouse Dell PowerEdge R720 server, equipped with top-tier components, and see how it stacks up against custom-built setups and cloud computing solutions. 🔍 Components Breakdown: Dell PowerEdge R720 Server: A robust foundation for demanding AI tasks. ...
iDRAC7 Setup and Access Guide on Dell PowerEdge R720 Server
มุมมอง 8816 หลายเดือนก่อน
Embark on a journey to harness the full potential of your Dell PowerEdge 12th Generation server with our comprehensive iDRAC7 setup and access guide! In this tutorial, we guide you through the intricate steps of configuring and accessing iDRAC7, the Remote Access Controller that empowers seamless server management. Whether you're a seasoned IT professional or a tech enthusiast, this video break...
Mastering iDRAC7: Unlocking Virtual Console & Enterprise Features on Dell PowerEdge R720
มุมมอง 6106 หลายเดือนก่อน
Dive into the world of enterprise server management with our comprehensive guide on setting up iDRAC7 for Dell PowerEdge R720 servers. In this tutorial, we'll walk you through the step-by-step process of configuring iDRAC7 to harness the full potential of virtual console and other advanced features. Whether you're a seasoned IT professional or a server enthusiast, our detailed instructions appl...
DIY Home Network Monitoring System - Raspberry Pi 4, Smart Switch, and Wireshark
มุมมอง 3.9K6 หลายเดือนก่อน
🌐 Join me on an exciting journey as we craft a powerful Home Network Monitoring System using a Raspberry Pi 4, a Smart Switch, and the Wireshark network protocol analyzer tool! 💻🛠️ In this step-by-step guide, I'll walk you through the process of setting up and configuring the Raspberry Pi 4 to serve as a dedicated network monitor. We'll harness the capabilities of a Smart Switch with port mirro...
Dell PowerEdge R720 GPU Deep Learning Upgrade: Installing Dual Tesla P40s with NVIDIA Drivers
มุมมอง 6K6 หลายเดือนก่อน
Dell PowerEdge R720 GPU Deep Learning Upgrade: Installing Dual Tesla P40s with NVIDIA Drivers
Dell PowerEdge R720XD GPU Upgrade: Installing Tesla P40 with NVIDIA Drivers
มุมมอง 7K7 หลายเดือนก่อน
Dell PowerEdge R720XD GPU Upgrade: Installing Tesla P40 with NVIDIA Drivers
Server Data Recovery: How to Save Your System & Extract Data from a Non-Bootable Server
มุมมอง 1177 หลายเดือนก่อน
Server Data Recovery: How to Save Your System & Extract Data from a Non-Bootable Server
Installing DUAL Tesla P100 GPU on Dell PowerEdge R720 Server with Driver Installation
มุมมอง 7K7 หลายเดือนก่อน
Installing DUAL Tesla P100 GPU on Dell PowerEdge R720 Server with Driver Installation
Installing Tesla P100 GPU on Dell PowerEdge R720 Server with Driver Installation
มุมมอง 6K7 หลายเดือนก่อน
Installing Tesla P100 GPU on Dell PowerEdge R720 Server with Driver Installation
AI/ML/DL GPU Buying Guide 2024: Get the Most AI Power for Your Budget
มุมมอง 53K7 หลายเดือนก่อน
AI/ML/DL GPU Buying Guide 2024: Get the Most AI Power for Your Budget
Detectron2 Setup with Docker: Simplify Computer Vision - Fix Bugs & Dependencies (2023)
มุมมอง 9748 หลายเดือนก่อน
Detectron2 Setup with Docker: Simplify Computer Vision - Fix Bugs & Dependencies (2023)
Neo4j - WebSocket Connection Failure - How to Fix!!!
มุมมอง 2K9 หลายเดือนก่อน
Neo4j - WebSocket Connection Failure - How to Fix!!!
Neo4j - Introduction to Querying with Real World AirBnB Data
มุมมอง 1349 หลายเดือนก่อน
Neo4j - Introduction to Querying with Real World AirBnB Data
2FA Setup For Linux Home Lab or Server - Cisco Duo
มุมมอง 1.2K10 หลายเดือนก่อน
2FA Setup For Linux Home Lab or Server - Cisco Duo
Why 3.5" Caddy (Tray) Lights Blink Amber After 2.5" SSD Replacement
มุมมอง 118ปีที่แล้ว
Why 3.5" Caddy (Tray) Lights Blink Amber After 2.5" SSD Replacement
How to Configure a Virtual Drive ?!
มุมมอง 145ปีที่แล้ว
How to Configure a Virtual Drive ?!
Upgrade Your Server's Old 3.5" HDD To 2.5" SSD
มุมมอง 1.9Kปีที่แล้ว
Upgrade Your Server's Old 3.5" HDD To 2.5" SSD
Great Budget PC Build For The Data Science/ML Beginner: P3 - Software
มุมมอง 679ปีที่แล้ว
Great Budget PC Build For The Data Science/ML Beginner: P3 - Software
Great Budget PC Build For The Data Science/ML Beginner: P2 - The Build
มุมมอง 1.6Kปีที่แล้ว
Great Budget PC Build For The Data Science/ML Beginner: P2 - The Build

ความคิดเห็น

  • @gorangagrawal
    @gorangagrawal 2 ชั่วโมงที่ผ่านมา

    If possible, please upload NVLink and PCIe extender video. It would be really helpful to understand them.

  • @Horesmi
    @Horesmi 7 ชั่วโมงที่ผ่านมา

    For some reason, 3090s are going out for $500 over here, and there are a lot of them on the market. Crypto crash or something? Anyway, that seems to change the calculations a lot in my case

    • @publicsectordirect982
      @publicsectordirect982 5 ชั่วโมงที่ผ่านมา

      Over where?

    • @Horesmi
      @Horesmi 5 ชั่วโมงที่ผ่านมา

      @@publicsectordirect982 Ukraine

    • @ericspecullaas2841
      @ericspecullaas2841 ชั่วโมงที่ผ่านมา

      ​@publicsectordirect982 my guess China......

    • @Horesmi
      @Horesmi ชั่วโมงที่ผ่านมา

      @@publicsectordirect982 Ukraine

  • @JimCareyMulligan
    @JimCareyMulligan 9 ชั่วโมงที่ผ่านมา

    Thank you for your work. Do you have any plans to test the tesla v100 16 GB? They are goes for half the price of the 3090 and support nvlink.

  • @werthersoriginal
    @werthersoriginal 9 ชั่วโมงที่ผ่านมา

    Oh wow, I'm in the market for the 3090 for LLMs but I've been eyeing the P40s because of their price. I saw your videos on the R720s and now I'm wondering if I can put a P40 in my R710.

  • @H0mework
    @H0mework 9 ชั่วโมงที่ผ่านมา

    I'm happy whenever you upload

  • @jaroman
    @jaroman 9 ชั่วโมงที่ผ่านมา

    PCI 3.0 vs PCI 4.0 makes a difference in this kind of setups?

  • @scentilatingone2148
    @scentilatingone2148 9 ชั่วโมงที่ผ่านมา

    Brilliant bud.

  • @TheNicCraft25-go8he
    @TheNicCraft25-go8he 23 ชั่วโมงที่ผ่านมา

    Ehrenmann

  • @GrossGeneralization
    @GrossGeneralization วันที่ผ่านมา

    Did you look into whether the 1U bump lid (supermicro part number MCP-230-41806-0N) clear an rtx 3090? Looks like that Zotac card is a lot lower than some other 3090 cards, you might get by with the MCP-230-41803-0N which was designed to clear the GTX cables. (Note that these were part numbers for the 4029GP-TRT2, but the chassis looks pretty much the same).

  • @tsclly2377
    @tsclly2377 วันที่ผ่านมา

    Yup, just got a couple og P40s for a ML350P... after investigating the NVIDIA site.. Slow, yes, but for cheap and something that can run on MS Win2012 Enterprise, it's the ticket. (old mining ETH machine) It will run in the garage without air-conditioning.... Slow but study.

  • @dustintunget4177
    @dustintunget4177 3 วันที่ผ่านมา

    This is still a solid video 5 months later. I ended up with a 4U ATX case and one of the epyc/supermicro combos from ebay, then 2xP40 + a 3060 12gb. Upside is that a 4U ATX rack case can use 120mm fans that can spin slower, and quieter...and you can fit more GPUs, and use standard PSU's. I'd like to see this kind of breakdown in the context of a EATX case (rack or desktop)...those results could be interesting. I think mine was in the realm of $1700...but that's not with much storage....I bought it over time so I don't have solid numbers. Great video, earned a sub!

  • @xxriverbeetxx1.065
    @xxriverbeetxx1.065 5 วันที่ผ่านมา

    Hi Iam really into servers and homela. which usecases does such a ai server have. I run stable diffusion on my pc but I don’t use it regularly. What are some ais you ran. I can’t imagine one which it is worth buying a server?

  • @GrossGeneralization
    @GrossGeneralization 5 วันที่ผ่านมา

    You probably found it already, but looks like one of your ram modules isn't fully seated (approx half way between both CPUs)

  • @tsclly2377
    @tsclly2377 5 วันที่ผ่านมา

    Heat kills.. so is the RTX4090 experience. I'm going slow and low wattage. Also type of SSD must be big petabyte write .. NVLinking is slow, 112Gb/s and is just dumping onto the Cudas/memory in the not really working like type GPU.. especially if that GPU is stuck into a 16lane that is really only 8 lane.. best to get a bigger Vram card like an A6000 @48GB instead of NVLink(2)ing two A5000 24GB, plus the link is going for premo$$$.. Me.. I'm going the 'Potato'/Orin route that doesnt' need a new air-conditioner/heat pump route, under 350W, and fuk the need for speed, I can wait 30 minutes instead of 3. It is not the speed, but the GPU ability to actually do the work and most can't. 48GBvram with 2000 cuda cores is the can do.. top TDP is like money spent. this review is best for MS BitNet 1.58.. so if you are Windows.. rock on... and I've read the NVLink Quadro cables work for RTX cards.. not the LoveLace...the need the LoveLace Link-2s..

    • @TheDataDaddi
      @TheDataDaddi 57 วินาทีที่ผ่านมา

      Hi there. Thanks so much for the comment! Yeah this is definitely a good point. I agree that more VRAM is typically better even if you actually have lower throughput, better to be able to do the task even if its slow than not because you dont have enough VRAM for your usecase.

  • @wasifmasood969
    @wasifmasood969 6 วันที่ผ่านมา

    Hi, I am facing a different challenge now. I have this machine at home and it creates a lot of noise. I have checked the fan setting, even on the "Optimal" mode it is loud but when I start some training, it goes really unbearable. I have checked various fan configurations, it unfortunately has none stating "Quite mode". What else can I do to make it some reasonable to sit next to, are there any high quality fans which can make a difference. Many thanks in advance?

    • @TheDataDaddi
      @TheDataDaddi 4 นาทีที่ผ่านมา

      Hi there. So I have a video actually related to this. th-cam.com/video/RUW3Ay5rsCY/w-d-xo.html Check this out. In the video, I should how to adjust the fans manually. You could put yours manually on the lowest setting and you may be able to turn them off completely. However, please be carefully when doing this. It is never good to run your GPUs or the rest of you system too hot for too long. Another alternative might be to get some snail blower fans and install them that make less noise then lower the naive fans to the lowest setting. Something like the following may work for you. www.ebay.com/itm/186459922797?mkcid=16&mkevt=1&mkrid=711-127632-2357-0&ssspo=2ucjrgqosnm&sssrc=2047675&ssuid=xv3fo9_stiq&widget_ver=artemis&media=COPY

  • @romeo-sierra
    @romeo-sierra 6 วันที่ผ่านมา

    Overall a good video. Not misleading or lazy. Don't take TH-cam comments to heart. People love to say negative things. It's often all they got in life. Only suggestion would be keeping making videos. You'll get better with experience.

    • @TheDataDaddi
      @TheDataDaddi 9 นาทีที่ผ่านมา

      Hi there. Thank you so much for the encouraging feedback. Really appreciate it!

  • @mateuslima788
    @mateuslima788 7 วันที่ผ่านมา

    Thank you for the video!

    • @TheDataDaddi
      @TheDataDaddi 10 นาทีที่ผ่านมา

      Hi there! Of course. So glad that you enjoyed the content!

  • @TheUserIWantedIsTakenAlready
    @TheUserIWantedIsTakenAlready 7 วันที่ผ่านมา

    This is awesome, that’s exactly what I was looking for!

    • @TheDataDaddi
      @TheDataDaddi 10 นาทีที่ผ่านมา

      Hi there. So glad this is what you needed! Thanks for the comment!

  • @HuzMS
    @HuzMS 10 วันที่ผ่านมา

    Thank you for your hard work. Was p40 or p100 the better choice? ALso were you using nvlink?

    • @TheDataDaddi
      @TheDataDaddi 11 นาทีที่ผ่านมา

      Hi there. Thanks so much for the kind words and the comment! It really depends on the use case, but overall for most people I would say that the P40 is actually the better choice. The only GPUs I currently have that support NVlink are my RTX 3090s. Yes, I have them connected via NVLink.

  • @b-ranthatway8066
    @b-ranthatway8066 10 วันที่ผ่านมา

    So would this mean I can't use a 7900XT to make AI meme pictures? I've actually been interested in the whole AI thing, even though I'm not a smart dude on tech. (I find it cool just because I can use my computer for something other than just gaming/streaming/video editing, but I'm going to try a resist a little against our developing AI overlords lol) I know the tech is still developing, but I thought it would be cool to use AI to create a Vtuber model to stream with. (even if it came out bad, I thought it would be a fun little experiment to do for some views and laughs) However one of my hardest parts to upgrade, in my mind, was a GPU. I know AMD is a step or two behind Nvidia (My last card and current card is a 1070) but when it comes to price, it's hard to beat. I just didn't know if something like a 7900XT or -XTX would at least make up for it vs a 4070 TI Super in terms of AI generation. (I still have no idea what app to use to even make use of my GPU to even make stuff with AI) Alright, enough rambling with the thoughts in my brain, I'll keep watching 👌

    • @TheDataDaddi
      @TheDataDaddi 13 นาทีที่ผ่านมา

      Hi there! Thanks so much for the comment. So, my take here is the NVIDIA GPUs are going to be much easier to work with at this stage. I have head from some of my viewers that AMD GPUs can and do work. It is just a lot more of a pain to work around bugs and the learning curve is steeper. NVIDIA is more or less plug and play when it comes to AI/ML/DL, but that is also why you pay a premium. I guess what I would say is if you want the easier route or don't have time to do much trouble shooting NVIDIA might be a better way to go. However, from the sounds of it you are more partial to AMD GPUs and have many other workloads besides just AI. In your case, it may be a better idea to go with AMD GPUs because you will get better price for performance for all of your other workloads then deal with the pain of setting up you AMD GPUs for your specific AI use case. The 7900XT is definitely a powerful card and can handle AI tasks, though you might need to use specific software or frameworks that support AMD GPUs, like ROCm. Creating a VTuber model sounds like an interesting project! I would recommend maybe starting with programs like DeepFaceLab for deepfake-style video or some stable diffusion flavors to generate images as a starting point. For generation, tools like Blender for 3D modeling could be helpful and for real-time animation you might could use VMagicMirror or VSeeFace which can utilize your GPU to bring your VTuber model to life. Hope this helps!

  • @agriculture7188
    @agriculture7188 12 วันที่ผ่านมา

    do you have a recommendation for what gpu(s) I should purchase for my R720? I mostly plan on running LLMs, Stable Diffusion, and image detection models. I don’t have a super high budget and was considering a dual P100 setup but wanted the opinion of someone a little more educated in the ML field.

    • @TheDataDaddi
      @TheDataDaddi 11 วันที่ผ่านมา

      Hi there! Thanks so much for the question. So for all of those things on a low budget. I would probably recommend the p100. You could also go with the p40 for more VRAM. The p100 will handle quantization more efficiently and have high significantly higher throughput which will be important when working with LLMs. The p40 has higher VRAM to start with so you can load larger models, but the fp16 performance is really bad so its through put will be a lot worse (theoretically). As budget options though, I think these could still put some of the smaller open source LLMs within reach for you to start experimenting with. Hope this helps!

  • @noth606
    @noth606 12 วันที่ผ่านมา

    GFLOPS is not calculated like shown in the video at 15:18, remove the Giga which we know, FLOating Point operations per Second, simply(there is some history for why this is used). It is somewhat archaic since a lot if other things are being done too, which aren't incorporated in this but in general most other operations take less cycles than a floating point one does because the comma needs special attention so to speak, 15*15 and 1,5*1,5 are the same thing except for tracking the comma separately with the result being 225 or 2,25. What I mean is the circuit needs additional logic to track commas or rather fractions so to speak, which is why we separate floating point from integer operations - additional hardware is required to track the comma "in top of" the integer type numerical operations. No idea if this makes any sense or is useful, I thought it would be simple to explain until I thought it through and realized I need to type this as opposed to scribble and show on a white board. I'm sure there is a good explanation for it out there, just trying to point to why since to a person doing math it's not as obvious as it is designing a circuit to do it.

    • @TheDataDaddi
      @TheDataDaddi 11 วันที่ผ่านมา

      Hi there. Thanks so much for the comment! This is great information. Thanks so much for sharing. I did know that floating point ops were different fundamentally than other operations though I was not 100% sure why. This makes a lot of sense! Also, if you know the correct formula for calculating FLOPs in general please let me know.

  • @iasplay224
    @iasplay224 16 วันที่ผ่านมา

    how did you download ubuntu on the dell server , I have installed the ubuntu on usb bootable then I installing on a ssd hard drive but I cannot boot into that hard drive did you have that issue or does someone know how to fix it

    • @TheDataDaddi
      @TheDataDaddi 15 วันที่ผ่านมา

      Hi there. Thanks so much for your comment. The process is normally as easy as: 1) Creating a bootable drive with whatever OS you want to install 2) Plug in the usb 3) Power on the machine and boot into BIOS 4) Adjust the bios settings so that the usb device is the first boot choice (although many times you dont even need to do this) 5) Let the machine boot normally, and you should be prompted to install the OS. 6) Install the OS 7) Restart machine and boot into BIOS 8) Adjust bios setting so that the new drive where the OS is installed is the first boot option 9) Let the machine boot normally, and you should have a working fresh install of whatever OS you choose.

    • @TheDataDaddi
      @TheDataDaddi 15 วันที่ผ่านมา

      Let me know if this does not work for you. I can try to help you trouble shoot from there.

    • @iasplay224
      @iasplay224 8 วันที่ผ่านมา

      @@TheDataDaddi I was able to solve the issue but the things is that I originally tried to plug in a 2.5 ssd drive to the back of the server and that didn't work in terms of booting, I was able to install the os on it but when restarting the bios said no bootable drive was found. my solution was to use a usb to Sata adapter and plug that in the internal usb port and that worked and one could then boot, but the drives are detected in ubuntu and can save files to them so it's weird

  • @wasifmasood969
    @wasifmasood969 17 วันที่ผ่านมา

    Hi, I have recently bought this system. I see there are two EPS 8-pin connectors (JPW3 and 5) on the mother board, in addition to the other 8-pin PIC power connectors. My question is that I have bought Tesla M10 32 GB GPU which requires an 8-pin EPS connector. Can I connect that GPU to one of these EPS connectors. The card needs an 8-pin 12V EPS connector. What would you suggest. Many thanks for your amazing support.

    • @TheDataDaddi
      @TheDataDaddi 15 วันที่ผ่านมา

      Hi there. Thanks so much for the question! You should be able to connect if they are the traditional PCIe 8 Pin (6+2) GPU connectors. The 4028GR-TRT has 8 12 EPS power cables that support most Tesla GPUs naturally. From what I can tell, the M10 uses the more traditional 8 pin connector. You could buy an adapter from 12 V EPS to PCIe 8 Pin (6+2) power. I think the following should work for you. a.co/d/hMUEDvs Hope this help you! Please let me know how it goes.

    • @wasifmasood969
      @wasifmasood969 13 วันที่ผ่านมา

      @@TheDataDaddi thanks for your prompt reply. I am wonder if instead of 6+2 pin, I can use 4+4 pin since I have it at home.

    • @TheDataDaddi
      @TheDataDaddi 11 วันที่ผ่านมา

      @@wasifmasood969 Unfortunately, I do not think the 4+4 pin will work in your case. You can certainly try, but it will likely not fit.

  • @sphinctoralcontrol
    @sphinctoralcontrol 17 วันที่ผ่านมา

    Tossing up between 3090s, A4000, and P40/P100 cards for my use case which would not exactly be ML/DL but rather local LLM usage hosted using something the likes of OLlama and various (I assume at least q4) models of higher parameters. I'm also dabbling with Stable Diffusion as well - at the moment I am shocked I'm able to run q4 quantized LLMs via LM Studio as well as Stable Diffusion models, on my little old aging M1 2020 Macbook Air with 16GB ram. I'm getting into the homelab idea, especially the idea of using a Proxmox server to spin up different VMs (including the Mac ecosystem) with way higher resources than what I'm working with currently. I'm also looking to integrate a NAS and other homelab services for media - but the GPU component is where I'm a little hung up - just what tier of card, exactly, is needed for this sort of use case? Am I nuts to think I could run some of the lesser quantized (as in, higher q number) LLMs on the low profile cards, as well as SD? It's been 10+ years since I've build a PC and am totally out of my element in terms of knowing just how good I've got it using the M series of chips - I've even been hearing of people running this sort of setup on a 192GB RAM M2 Ultra Mac Mini Studio, but would really love to get out of the Apple hardware if possible. I realize this was several questions by now... but, to distill this down, GPU thoughts? lol

    • @TheDataDaddi
      @TheDataDaddi 15 วันที่ผ่านมา

      Hi there. Thanks so much for you question! Yeah so this is a really good question. It really depends on the size of model you are trying to run. For example, to host Llama2 70B for FP16 you need approximately 140GB of VRAM. However, you could run quantized versions with much less. Or you could always work with the smaller model sizes. In terms of GPUs, I would recommend GPUs that have at least 24GB VRAM. I have been looking at this a lot for my next build, and I think I actually like the RTX titan best. The RTX 3090 would also be a good choice its FP16 performance just isn't as good. I think the P40/P100 are also great GPUs for the price, but for LLMs specifically they may not be the greatest options because the p100 has only 16 GB of VRAM and the p40 has very poor FP 16 performance. Another off the wall option is to look at the V100 SMX2 32GB. Since these are are SMX2, they are cheaper, but there are a lot fewer servers that they will fit in. The only one I know of off the top of my head is the Dell C4140/C4130. From my research, they the SMX2 GPUs are also fairly tricky to install. Anyway, these are the routes I would go to make a rig to host these models locally. I will eventually build a cluster to host and train these models locally so stay tuned for videos to come on that subject if you are interested

  • @Meoraclee
    @Meoraclee 20 วันที่ผ่านมา

    Um Hi daddy, Im having a trouble with building a pc to train ai (play game sometimes). With 2500$ budget should I aim for 2x3090 to run 48 GB VRAM or 2x4060ti 24GB VRAM ? Is there any better option in my case ?

    • @TheDataDaddi
      @TheDataDaddi 15 วันที่ผ่านมา

      Hi there. Thanks so much for the question! I think it depends a lot on your use case, but I would say if you plan on using it primarily for AI workloads the 3090s would be a better choice because of the higher VRAM and ability to support NVLINK. However, if you want to focus more on gaming and some AI work loads I would choose the 4060s and put more money towards other components like a better CPU. You could also just go with a single RTX 4090. This would give you great performance for AI workloads and gaming with budget enough for other high quality components.

  • @ThePandalars
    @ThePandalars 20 วันที่ผ่านมา

    perun much?

    • @TheDataDaddi
      @TheDataDaddi 15 วันที่ผ่านมา

      Hi there. Thanks so much for the comment! I am not sure I understand the question. If you could give me, a bit more context that would be great!

  • @rohithkyla7595
    @rohithkyla7595 20 วันที่ผ่านมา

    keen about the comparison video!

    • @TheDataDaddi
      @TheDataDaddi 15 วันที่ผ่านมา

      This will be coming very soon! Stay tuned.

  • @vampritt
    @vampritt 24 วันที่ผ่านมา

    omg so detailed.. thank you for your time.

    • @TheDataDaddi
      @TheDataDaddi 21 วันที่ผ่านมา

      Hi there. Thanks so much for you comment. So glad that the content was useful for you!

  • @cyklondx
    @cyklondx 24 วันที่ผ่านมา

    i think you missed out on parts - actual performance per model, and if one can use fp32, fp16 or int8 or tensor. P40 is terrible options for any ai workload due to amount of time one would have to wait... and its power requirement.

    • @TheDataDaddi
      @TheDataDaddi 21 วันที่ผ่านมา

      Hi there. Thanks so much for your comment! I would agree that for anything below fp32 operations these GPUs would be quite slow. However, the GPU is less than $200 dollars for 24GB of VRAM. So, if you are wanting to experiment with larger models cheaply, I think these GPUs still have good value.

  • @chimpera1
    @chimpera1 25 วันที่ผ่านมา

    I'm thinking of building one of these with 8xp100. My knowledge is fairly limited but I want to explore llms. Would you recommend

    • @TheDataDaddi
      @TheDataDaddi 21 วันที่ผ่านมา

      Hi there! I think this would be a good cheaper way to start experimenting! Do be aware through that training or fine tuning most of the largeer open source LLMs will be out of reach even with a setup of this magnitude. However, you could likely host some of the smaller one or quantized versions locally. Hope this helps. Cheers!

  • @chentecorreo
    @chentecorreo 26 วันที่ผ่านมา

    Excelent work!

    • @TheDataDaddi
      @TheDataDaddi 21 วันที่ผ่านมา

      Hi there. Thanks so much for the comment! Really appreciate your positive feedback!

  • @jjolleta
    @jjolleta 27 วันที่ผ่านมา

    I see that cable management and I kinda wonder, what about thermals....... You should find a way to make those cables even more custom for them to fit better and as in the second card below the air ducting. This is just a suggestion, hope you can manage to give those cards a better lifespan. Greetings !!!!

    • @TheDataDaddi
      @TheDataDaddi 21 วันที่ผ่านมา

      Hi there! So I have actually found a better cabling strategy. The updated cabling should be included in the video description. Please have a look there if you are interested in going that route. Cheers!

  • @jaystannard
    @jaystannard หลายเดือนก่อน

    This is a cool little project you did.

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. I am so glad you enjoyed the content! Really really appreciate the donation. Really helps the channel!

  • @vap0rtranz
    @vap0rtranz หลายเดือนก่อน

    I made a spreadsheet too, but yours is thorough! I'd also come to similar conclusions as you: that the P40 / P100 were cheap ways to get medium size LLM models into a GPU with decent tokens/second. Your spreadsheet would have saved me time if I'd known about it! At least there's some independent confirmation of your conclusions. There's a lot of detail to add, like how fast/slow models are on certain GPUs ... perhaps another vid on that to save me the effort? :P

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks so much for the comment! I am glad to here you can confirm! Especially as hardware prices keep increasing I think these are actually becoming even more relevant for those who are budget conscious. Funny you mention this. I am actually working right now on a benchmarking suite to enable reliable comparison between GPUs for different models. There is not a reliable open source benchmarking solution for GPUs so I am trying to create one (or make steps toward it at least). As soon as I get something decent, I will make a video series on it and start using it to benchmark GPUs in a real way with respect to individual models.

  • @kellyheflin5931
    @kellyheflin5931 หลายเดือนก่อน

    What dimensions are necessary for the Supermicro SuperServer 4028GR-TRT to fit in a mid-sized server rack? I'm grateful. Thank you.

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thank you so much for the comment! So, I forget what length I set my rack at, but the Supermicro SuperServer 4028GR-TRT is 29" long so I would set it about 3" to 5" longer than this to comfortably fit the server. To my knowledge, servers are pretty much all the same width. One thing to keep in mind though is not all servers are the same length so if you every buy others in the future they may be longer so it is a good idea to set you rack a few inches bit deeper than the longest server you plan on housing.

  • @jack-nguyen
    @jack-nguyen หลายเดือนก่อน

    i am building one myself with 2x rtx 4090, is there any suggestion for the cpu?

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there! Thanks for the questions. Couple of suggestions/comments here. 1) Pretty much all consumer grade motherboards and CPUs that i know of besides the threadripper series CPUs and compatible motherboards will not be able to support both GPUs at the full x16 lane bandwidth (this route is incredibly expensive). Not having the full x16 lanes for both GPUs might be okay with you. I just wanted to make you aware. With some CPUs you can get have configurations like x16x8 or x8x8. 2) For the CPU, it also depends one whether you want to go the DDR4 or DDR5 route with respect to RAM. I would personally recommend DDR4. It is cheaper, you have more options, and is supposedly more stable at this stage (do some research of your own here as things change very fast these days). I think both AMD and Intel CPUs are fine. I would recommend at least 12 cores, but the exact CPU really depends on you budget. Some cost effective suggestions might be: AMD Ryzen 9 5950X Intel Core i9-12900K

    • @jack-nguyen
      @jack-nguyen หลายเดือนก่อน

      @@TheDataDaddi I also struggle finding the motherboard as you mentioned. I found a solution that using the X11DPI-N mainboard and I have to use intel xeon gold 6138 in this way which only support pcie 3.0. Thank you for your response and I really appreciate it.

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      @@jack-nguyen Yeah I did as well when I was researching for this video. Eventually, I just switched over to using servers because this normally isn't an issue with server mobos and CPUs. I like the idea though! That should CPU and mobo should work great for you. What case/chasis are you planning on using? Of course man! I am always happy to help.

  • @mrbabyhugh
    @mrbabyhugh หลายเดือนก่อน

    29:11 haha, that's the exact card I am looking at. Comparing it with ARC a770 actually.

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks so much for the comment! I have also been interested in non NVIDIA solutions. The ARC GPUs have certain interested me. However, I would caution you. If you leave the NVIDIA ecosystem, it is like going into the wild west so just make sure you are prepared. Here is a Reddit thread that might shed some light. www.reddit.com/r/MachineLearning/comments/z8k1lb/does_anyone_uses_intel_arc_a770_gpu_for_machine/ If you do decide to go the ARC route, please let me know how it goes for you. I would be super curious to better understand where those GPUs are in terms of AI/ML/DL applications.

  • @PreparelikeJoseph
    @PreparelikeJoseph หลายเดือนก่อน

    Is using a server build better than a PC build or were parts just cheaper ?

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks so much for the question! The reasons I prefer servers over OC builds: 1) Price for compute is almost always better (main reason) 2) Ability to support more cores and higher CPU RAM 3) Remote management tools like IPMI or IDRAC 4) Generally more stable and built to run forever with out being turned off 5) Built in redundant power supplies and failovers That said. A custom build is always going to be more flexible and likely give you the ability to have the latest and greatest. Overall though, I like refurbed servers because I find they provide the price to performance.

  • @SanjaySharma-ov1kf
    @SanjaySharma-ov1kf หลายเดือนก่อน

    Hi @TheDataDaddi, I have couple of questions if you don't mind to help. One I want to boot from PCIE card with NVME SSD, but the NVME SSS is not recognized in the BIOS, but it works when I try to access using Ubuntu. And second issue is that I am not use the 3060 and 3080 GPU on this server, it seem that the power cable for GPU is different on server and not compatible with 3060 and 3080 GPU. I tired new PCIE power cable from Amazon, but it didn't help. Can you please help on these two issues?

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks so much for your questions! 1) For the NVME issue you will likely need to go into the bios and enable pcie bifurcation x4x4 or x4x4x4x4 depending on which pcie slot you have it in. If I remember correctly you will need to boot into bios. Then go to chipset configuration>north bridge and here you can change the pcie slot configurations. If you google super micro 4028GR-TRT block diagram this should should you which pcie slots are which. This this and let me know if it works for you. 2) The native power cable in the server is a male 12V EPS connector. This is designed to be used with Telsa series GPUs. However, you can by an adapter that will convert into the 8 pin connector that the RTX series GPUs expect. I believe the following should work for you: a.co/d/fQ9ncia Please let me know if you have any other questions!

    • @SanjaySharma-ov1kf
      @SanjaySharma-ov1kf หลายเดือนก่อน

      @@TheDataDaddi Thank you for the quick response.Really appreciate your help and support. I did change the BIOS setting, Under North bridge 4x4 and 4x4x4x4 but still not able to see the NVME SSD in the boot option. Can you please help to resolve the issue? I have ordered the cable from Amazon, hopefully that will fix the 3080 GPU issue :)

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      @@SanjaySharma-ov1kf Sure! Also make sure that under PCIE/PCI/PnP section the PCIE slot is enabled as UEFI. Another issue might be that you have not have set the bifurcation for the correct PCIE slot. Try it for other possible candidates if you have not already. Even with the block diagrams it can be a bit confusing to figure out which PCIE slot is which. Unfortunately, it could also be bad hardware. I order 1 4 slot PCIE to NVME adapter, and it did not work originally due to a hardware issue.

    • @SanjaySharma-ov1kf
      @SanjaySharma-ov1kf หลายเดือนก่อน

      @@TheDataDaddi Thanks for the hehlp. I will try another options as well for bifurcation. Under PCIE/PCI/PNP I do not have option UEFI but EFI. The PCIE slot is working as I did install Ubuntu on it by booting from Ventoy USB but the NVME drive is not visible while booting.

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      @@SanjaySharma-ov1kf Sorry I meant EFI. How are you connecting the NVME drive to the PCIE slot?

  • @Rico0333
    @Rico0333 หลายเดือนก่อน

    Does the GPU have a CPU Ram requirement? I'm getting not enough resources error in windows.

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks for the question! So in theory there is not hard and fast requirement for a CPU RAM based on a particular GPU. That said you need enough RAM to hold whatever it is you are trying to load on your GPUs. For example, if you have 16GB of CPU RAM and 32GB of GPU VRAM. If you try to load a 24GB model onto the GPUs you will likely get errors, because even though you have enough GPU VRAM to hold the model you can't fit it into CPU RAM first in order to load it. I am not sure if this is your problem? If not please let me know, and I can dig deeper to try to help you figure out where the error may be coming from.

  •  หลายเดือนก่อน

    In wich OS did you run the nvidia-smi command ?

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks for the question! I am using Ubuntu 22.04 for all of my servers at the current moment. Might switch over to NixOS soon though.

  • @publicsectordirect982
    @publicsectordirect982 หลายเดือนก่อน

    Hello I'm new to your channel. Really useful information thanks! I'm just getting into ml and was wondering if this set up would allow a 40b llm to be loaded? Or what might be the best solution to run 4x tesla p40? Thanks for any potential tips

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks so much for the comment and so glad you are enjoying the content! Take what I say with a grain of salt because I have not actually tried this myself, but for a 40 billion parameter model at half precision it would roughly require about 80GB to load the associated parameters then there will be some extra memory overhead for other process and space to load data to run the actual inference. So, to be safe lets approximate 100GB ish of VRAM will be needed. With 4 P40 GPUs (4 x 24 = 96GB) you would in theory be able to run inference locally. Caveat here is you would need to use model sharding to split the overall model across all of the GPUs. This should also be theoretically possible. One other thing to note is that the half precision of the P40s is very low. This will likely mean that you will have much longer inference times. Some other options, you could consider would be going with 4 3090s instead or going with 6 (or more) P100s. This should give you the VRAM you need for your use case while also providing better performance. It seems like the 6 p100s would be the best balance of cost and performance. You could also consider using a smaller model and running it at the full precision to better take advantage of the p40s design. So, it seems that for your use case you will need a minimum of 4 GPUs so you will need a different server than the one discussed in this video that can hold at least 4 2 slot GPUs. If you would like some recommendations, here please let me know.

  • @DarrenTarmey
    @DarrenTarmey หลายเดือนก่อน

    Can I get access to this spreadsheet

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks so much for reaching out. You are should be able to download a local copy and change it in anyway that you see fit. I acknowledge that the spreadsheet needs to be updated. I am working right now on a website actually that lists GPU specs and then keeps track of historical price trends. So all of the information in the spread sheet and more should be available updated on a daily basis soon. Unfortunately, I am not comfortable giving you direct editable access to the original version in the Google drive. I apologize. It is nothing personal. I just don't know you well enough.

  • @iskiiwizz536
    @iskiiwizz536 หลายเดือนก่อน

    Thanks for the work man ! Love from France

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks so much for the comment. So glad you enjoyed the video!

  • @emiribrahimbegovic813
    @emiribrahimbegovic813 หลายเดือนก่อน

    what kind of AI work do you do? are you able to run and/or train on llama3 with 70B parameter? also wow that is loud, is it this server only or all your servers together? I was thinking to do the same thing in my garage (which coincidentally is the place I do most of the work from), but this might be too loud for me.

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks so much for the question. Currently, for my main body of research around browser security I mainly work in computer vision (ConvNets) and also with some of the newer multimodal models like CLIP and FLAVA. In a separate project, I am working with Graph Neural Nets and Graph Databases to ingest and analyze the entire BTC ledger. As of yet, I have not played around much with any of the open source LLMs on this machine because frankly I just have not had the time. It is high on my list of things to do though. So, unfortunately, I cannot tell you exactly what works and what doesn't concretely because I do not have real data to support it. What I can tell you though is that even with 8 24GB GPUs will only get you to 192 GB VRAM. Llama3 70B will need about 140GB for half precision (back of the envelope here) for the model parameters. So pre training or fine tuning are likely out of reach. Honestly, it seems like if you really wanted to pre train or fine tune any of the really large LLMs you would need a distributed cluster. I would love to do this one day, but that will require a lot more time and funds. lol. That said, I do believe that hosting the Llama3 70B locally for inference with model sharding (splitting parts of the model across different GPUs) should be possible with the setup in the video (depending on what GPUs you choose to add of course). Especially if you are okay with some level of quantization, I think it should be highly manageable on this rig. I will try my best to do some experimentation here very soon and make a video on the results. You are not the only who is very interested in this use case. Yes. Lol. This server is loud indeed. It is probably the loudest one I have ever worked with. The noise is not unbearable (for me at least anyway) unless the machine is booting or under heavy load (running 4+ GPUs at max capacity at once). I would definitely not suggest putting it anywhere you will be working all the time as it could get annoying especially if you are constantly running experiments. What I normally do is if I need to work next to the server for extended periods of time, I just wear headphones (mine are over the ear and noise canceling, but any should probably help). This might help you if you don't have any other options, but to work in the garage with the servers frequently.

    • @emiribrahimbegovic813
      @emiribrahimbegovic813 หลายเดือนก่อน

      @@TheDataDaddi thank you so much for responding, great stuff, great channel.

  • @mahdi-oe6mk
    @mahdi-oe6mk หลายเดือนก่อน

    Wonderful video, i started with machine learning and did some small projects for myself now i want to dive deeper into it, I have 2 config on the table: 3090 second handed with i5 13500 And 4060 ti 16gb with i5 13400 The 4060 ti setup is 330-350 dollars cheaper. Which one do you suggest do you think 3090 and i5 13500 would be much greater or not?

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there! So glad you enjoyed the video! I think both are good options, but I think it depends on your use case and the size of VRAM that you need. The VRAM size of the 4060 ti will be enough for most ML/DL outside of some diffusion related and LLM stuff. If you are wanted to go for larger models then the larger VRAM of the 3090 might be worth the extra money. Personally, I love the 3090s, and I normally always recommend getting the most VRAM you can afford. However, if you are not planning on using the extra VRAM then the 4060 ti will also be a great option. Hope this helps!

  • @wasifmasood969
    @wasifmasood969 หลายเดือนก่อน

    Really awesome stuff, can you pls also recommend which type of server rack should be bought for this?

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. So glad to hear you are enjoying the content! This is the rack I purchased (link below). I have 3 2U servers, the 4U 4028GR-TRT, a 1U switch, and a 1U PSU. It has worked really and been very solid. However, it is a bit expensive. You could probably find something cheaper possibly on EBAY or Facebook market place. www.ebay.com/itm/134077093756

    • @wasifmasood969
      @wasifmasood969 หลายเดือนก่อน

      @@TheDataDaddi thanks, can you pls recheck the link you shared?

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      @@wasifmasood969 Opps sorry. Wrong link. Here is the correct one. a.co/d/18HuxQG

  • @skrillexdj79
    @skrillexdj79 หลายเดือนก่อน

    Is this the same for the telsa k80 for the cable to power in the server

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks for the question! Yes, the same cabling should work for the K80s as well.

  • @gurshansingh6869
    @gurshansingh6869 หลายเดือนก่อน

    Are nvidia gpu better than amd for data science I just rejected an 7900 GRE for 4070 so i hope it is

    • @TheDataDaddi
      @TheDataDaddi หลายเดือนก่อน

      Hi there. Thanks for the comment! I actually just had an in-depth conversation with one of my viewer recently who when the AMD route. He basically said there was a steep learning curve and still a good number of bugs. That said from what I understand AMD GPUs are becoming more viable alternatives for particular AI/ML/DL applications. At this stage though, I still think its too early for most people. I still believe NVIDIA GPUs are the way to go for now.