Unboxing 100Gb NIC | How to setup Mellanox CX455A in CentOS 8

แชร์
ฝัง
  • เผยแพร่เมื่อ 15 ก.ย. 2024

ความคิดเห็น • 57

  • @HAKA...
    @HAKA... 3 ปีที่แล้ว +24

    For anyone thinking about 10/40/100Gbps you need to think about how fast your storage runs, lot of people install 10+Gbps and then complain that the speeds are not what they see in iperf.

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว +2

      Good point!

    • @drtweak87
      @drtweak87 3 ปีที่แล้ว +1

      Was about to say the same thing! Unless you got 100+ Gb of Storage, and to get that you probably need some mass array of SSD's, no point of doing 100Gb.

    • @fiskfisk33
      @fiskfisk33 2 ปีที่แล้ว +2

      I upgraded to 10 at home because the 1G was really easy to saturate, even with my spinning rust. I'm not close to saturating the 10 but it was still a worthwile upgrade

  • @andrewkaplan149
    @andrewkaplan149 3 ปีที่แล้ว +4

    That little kitty background is amazing. Thanks!

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว

      Thanks for watching! :-)

  • @naveedrizvi07
    @naveedrizvi07 3 หลายเดือนก่อน +1

    Great video, I Loved it. I really liked the fact that started from the scratch.. unboxing card, physically installing it on the server. These small things means a lot for new guys, like me. THANKS!!!

    • @ArtofServer
      @ArtofServer  3 หลายเดือนก่อน

      Glad you liked it!

  • @RootSwitch
    @RootSwitch 3 ปีที่แล้ว +6

    Great video. Was going to say "the OFED driver install offers to update the firmware", but then you covered that using the HP specific firmware. The built in FW update does work pretty slick for OEM Mellanox ConnectX cards in my experience.
    While admittedly 100Gb/s direct attach IB isnt terribly practical for most uses, it is a solid way to simulate a 2 node HPC or machine learning environment that leverages RDMA. With IB cards, OFED, CUDA, NCCL, OpenSM, and an Nvidia GPU on both servers, you basically have the basis for a mini compute cluster. The 40Gb/s ConnectX3 cards are cheap. Helped me a ton when I was trying to wrap my mind around IB and a 3 node Nvidia DGX2 cluster I assisted in deploying.

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว

      Did you have that in a ring topology or with a switch in a hub and spoke topology?
      Glad you enjoyed it. Thanks for watching!

    • @RootSwitch
      @RootSwitch 3 ปีที่แล้ว +1

      @@ArtofServer So I wound up only using 2 nodes so they were directly attached with just one IB port. Nvidias NCCL for testing GPU to GPU communication assumes that every node can talk to eachother over all active IB ports, so I never got a ring of 3 nodes to work. I thought about picking up a Mellanox branded 40Gb IB switch to emulate the 100Gb switch that the DGX deployment had, but ultimately decided against it.

  • @501Bakky
    @501Bakky 3 ปีที่แล้ว +2

    Very cool!

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว +1

      Thanks for watching!

  • @pg_usa
    @pg_usa 2 หลายเดือนก่อน +1

    Each lane 25Gbit and the connector aggregate to 4x25Gb so if you try with 4 parallel you get the same results like 8 lane. Or up to the cable it's qsfp or osfp. That's why you don't get much than 25Gbit on one lane.

    • @ArtofServer
      @ArtofServer  หลายเดือนก่อน

      Thanks for sharing your thoughts!

  • @ryanmalone2681
    @ryanmalone2681 3 หลายเดือนก่อน +1

    There are plenty of 100G switches for $1,000-1,600, but they look complicated AF to setup and god help you if something goes wrong. I was looking at an Arista 100G switch and the user manual was over 3,000 pages!

    • @ArtofServer
      @ArtofServer  2 หลายเดือนก่อน

      Well, depends if those 3000 pages are mostly pictures or text, and font size. LOL I have not messed with 100G switch so I have no clue.

    • @ryanmalone2681
      @ryanmalone2681 2 หลายเดือนก่อน

      @@ArtofServer Small text and absolutely no diagrams. Google it. Anything is possible, but how much time do you have to research and try new things is the real question.

  • @guillepunx
    @guillepunx 3 ปีที่แล้ว +1

    I don't know why but youtube stop to show me when you publish a new video. I had to open your channel to see that you have been uploading videos for the last month. And I'm subscribed to your channel. :|

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว +1

      Well, thanks for checking my channel to see new videos! Make sure you hit the notification bell next to the subscribe button and select "all". If you already did that, could just be a glitch on TH-cam.

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว +1

      By the way, I try to release 1 video every week on Friday at 6:30am pacific time.

  • @andriitarykin9567
    @andriitarykin9567 3 ปีที่แล้ว +1

    Thank you!

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว

      Thanks for watching!

  • @drtweak87
    @drtweak87 3 ปีที่แล้ว +2

    One other thing to point out form what i remember on LTTs channel is also CPU bottlenecking. I remember when they were doing either 10 or 40Gb networking they had a LOT of issues getting there and issue was due to multiple things needed to be configured to get that like enabling jumbo frames and other tweaking and also that the CPU's couldn't even handle it and had to upgrade the CPU as well. But sure what CPUs you got but you got two, but then wouldn't all that bandwidth be tied to the CPU on which the card is tired to on the PCIe lanes?? at least one would think.

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว +1

      yes, as you saw in this video, a single thread iperf run could not come close to the 100Gbps. if you want to push more throughput per thread, then you definitely need to do some more tuning; things such as NUMA / core affinity (to avoid QPI latency), interrupt affinity, jumbo frames, increase socket buffers (to allow large TCP window sizes), disabling power management functions to sustain high clock speeds, etc.
      the benchmark run in this video was done without any of that tuning, but obviously takes advantages of using multiple threads to transfer data in parallel, and was able to hit the numbers i showed.

  • @andyhello23
    @andyhello23 3 ปีที่แล้ว +2

    Kind of mute at the moment, unless you are running loads of ssds or nvme in some sort of raid. Your hard disks simply will not take advantage of these speeds.
    Neat to know this tech is around though.
    Thanks for the video.

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว +1

      Good point about storage I/O being matched to the network I/O. NVMe is getting fairly cheap now, so it's not too hard to get fast storage that can handle 90-100Gbps I/O.

    • @ewenchan1239
      @ewenchan1239 3 ปีที่แล้ว +2

      Your point is well taken, but that isn't the only thing that this is used for.
      For me, I use it to solve distributed sparse matrices where the problem is split between my nodes, and the matrix resides in RAM (128 GB per node, 512 GB total) and so, in order to "link" the memory together, fast enough, you will need something like this to accomplish that.
      My finite element analysis solutions can use about 80% of the 100 Gbps bandwidth capacity when it is running a simulation/working on solving a problem for me, and the results can be relatively small compared to the total volume of data that gets pushed through my Infiniband network during the course of the solution process.

  • @fengchouli4895
    @fengchouli4895 3 ปีที่แล้ว +1

    Sigh, 100Gb NICS are fast but expensive. 40/56Gb might be more affordable. BTW, I found you used HP NICs/HBAs in your videos. Thought I am using DELL's machines too, is there any chance that we may see videos about HP servers in the future?

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว +2

      If I happen to acquire some HPE servers from decommissioned DC, maybe. But I'm not a big fan of HPE servers so I wouldn't go out of my way to get some.

    • @ewenchan1239
      @ewenchan1239 3 ปีที่แล้ว +2

      On a $/Gbps basis, 100G actually isn't that bad.
      I think that it's actually a better deal than 10GbE.
      But that will depend on whether you actually need (or want) 100 Gbps.
      I have a micro HPC cluster at home, in my basement, and at least one of my applications WILL use around 80% of the 100 Gbps capacity when it is working on solving a problem for me, so unlike probably most people who might only use the capacity ONCE in a while, when I am running a simulation, I will use the capacity each and every single time and for a while, my micro cluster was pretty busy that I had simulations scheduled out 4 months in advance.
      So it is really going to depend on whether you will be able to make use of it.
      The most painful and expensive part, I think, is actually the switch (depending on whether you're going to be getting an Infiniband switch or an ethernet switch because for some stupid reason, Mellanox decided to charge customers an extra 50% for 100GbE vs. 100 Gbps Infiniband despite the fact that they own the technology where you can switch the ports from IB to ETH, which they could have just included that feature on their switches so that you can run both, but they want more money from you, so they DIDN'T do that).
      And then on top of that, if you are getting an Infiniband switch from Mellanox, you have to decide on whether you are going to be getting a managed switch or an externally managed switch. I initially bought a used, managed switch, but it was having a problem where it would reboot itself every hour, on the hour, so I returned that and bought a used externally managed switch instead.
      The upside of the externally managed switch was that it was cheaper than the managed switch. The only downside was that I had to have a Linux system run the subnet manager (OpenSM) and THAT ONLY runs in Linux.
      So, say if you want to run an all Windows lab, you'll either need to deploy a Linux system (like this) JUST to run the subnet manager OR you'll have to pay the premium to get the 100 GbE ethernet switch instead if you aren't going to be running Infiniband on Windows.
      The second most expensive thing that you will pay for will be the cables because if you have short distances, then the DAC that is shown will work, but if you need to run longer distances, then you need at least passive optic fibre cables, which costs more than the DAC cables.

  • @dianputrijkt
    @dianputrijkt 3 ปีที่แล้ว +1

    :-) Omg thanks for sharing

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว

      Thanks for watching!

  • @kaagyk3386
    @kaagyk3386 3 ปีที่แล้ว

    wait for RoCE ,,hope next video

  • @RepaireroftheBreach
    @RepaireroftheBreach ปีที่แล้ว

    I am thinking to move to Ubunto to achieve the 100G speed of my Network. Windows is only giving me 50 Gb/s write and 14 Gb/s read. And I need the Read speeds the most. I think SMB protocol may be at fault here. What network protocol are you using in CentOS?

  • @II_superluminal_II
    @II_superluminal_II 2 ปีที่แล้ว +1

    Hey is this mellanox CX455A compatible with the Poweredge R820 by any-chance, I have 1 R82 and 1 R920 and wanted to interconnect them through a 100gb/s mellanox link and then connect the master to a 10gb switch for a HPC cluster in my basement to run some simulations. I was wondering if this would at all possible, Couldn't find if the CX455A would work in the R820 PCIE-3 slot???? thanks for making these videos, you helped me rescue my R820 :) you should make a discord channel or something

    • @ArtofServer
      @ArtofServer  2 ปีที่แล้ว +1

      What makes you think it would not be compatible? It's just a PCIe device...

    • @II_superluminal_II
      @II_superluminal_II 2 ปีที่แล้ว

      @@ArtofServer idk especially with dell servers, everything seems finicky. Never owned a server, wanted to use it as a home security and homelab setup

    • @ArtofServer
      @ArtofServer  2 ปีที่แล้ว +1

      @@II_superluminal_II I've never had a PCIe card that didn't work in a Dell server. I've had issues like that in HPE stuff, but not Dell.

    • @Wolgorboy
      @Wolgorboy 2 ปีที่แล้ว

      @@ArtofServer Well, I just managed to buy a HPE 620QSFP28, that only works in ProLiant servers :( I hope Mellanox will work

  • @learnwot4131
    @learnwot4131 3 ปีที่แล้ว +4

    Sadly I can give you only one thumb... Tnx

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว

      Thanks! Well, as long as you hit the like button an odd number of times, I'm ok with it! ;-)

  • @ritzmannaxie284
    @ritzmannaxie284 3 ปีที่แล้ว

    Hi, is there any 40gb cheap infiniband card, like Mellanox, with runs on Ubuntu 20.04 ?

  • @pcb7377
    @pcb7377 10 หลายเดือนก่อน

    What happens if you add information to the DAC cable that it is 2 times longer (the cable is 1 m - we sew it to be 2 m).
    Will it work

    • @ArtofServer
      @ArtofServer  10 หลายเดือนก่อน

      what do you mean by "add information to the DAC cable" ?

    • @pcb7377
      @pcb7377 10 หลายเดือนก่อน

      @@ArtofServer Thank you for responding!
      Each DAC cable at the ends has an EEPROM chip. In it lies information about the cable, about its length! I want to change the value in the field of the cable length (I want to increase 2 times). The question is whether such a cable will work. A cable with a modified (programmatically) length 2 times. There are people who can try it?

  • @FaizanAli-gg5qu
    @FaizanAli-gg5qu 3 ปีที่แล้ว +1

    dear sir
    Art of Server that is good thing you're doing for us,we are waiting your complete networking series make the all videos sequences wise, bring more complete networking series such windows server,cisco,huawei,virtualliztion,free nas,firewall and so on.l mean to share your IT knowledge on your youtube channel that you know, many people want learn complete networking series make all complete series step by step.if you will fellow my idea soon you will get more subscribes.add social media links for ask question & help .

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว

      Thanks for the suggestions! :-)

  • @richardwu8225
    @richardwu8225 ปีที่แล้ว

    Can I connect two nodes with infiniband mode? or it needs the infiniband switch? Thanks!

    • @ArtofServer
      @ArtofServer  ปีที่แล้ว +1

      I think so, but I don't know much about infiniband.

  • @merlin3717
    @merlin3717 3 ปีที่แล้ว

    Hi I don’t suppose you know how to get the dell repository manager on Ubuntu?

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว

      I don't use Ubuntu, but I have used DRM in Fedora.. it should work the same in any Linux OS as it's a monstrous java application.

  • @guywhoknows
    @guywhoknows 3 ปีที่แล้ว

    Well there goes high speed storage and system bus...
    Tbh. As files size has grown and codecs are not greater, mass data migration is somewhat a pain.
    I think, a 4 X 10gbe would "deal" is a lot of useage in common computing And in enterprise,.
    It was funny between my server and work station...
    I just install DAS to the work station, and echo for offline backup. But sure TCP/UDP is a good way to transfer files.
    What is the range?

    • @ArtofServer
      @ArtofServer  3 ปีที่แล้ว

      SFP28 DAC cables can go up to 5-meters I believe. If you use optical transceivers, obviously the range can be much longer.

    • @guywhoknows
      @guywhoknows 3 ปีที่แล้ว

      @@ArtofServer I assume that the standard cable and DAC would make the range lower, as with higher data rates usually means shorter distances.
      Passive and active... Would make a difference.