Proxmox 8 cluster setup with ceph and HA

แชร์
ฝัง
  • เผยแพร่เมื่อ 18 ธ.ค. 2024

ความคิดเห็น •

  • @YofuMedia
    @YofuMedia ปีที่แล้ว +5

    Nice video man... thanks for sharing!!!🎉🎉🎉🎉

  • @raulgil8207
    @raulgil8207 ปีที่แล้ว +27

    In 14 minutes you have explained and demonstrated better than many 1 hour videos, precise and without many entanglements, thank you very much

    • @distrodomain
      @distrodomain  ปีที่แล้ว +2

      You're very welcome, I'm glad you find it useful! Thank You for watching :)

    • @benardmensah7688
      @benardmensah7688 11 หลายเดือนก่อน +2

      He certainly did

    • @hidayatmn6688
      @hidayatmn6688 10 หลายเดือนก่อน +1

      Agree this comment, thanks for the video

  • @johnmarotta6879
    @johnmarotta6879 10 วันที่ผ่านมา +1

    Awesome video. Just need a video on cluster recovery now. :)

    • @distrodomain
      @distrodomain  3 วันที่ผ่านมา +1

      Yes, I will work on that, thanks for the suggestion!

  • @eintyp9201
    @eintyp9201 10 หลายเดือนก่อน +1

    Finally a video like this. Not too much bla bla but not too fast. in 15 minutes all infos you need

    • @distrodomain
      @distrodomain  10 หลายเดือนก่อน

      I'm glad it was helpful!, thank you for watching

  • @IndigoVFX
    @IndigoVFX 3 หลายเดือนก่อน +1

    Instant subscription. Concise pertinent and well paced information.
    Thank you - now I’m off to hunt down some more of your content 😁

    • @distrodomain
      @distrodomain  3 หลายเดือนก่อน

      Welcome aboard! I'm glad you found it useful, thank you for watching

  • @Zbyszekakaduk
    @Zbyszekakaduk 9 หลายเดือนก่อน +2

    ...a big thumbs up from me :) Your video helped me a lot. Others are right, a pill of specific knowledge in 15 minutes. And I don't speak English well and I've been 60 for a month :)
    Best regards, Sedece :)

    • @distrodomain
      @distrodomain  9 หลายเดือนก่อน

      Glad it was helpful! thank you for watching :)

    • @Zbyszekakaduk
      @Zbyszekakaduk 9 หลายเดือนก่อน +1

      Sorry, I have to add something :)
      What I missed in your video was the appropriate network configuration that will enable the best operation of the HA cluster. I had to delve into the documentation. It's not a defect of the video, but I had a problem with data transfer. They recommend 10Gbps or more...
      Regards.

  • @selcukkaraca6154
    @selcukkaraca6154 11 หลายเดือนก่อน +1

    great video. Now it is completely clear to configure HA Proxmox VE with CEPH!

    • @distrodomain
      @distrodomain  11 หลายเดือนก่อน

      Glad to hear, thank you for watching! :)

  • @KewlkatNDahat
    @KewlkatNDahat 6 หลายเดือนก่อน +1

    Just got into Proxmox and was wondering how you go about clustering....etc..... Excellent Video to get me started

    • @distrodomain
      @distrodomain  6 หลายเดือนก่อน

      I'm glad it was helpful, thank you for watching!

  • @nitka_it
    @nitka_it ปีที่แล้ว +6

    LOL Great Job 👍 Clear without bullsh*t. Thank You v.much.

  • @scorpion_lux
    @scorpion_lux 5 หลายเดือนก่อน +1

    What’s. Great content. It was what I was looking for. For my home lab thank you so much.

    • @distrodomain
      @distrodomain  5 หลายเดือนก่อน +1

      Glad it was helpful, thank you for wtching!

  • @JeffersonSantos-mi1vh
    @JeffersonSantos-mi1vh 9 หลายเดือนก่อน +1

    Great video, straight to the point! Liked and Subscribed!

    • @distrodomain
      @distrodomain  9 หลายเดือนก่อน

      I'm glad it was helpful! thanks for watching :)

  • @mattb474
    @mattb474 9 หลายเดือนก่อน +1

    Just a note when you tested migration - there was a single ping downtime, MacOS pings miss a seq number (before going to host unreachable) - they dont show timeouts like Windows. Still a great demonstration though.

    • @distrodomain
      @distrodomain  8 หลายเดือนก่อน

      Yea I realized after the fact, indeed I lost 1 ping during the migration not too shabby tho, thank you for watching! :)

  • @subashjena2326
    @subashjena2326 2 หลายเดือนก่อน +1

    Thank you for this video. My question is why you have taken 3 nodes? Why not 2?

    • @distrodomain
      @distrodomain  2 หลายเดือนก่อน

      3 nodes is usally the standard for redundancy in producrion enviromnets to more easliy do maintenances on 1 node while offloading all the vms to the other 2, but a 2 node cluster is also possible, thank you for watching!

    • @subashjena2326
      @subashjena2326 หลายเดือนก่อน

      @@distrodomain Thank you for smart support.

  • @Uni12355
    @Uni12355 11 หลายเดือนก่อน +1

    thanks for the explanation & appreciated with the tutorials . I just get confused with the HA , I thought setting up & join cluster the proxmox will have cluster . creating zfs storage in one of the node wont have HA right ? So I need to create OSD drives pool only it will have HA function ? If OSD drive is conventional hard drive lets say 16TB non ssd drive will the performance drop & have issues ? Thanks

    • @distrodomain
      @distrodomain  11 หลายเดือนก่อน

      HA will only work if you have ceph configured with more than 1 drive, or if you are using zfs replication across multiple drives, for the 16tb drive your vms sill have the performance of 1 drive so around 100MB/s r/w. Thank You for watching :)

  • @loupitou06fl
    @loupitou06fl 4 หลายเดือนก่อน

    Excellent video, clear and concise. 2 questions, can you HA a VM that has a PCI passthru if you have it on several but not all nodes ? And what happen if you have more storage on one node than others in ceph ?

    • @distrodomain
      @distrodomain  3 หลายเดือนก่อน

      for pcie passthrough I haven't fully tested it but I believe is not supported, ceph will make the volumes as big as your smaller drive that's why you can almost identical drives on each node, thank you for watching!

  • @gustavoroo
    @gustavoroo 8 หลายเดือนก่อน +2

    in my cluster... the VM shuts down and then turns on during the migration. why?

    • @distrodomain
      @distrodomain  8 หลายเดือนก่อน +1

      do you have dedicated storage on each node for ceph, what does your setup looks like, thank you for watching! :)

    • @benjaminwolke9280
      @benjaminwolke9280 7 หลายเดือนก่อน +2

      If you shut down the node, the VM has to restart. It's also shown in this video, that Rocky VM is restarting. Only if all nodes are up, you can do live migration without restarting. The reason is the content in the RAM of the machine with the node that's shutting down. It cannot be copied if there is power loss in one node. On a planned shutdown you can give a setting in the last HA part of the video, that you set it to 'migrate', if I remember correctly.

    • @benjaminwolke9280
      @benjaminwolke9280 7 หลายเดือนก่อน +2

      13:35 request state, I think you can set it to 'migrate' there. Not 100% sure. But if I remember correctly, the VMs are auto-migrated before the machine is shutting down.

    • @distrodomain
      @distrodomain  7 หลายเดือนก่อน

      Yes if the node loses all power all of a sudden the vm needs to restart, the contents of ram would not have time for migrate, but the vm harddrive should be replicated in the other nodes with ceph, so the vm should be able to boot back up. I will do some testings to confirm. Thank You for watching ! :)

  • @scorpion_lux
    @scorpion_lux 4 หลายเดือนก่อน +1

    I have a question ? can it be possible when a node fail can be transfer just for a specific node? and where i can change that ? thanks for you support on your videos

    • @distrodomain
      @distrodomain  4 หลายเดือนก่อน +1

      Yes is possible you can play around with the HA groups this will enable you to spesify a host, thanks for watching!

  • @spvideos4088
    @spvideos4088 9 หลายเดือนก่อน +1

    One of the best demonstrations I've ever seen. Even though I've never worked on Proxmox before, I understood everything you just did. The only drawback is if you use the Community Edition in a disconnected network (no internet connection), how are you going to install Ceph? Can you shed some light on that?

    • @distrodomain
      @distrodomain  8 หลายเดือนก่อน

      You would need internet to update and install paackages fisrt, if that is not an option then creating a local repo of the proxmox repository and configuring that new repo on the nodes.

  • @jesusleguiza77
    @jesusleguiza77 2 หลายเดือนก่อน

    Hi, what configuration should I apply to enable the minimum number of available copies to be 1 for a cluster with 2 nodes and a qdevice? Cheers.

  • @cammac648
    @cammac648 11 หลายเดือนก่อน +1

    great tutorial clear and concise 👍

    • @distrodomain
      @distrodomain  11 หลายเดือนก่อน

      Thank you for watching! :)

  • @emanuelistudor7994
    @emanuelistudor7994 11 หลายเดือนก่อน +1

    nice work, noob question coming.. in this example you used 3x 480GB discs for your 3 OSD's ... does that give you a total useable pool size of roughly 1.4TB (stripe)? Or Is it giving you the useable total pool size of 1 drive? (i.e. ceph consumes 3 HD's but only 1 HD useable total space)? I ask before doing this on 3x 24-bay servers... want to make sure I have enough useable storage.

    • @distrodomain
      @distrodomain  11 หลายเดือนก่อน

      It gives the total amount of 1 drive, think about it as a 3 way mirror where the data and vm disks are replicated on the 3 drives, thank you for watching :), let me know if you have any other questions.

    • @emanuelistudor7994
      @emanuelistudor7994 11 หลายเดือนก่อน +1

      @@distrodomain my understanding was that ceph got better / faster / larger with each node. Does adding a 4th 480GB node make it a 4-way mirror? If so how would one ever go about increasing the pool size aside from adding 1 extra hdd to every node? What happens if you only add 1 extra hdd to one node but not the others? thanks again!

    • @distrodomain
      @distrodomain  11 หลายเดือนก่อน

      you can add additional disks to each node like another 480gb ssd, to have 2 on each node, for this new drive you create an osd the same way and add it to the pool, to keep things consistent I recommend using same amount of drive and capacity in each node.

  • @richardgomes5420
    @richardgomes5420 ปีที่แล้ว +1

    @Distro Domain: You didn't show the creation of the VM. You showed that the migration is immediate but does not mention why. I suppose that you've placed the disk(s) of it into pve-storage.

    • @distrodomain
      @distrodomain  ปีที่แล้ว +1

      When you ceate the vm you want to use the pve-storage pool that we create in the video this is the ceph pool containing the 3 osd devices that correspond to the the 3 drives, 1 in each node. Thank You for watching :)

  • @mattiavadala7870
    @mattiavadala7870 ปีที่แล้ว +1

    Nice VIdeo! When you create a new VM which storage you have to select? the dault one or the Ceph ppol? and when you mark a VM for HA, the node will move it to the Ceph pool? thanks for your work!!

    • @distrodomain
      @distrodomain  ปีที่แล้ว +2

      Ideally if you create a ceph configuration like this, you should put all your vms in the ceph pool, if you have a vm where the vm drive is not in a ceph pool you need to move the vm drive to the ceph pool before the vm can be HA. Thanks for watching!

  • @mohammedfaisal6946
    @mohammedfaisal6946 4 หลายเดือนก่อน +1

    i have a 3 node cluster with proxmox 8.2.2 they have public IP assigned three servers in three different locations
    my problem is I have installed ceph but the public IP on each of the server is different and has different network how do I resolve this and create a the ceph cluster ..and if not how do I create a HA storage accrose these three nodes ..i have tried NFS but it takes lot of time for migration

    • @distrodomain
      @distrodomain  4 หลายเดือนก่อน

      For ceph you need high bandwith like 10g and low latancy for the replication depending on how taxed the cluster will be, you a network separate from the main network, you would need to have a second public ip for each node just dedicated for ceph, you can also try zfs replication, for multiple locations I would recommend having a cluster of 2 nodes in each locations and then, have a migration plan for each location if you need to move vms around. thank you for watching!

  • @primenetwork27
    @primenetwork27 9 หลายเดือนก่อน +1

    i have a 3 dell server and config raid in my DELL PERC , i have a problem creating a pool in ceph

    • @distrodomain
      @distrodomain  8 หลายเดือนก่อน

      What are the issues you are encountering?

  • @Elrevisor2k
    @Elrevisor2k ปีที่แล้ว +1

    are you installing the cluster on 3 PCs or 2 PC and raspberry? could be different devices? so in case 1 node goes down the low power node is up and running? Ceph looks great to use the same storage on the 3 clusters nice!!

    • @distrodomain
      @distrodomain  ปีที่แล้ว +1

      Is recommended to have the same machines with same storage for the nodes, but it can work with different ones, as long as the rest of the nodes can take the load to spin the vms that went down back up again.

  • @RiyasudeenMohamedthaha
    @RiyasudeenMohamedthaha 4 หลายเดือนก่อน +1

    before cluster config. i have created VM in both node is it possible to create cluster

    • @distrodomain
      @distrodomain  4 หลายเดือนก่อน

      you shoule be able to create cluster with vm already running, let me know if you get errors, thank you for watching!

    • @RiyasudeenMohamedthaha
      @RiyasudeenMohamedthaha 4 หลายเดือนก่อน

      @@distrodomain Thanks for reply. detected the following error(s):
      * this host already contains virtual guests
      TASK ERROR: Check if node may join a cluster failed! - in this case i have created VM both node even i stopped Vm clster geeting error

    • @RiyasudeenMohamedthaha
      @RiyasudeenMohamedthaha 4 หลายเดือนก่อน

      @@distrodomain Yes i got different error after configure cluster. Both node have same VM_ID name Like (VM100) in this case cannot allow to run the machine Same ID either any one getting work rest of the not support to run this machine - But my understanding i need to change the VMID both machine and disk

  • @dundydunker
    @dundydunker ปีที่แล้ว

    How would this compare to using the another machine for the shared storage running TruNAS or something. Would it be more optimized or less? For example you have 3 HA prox cluster then have another machine for the shared storage (or several if that’s HA as well).

    • @distrodomain
      @distrodomain  ปีที่แล้ว +1

      For an hyperconverge setup you cound create 3 iscsi targets with a dedicated vdev for each node, this way linux will see the them as a block device and you can run the ceph setup on those block devices, I have not tried this setup but I'm curious about the outcome, if you only have a 1g network you might encounter performance issues, you'll be limited to 125Mb/s read/write on the vm disks. Thank You for watching :)

    • @dundydunker
      @dundydunker ปีที่แล้ว

      @@distrodomain Interesting, I will have to try that out. Thanks for the reply

    • @DigisDen
      @DigisDen 11 หลายเดือนก่อน +1

      I have a TrueNAS as well as ceph, I have dedicated 10Gb links for it too. Nevertheless, the problem with the TrueNAS is that it then becomes another single point of failure, I'd rather use a slower but more resilient ceph than the TrueNAS

  • @kristof9497
    @kristof9497 11 หลายเดือนก่อน +1

    Thanks, great video.

    • @distrodomain
      @distrodomain  11 หลายเดือนก่อน

      Glad you liked it! Thank you for watching :)

  • @mentalplayground
    @mentalplayground หลายเดือนก่อน

    This reminds me Gluster nightmare.

  • @munizda
    @munizda 5 หลายเดือนก่อน +1

    This is a gem.

    • @distrodomain
      @distrodomain  5 หลายเดือนก่อน

      I'm glad you enjoyed it! thank you for watching

  • @BoonyaritSuphannork
    @BoonyaritSuphannork 3 หลายเดือนก่อน +1

    Hi, How do i change ip on node and cluster have work 😂

    • @distrodomain
      @distrodomain  3 หลายเดือนก่อน

      check under these files, /etc/network/interfaces, /etc/hosts, /etc/pve/corosync.conf, thank you for watching!

  • @crc1219
    @crc1219 ปีที่แล้ว

    How did you replicate the 500GB SSD to each of the three nodes? Thanks for the video.

    • @distrodomain
      @distrodomain  ปีที่แล้ว +1

      In the part of the video were I am setting up de OSDs and the ceph pool that's where we are setting up the replication the 3 500gb drives are part of the pool and the ceph monitors keeps track of the replications. Thank you for watching!

    • @crc1219
      @crc1219 ปีที่แล้ว

      @@distrodomain Got it. You have a 500GB SSD drive on each node. Thanks.

    • @distrodomain
      @distrodomain  ปีที่แล้ว +1

      @@crc1219 Yes exactly, each node has the same storage 1 256g nvme ssd for the OS, and 1 500g ssd for ceph replication used for storing the VMs.

  • @syncbyte1782
    @syncbyte1782 4 หลายเดือนก่อน +1

    it dont work for me,
    cant install Cluster

    • @distrodomain
      @distrodomain  4 หลายเดือนก่อน

      At what step are you getting stuck?

  • @BMW750Ldx
    @BMW750Ldx ปีที่แล้ว

    great video....thanks 🤗👍

  • @rhb.digital
    @rhb.digital 2 หลายเดือนก่อน

    nice video... thnx man

  • @willbaker641
    @willbaker641 ปีที่แล้ว

    Can this be done with just two nodes?

    • @distrodomain
      @distrodomain  ปีที่แล้ว

      Yes this can be done with 2 nodes, and then later add nodes and grow your ceph pool. Thank you for watching :)

  • @rasyidrosada3067
    @rasyidrosada3067 ปีที่แล้ว

    what happens if ceph manager goes down?

    • @distrodomain
      @distrodomain  ปีที่แล้ว

      The manager provides the dashboard, it collects and distributes statistic data.
      They are needed for a good running ceph system, but they are not essential. So if all your managers would go down. Then there would be no immediate problem. You just start another manager on a node that is still up.

    • @rasyidrosada3067
      @rasyidrosada3067 ปีที่แล้ว

      thanks for explain

  • @handfootshrink
    @handfootshrink 5 หลายเดือนก่อน +1

    genius!

    • @distrodomain
      @distrodomain  5 หลายเดือนก่อน

      Thank You! :)

  • @RahulWaikar
    @RahulWaikar 10 หลายเดือนก่อน +1

    it working👍thanks

    • @distrodomain
      @distrodomain  10 หลายเดือนก่อน

      Nice, thank you for watching!

  • @karloa7194
    @karloa7194 ปีที่แล้ว

    I do not have ceph enabled, but I'm using zfs replication with HA enabled. For VM migration, no pings were lost. For the down nose, I get ~3 minutes then I got the ping started responding again.
    Each of my node is an Intel NUC8. How much overhead does the ceph adds to the PVE vs ZFS replication?

    • @distrodomain
      @distrodomain  ปีที่แล้ว +2

      If it's for homelab there is not much diference tho ceph is made more for enterprise depolyment where the biggest difference is that ceph has data redundancy on block or object level where ZFS does redundancy with whole disks. Need more space on ceph; Just add more disks, it will rebalance itself. Need to retire old disks? Just pull them out and the cluster will rebalance itself. Same for adding and removing nodes.
      Also is recommended with ceph to have high throughput network for rebuilds and rebalancing. I run it on a 1g network and I have had no issues but many users recommend a dedicated 10g switch, and 10g nics. Thank You for watching :)

  • @centurion7722
    @centurion7722 10 หลายเดือนก่อน +1

    plain and simple 👍

    • @distrodomain
      @distrodomain  10 หลายเดือนก่อน +1

      I'm glad it was helpful! :)