Hunyuan Video Day 1 Experiment With H100 GPU Server

แชร์
ฝัง
  • เผยแพร่เมื่อ 22 ม.ค. 2025

ความคิดเห็น • 50

  • @AIBusinessIdeasBenji
    @AIBusinessIdeasBenji  หลายเดือนก่อน +4

    Resources:
    Hunyuan Video Diffusion Model
    Model Page aivideo.hunyuan.tencent.com/
    Github : github.com/Tencent/HunyuanVideo
    Model Weights: huggingface.co/tencent/HunyuanVideo
    Hunyuan Video Installation In GPU Server : www.patreon.com/posts/hunyuan-video-in-117351026

    • @DonutMafter
      @DonutMafter หลายเดือนก่อน

      Nice video :) What cost per generation would you estimate?

  • @Rynwlms
    @Rynwlms 5 วันที่ผ่านมา

    Hi. I want to let you know that you do good work. And your videos are extremely helpful and valuable. Thank you. And keep it up!

  • @FusionDeveloper
    @FusionDeveloper หลายเดือนก่อน +10

    LTX video with the STSG enhancing and proper prompts is AMAZING and FAST and Local!
    Even runs on 11 and 12 gb VRAM (barely). There is a quant model that cuts down about 1gb VRAM.
    I run it on a 1080 to 11gb!
    Set steps to 20, 512x512 at 25 to 97 length for super fast outputs.
    There is image to video also if you use a custom node with guidance.
    Edit:
    LTX is the video model itself.
    STSG = SpatioTemporal Skip Guidance is a set of nodes that increases video quality by a lot.
    "Add LTX Latent Guide" is another node, which makes it so that Image to Video works.

    • @michaelbonomo6460
      @michaelbonomo6460 หลายเดือนก่อน

      What is STSG?

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน +1

      @@FusionDeveloper yes, for Local PC AI video, I suggest using LTX also. A good option. And a lot of add-on features to use

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน +1

      @@michaelbonomo6460
      Let talk about this one later 😉 using LTX and Cog are able to see the comparison with CFG junhahyung.github.io/STGuidance/

    • @FusionDeveloper
      @FusionDeveloper หลายเดือนก่อน

      ​@@michaelbonomo6460 LTX is the video model itself.
      STSG = SpatioTemporal Skip Guidance is a set of nodes that increases video quality by a lot.
      "Add LTX Latent Guide" is another node, which makes it so that Image to Video works.

    • @michaelbonomo6460
      @michaelbonomo6460 หลายเดือนก่อน

      @@FusionDeveloper noted. I've used cog and LTX a bit but I didn't know people were using a different node.

  • @PolyPhaze
    @PolyPhaze หลายเดือนก่อน +1

    Amazing content, Benji! Thanks for sharing it!

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน

      Glad you liked it! Will keep updating with this AI model , this one truly have potential VS

  • @xXWillyxWonkaXx
    @xXWillyxWonkaXx 18 วันที่ผ่านมา

    Amazing

  • @GarryMobi
    @GarryMobi หลายเดือนก่อน +3

    "An AI you can run at home" - Great! - "All you need is a GPU with 80Gig of memory!"

  • @AnonymerTraum
    @AnonymerTraum หลายเดือนก่อน +1

    du you know, is it possible to generate a 8min. 1080p video on a h200 running hunyuan video with the tiling strategie. or can it still run out of memory?

  • @Jet-Wingsio
    @Jet-Wingsio หลายเดือนก่อน

    wow this is so cool i feel like this would be so expensive

  • @garychap8384
    @garychap8384 หลายเดือนก่อน

    Can linked cards pool their vram and allow for a larger model? Or do you just get faster processing and the models have to fit on a single cards vram?
    I guess that one or two seconds locally, with a given seed, lets you know what you're generating... which, if you're budget constrained, means you're not moving a job to a paid farm until your confidence is high. But are there any unavoidable parameter changes between local and farm jobs? If so, that'd really suck because it would means your entire flow from conception to final needs to be farmed out : /

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน

      Link grpahic card at home wouldn't even get a lot faster. It also consider the GPU clip processor, clips that it use, etc...
      For AI Video, No one would only use Text2video method and generate video for production. So for the approach to try to gather a list of seed number, it haven't work effectively.
      By only running txt2vid, it is just a gambling play.

    • @garychap8384
      @garychap8384 หลายเดือนก่อน

      @@AIBusinessIdeasBenji Ah, that makes sense. Shame about linked cards though : /
      But one day... retired server farm H200's will be on eBay, selling in their 1000's as farms move towards higher densities and lower TDPs ; )
      ... on that day, we'll get ourselves a bargain ; )))))

    • @JohnEliot1978
      @JohnEliot1978 หลายเดือนก่อน

      Multi-gpu inference is on their open source implementation plan, but not ticked done yet

    • @ToothlessMusic
      @ToothlessMusic หลายเดือนก่อน

      You CAN run multigpu on this, but its not "distributed", I modified the comfyui nodes so I can set device on the sampler, decoder, and text encoder, allowing me to load each one to a different gpu, just to prevent the constant loading and unloading of the models, not everyone would need to do this, but my 2 4090's share a single 1x pcie slot via a riser switch with a 4080 on the board so it was taking me longer to load/unload then it did to generate. I also noticed on the decoder, auto_tile_size is default to on. turn that off and set sizes to 128, drops from needing 15gb to decode to about 4gb

  • @FluentlyNow
    @FluentlyNow หลายเดือนก่อน

    Thanks man! What about the cost per minute generated? Did you calculate it?

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน

      In server about $0.5 for a 121 frames 50 steps sampling.

  • @jcinewilliams8819
    @jcinewilliams8819 หลายเดือนก่อน

    Thank you for making this video I am really interested in this? I do have a beefy setup so I will have no problems installing it. However, I do have AP. C and you said that you're using AP. C, then, in order to get it to work, I would have to use sage.
    If it isn't too big of a ask, could you make an instructional video? Or how to go about install it? This AI video generator and sage on p. C. Thanks, that would be greatly appreciated. BTW this was a great review.

  • @geoffphillips5293
    @geoffphillips5293 หลายเดือนก่อน

    A week is a long time in AI, so now we have Kijai's wonderful comfyui adaption with temporal tiling that works on 24GB cards, or less. Works fine on Windows, but can't get it to work on Runpod, because of torch differences.

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน

      How many second you can generate with Comfyui wrapper node?

    • @geoffphillips5293
      @geoffphillips5293 หลายเดือนก่อน

      @@AIBusinessIdeasBenji It's stable at about 89 frames. After that if you try to go higher, it sometimes seems to lock up during the tiling part. I did find a runpod method too - look for hunyuan lab comfyu, or something like that. But it takes about 15 minutes to start up, and you have to copy to workspace to save your results using a console. Pretty good though! And you can reduce the size to pretty much anything to get faster results, say 512x512. It's also pretty good video to video, adding more life into say a Cogvideo creation.

    • @vpakarinen
      @vpakarinen 3 วันที่ผ่านมา

      ​512x512 is very low resolution, it needs to be at least 960x544 or even 1280x720​@@geoffphillips5293

  • @CharlenoPires
    @CharlenoPires หลายเดือนก่อน +1

    What is this hosting server?

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน

      @@CharlenoPires i did it in Runpod for code implementation, and also try with Replicate.com for API request

  • @JohnEliot1978
    @JohnEliot1978 หลายเดือนก่อน

    anyone tried this on runpod? have a deployment template?

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน

      They have a docker image. Or use Python code, just copy and paste.

  • @michaelbonomo6460
    @michaelbonomo6460 หลายเดือนก่อน

    Tested the lower quality version. Definitely not as good as what we're seeing from tencent. Does the full version that basically requires an H100 not have a standalone UI to use outside of ComfyUI? I'd rather not sign up for yet another website.

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน

      Theres no AI model have a standalone UI program for it man...

    • @michaelbonomo6460
      @michaelbonomo6460 หลายเดือนก่อน

      @AIBusinessIdeasBenji hate that. Was hoping I could run the full version like AI Toolkit or FluxGym in a standalone program. Lame. Hope they crank through their roadmap fast so we can see what it can do on our own servers.

    • @TheEconoVision
      @TheEconoVision หลายเดือนก่อน

      ​@@michaelbonomo6460AI toolkit, Fluxgym is not even made by Black Forest Labs anyway. Most tools and web ui made by other people.

  • @MilesBellas
    @MilesBellas หลายเดือนก่อน

    3D VAE : wow

  • @mickelodiansurname9578
    @mickelodiansurname9578 หลายเดือนก่อน

    Its not up to production level yet, then again its not a million miles away... for little insets it might be good

  • @FusionDeveloper
    @FusionDeveloper หลายเดือนก่อน

    Anyone getting into AI and planning to save up to buy a 8gb vram card, are making a huge mistake.
    16gb VRAM should be your minimum goal.
    Model size will continue to grow.
    Get the most vram possible.
    VRAM is more important than cuda cores (within reason).
    The 5090 should be coming out January 2025, approximately.

    • @AIBusinessIdeasBenji
      @AIBusinessIdeasBenji  หลายเดือนก่อน +1

      I will say 24GB VRam is the minimum for nowadays. Many diffusion models, lora training, etc... basically useup to 20GB VRam now.

  • @emimix
    @emimix หลายเดือนก่อน +1

    This whole AI never sleeping thing is getting old...

    • @GustavoTommaso-xk8vk
      @GustavoTommaso-xk8vk หลายเดือนก่อน

      “AI is addicted to coke.. this week we have some great news…”

    • @minimal2224
      @minimal2224 หลายเดือนก่อน

      Haterrr lol and he’s freaking right though

  • @aishatabubakar5872
    @aishatabubakar5872 หลายเดือนก่อน

    can you speak in layman terms this feels like a whole different language, theere should be a process non coders or techies can use

  • @Sujal-ow7cj
    @Sujal-ow7cj หลายเดือนก่อน

    Hat

  • @praiseandworship_com
    @praiseandworship_com วันที่ผ่านมา

    HELP!!!! $$$- Anyone out there want to help me with a 6 minute music video? Paid gig. good artist!! good cause