LTX Video In ComfyUI - The Fastest AI Video Model Run Locally - Tutorial Guide

แชร์
ฝัง
  • เผยแพร่เมื่อ 4 ก.พ. 2025
  • LTX Video In ComfyUI - The Fastest AI Video Model Run Locally - Tutorial Guide
    Hey there amazing people! 🎥 Exciting news in the AI video generation world! Today, I'm diving deep into the groundbreaking LTX video model, now fully integrated with ComfyUI. This game-changing model is a real powerhouse that's surprisingly lightweight and runs smoothly on consumer PCs, capable of generating high-quality 24 FPS videos in mere seconds on an NVIDIA 4090.
    Comfy Blog blog.comfy.org...
    LTX Video Model HF huggingface.co...
    Freebie Workflows: www.patreon.co...
    For My Patreon Supporters: www.patreon.co...
    For Low/No GPU User, You Can Try Run On The Cloud
    ----------------------------------------
    ComfyUI - LTX Video
    home.mimicpc.c...
    ComfyUI - CogVideoX 5B 1.5 Both t2v & i2v
    home.mimicpc.c...
    FLUX.1 Tools - Redux: Adaptive Image and Prompt Mixing:
    home.mimicpc.c...
    FLUX.1 Tools - Fill: Smart Inpainting
    home.mimicpc.c...
    FLUX.1 Tools - Fill: Smart Outpainting
    home.mimicpc.c...
    In this comprehensive breakdown, I'll show you exactly how to set up and use this incredible tool, from installing the model to creating stunning video content. We'll explore both text-to-video and image-to-video capabilities, with real demonstrations of landscape scenes, character animations, and cinematic sequences. I'll share some pro tips on optimizing your prompts and settings to achieve the best possible results, including how to extend your videos beyond the standard 5-second limit.
    What sets this model apart is its exceptional performance-to-resource ratio and the remarkable quality of its outputs. Whether you're a content creator, filmmaker, or AI enthusiast, you'll be amazed by how this 2 billion parameter model handles complex scenes, character movements, and environmental effects - all while maintaining impressive coherence and stability. Join me as we explore this exciting new frontier in AI video generation!
    If You Like tutorial like this, You Can Support Our Work In Patreon:
    / aifuturetech
    Discord : / discord

ความคิดเห็น • 95

  • @TheFutureThinker
    @TheFutureThinker  2 หลายเดือนก่อน +5

    Workflows Run On The Cloud (For Low/No GPU Users): home.mimicpc.com/app-image-share?key=d1ea3605ebe54306bc6876b6af49a85a&fpr=benji
    Freebie Workflows: www.patreon.com/posts/116627608/?
    For My Patreon Supporters: www.patreon.com/posts/116627998/?

    • @ssj3mohan
      @ssj3mohan 2 หลายเดือนก่อน +1

      amazing stuff for sure

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน +1

      @ssj3mohan hope this helps

  • @damarcta
    @damarcta 2 หลายเดือนก่อน +3

    I have to say that I'm grateful for how quickly the videos are uploaded. I use ComfyUI extensively, PDXL, Flux, CogVideoX, etc., and I never keep track of when new updates are released. So, thank you for the quick updates!

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน +4

      this is my hobby :) finding new AI and crafting something.

  • @content1
    @content1 28 วันที่ผ่านมา

    Thank you for your video. Originally I had the problem of missing nodes that were not available in the missing nodes then I had to upgrade comfy and now it works.

  • @crazyleafdesignweb
    @crazyleafdesignweb 2 หลายเดือนก่อน +6

    Looks promising for local AI Video. It feels like last year when AnimateDiff began.

  • @cgonv
    @cgonv หลายเดือนก่อน

    I managed to make it work very well on an 8GB Nvidia, as fast as on your 24GB one! Fantastic, thank you very much for your help! That was revolutionary!

  • @jac001
    @jac001 2 หลายเดือนก่อน +7

    1min 13 sec on a 3060 12gb, default text to video scene, w/ Ponokio. Happy to have arrived in this reality finally, and it only gets better from here!

  • @jorgemiranda2613
    @jorgemiranda2613 2 หลายเดือนก่อน

    Cool content!! Thanks for keeping us updated ! Subscribed !

  • @bause6182
    @bause6182 2 หลายเดือนก่อน +3

    I can't wait to see controlnets and finetunes or consistent carachters on this model

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน +1

      There are v2v module for this AI, where it use a reference video to control the motion.

    • @VFXShawn
      @VFXShawn 2 หลายเดือนก่อน +3

      @@TheFutureThinker Where can we find that v2v module?

  • @joshuadelorimier1619
    @joshuadelorimier1619 2 หลายเดือนก่อน +1

    For how fast it is it's incredible. Can do portrait and landscape but I also got some decent three character shots no dramatic action yet. One minute renders on a 4070 image to video.

    • @jac001
      @jac001 2 หลายเดือนก่อน

      1m13s on 3060 12gb.

  • @kdzvocalcovers3516
    @kdzvocalcovers3516 15 วันที่ผ่านมา

    the longer the video the more deformation and morphing will occur...5 seconds max..to render more accurate motion and transitions....seems to be true ...video extend=mutations in the last 5 seconds i think..open source is still great...glitches and all..thanks for the great tutorial

  • @Blenderlands
    @Blenderlands 2 หลายเดือนก่อน

    Is it good for stylish animation videos?

  • @mr.entezaee
    @mr.entezaee หลายเดือนก่อน

    size mismatch for decoder.up_blocks.6.res_blocks.2.conv2.conv.bias: copying a param with shape torch.Size([128]) from checkpoint, the shape in current model is torch.Size([256]).

  • @personmuc943
    @personmuc943 หลายเดือนก่อน +2

    Why am I getting an error message after it finishes generating saying "LTXVModel.forward() missing 1 required positional argument: 'attention_mask'"? the video never shows up

  • @CoreyJohnson193
    @CoreyJohnson193 2 หลายเดือนก่อน

    If only we could use ControlNet with this... Perfect combo. Still waiting for X Portrait 2; combined would be illegal!

  • @WildCrashBNG
    @WildCrashBNG หลายเดือนก่อน

    Sir, which python version is required for ComfyUI? Also, when I install LTX video from the manager, it says (IMPORT FAILED). What could be the reason for this?

  • @PyruxNetworks
    @PyruxNetworks 2 หลายเดือนก่อน

    Have you tried the SkipLayerGuidanceDit node in comfyui? outputs seems better with it

  • @jcinewilliams8819
    @jcinewilliams8819 26 วันที่ผ่านมา

    How do we save our generated videos?

  • @ForeverNot-wv4sz
    @ForeverNot-wv4sz 2 หลายเดือนก่อน

    I'm wondering when/if we're ever going to get something akin to animate diff 1.5 lcm in sdxl terms. I can see we have motion models now for sdxl but we can't take use any other sdxl model with it, unlike 1.5 where it loads the motion model as a unet for the 1.5 model selected. We have the tools now for sdxl to make really fast images with lcm and dmd lora so we can use lcm on any non-lighting sdxl model. Unless I'm just uninformed and we already have something like this and I just missed it with all the new tech coming out so quickly to keep up with it all

  • @howtowebit8033
    @howtowebit8033 2 หลายเดือนก่อน

    Does it work in Colab?

  • @eveekiviblog7361
    @eveekiviblog7361 2 หลายเดือนก่อน

    Is it possible to make good video from inage? Or descent results are only with t2v?

  • @francsharma7276
    @francsharma7276 2 หลายเดือนก่อน +4

    on my 8gb 3070ti text to video work absolutely great

    • @navaneeth8260
      @navaneeth8260 หลายเดือนก่อน

      is LTX paid?

    • @francsharma7276
      @francsharma7276 หลายเดือนก่อน

      @navaneeth8260 no its open-source and now witn stg diffusion you can generate kling level videos through image to video, on 8gb vram

  • @BlackMatt2k
    @BlackMatt2k 2 หลายเดือนก่อน

    Can it do equirectangular projection?

  • @guangxing9010
    @guangxing9010 19 วันที่ผ่านมา

    How to convert to MP4

  • @CSBRHO
    @CSBRHO 2 หลายเดือนก่อน

    Thank you my friend! U know how to upscale the video after the process?

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน

      In Comfyui, you can use Upscale by Model.
      Or another way use AI upscale software like Topaz.

  • @aivideos322
    @aivideos322 2 หลายเดือนก่อน

    your hard drive must be huge, i have run out of space for all these new models.

    • @TheFutureThinker
      @TheFutureThinker  หลายเดือนก่อน

      I deleted some old one , AI files over 1 month no using I deleted it.

  • @The_Python_Turtle
    @The_Python_Turtle 2 หลายเดือนก่อน

    thank you. Do you know how to save the output .webp as a video file so you can import into a video editor? I tried saving the .webp but it is just a static image. Was thinking there is a comfy UI that might be able to do this

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน +1

      Change the output image connect to Video Combine, and use the option out as mp4 format. So you don't have to convert one file to another format.

  • @Radarhacke
    @Radarhacke 2 หลายเดือนก่อน +1

    Did not work, cant get the nodes: LTXVConditioning, LTXVScheduler, LTXVImgToVideo. Updated All Comfyui and Manager. Also tried manual Install, same Result: When loading the graph, the following node types were not found: LTXVConditioning, LTXVScheduler, LTXVImgToVideo. Also cant select LTX in Clip Loader, you see, nothing worked for me.

    • @geoffphillips5293
      @geoffphillips5293 2 หลายเดือนก่อน

      I needed to do a full comfyui upgrade using manager, I also needed to do "FIX NODE" right clicking the one that loads the model. And of course refresh the webpage and restart the server. Plus you need the git clone line and model as per the git hub page.

    • @content1
      @content1 29 วันที่ผ่านมา

      I have same problem

  • @AB-wf8ek
    @AB-wf8ek 2 หลายเดือนก่อน +1

    It seems like a lot of people using these tools don't realize, having immediate access to these models is way beyond most peoples' experience in the past.
    In the 3D rendering community, if you watched a video on the latest simulation and rendering research, you'd be lucky if you actually got to use it in 5 years when a commercial application got around to integrating it, and it certainly wouldn't be free.

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน +1

      That's right. AI model , a same model put in commercial and an open source version. Its not the same. Like Mochi, and Pyramid Flow.

  • @contrarian8870
    @contrarian8870 2 หลายเดือนก่อน +1

    I recently saw the CogX video model in Comfywith the "orbit_left" and "orbit_up" LORAs from DimensionX, to make simple 3D clips (no complex motion). Can LTX be used with these LORAs instead of CogX to speed things up?

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน

      Is like can a sea lion merge with African lion? They are both call lion 🦁

  • @cr_cryptic
    @cr_cryptic 2 หลายเดือนก่อน +1

    13:31, lucky fuxkr… Mine’s been taking like an hour & it hasn’t even made it to step 1/30 yet. 🤦‍♂️ Why? 😭

  • @hicks100
    @hicks100 2 หลายเดือนก่อน

    Excuse me:
    1. Can LTX Video generate animated videos? Can it be used with Lora and SDXL stylized prompt words?
    2. In image to videos, the prompt words are automatically analyzed and generated. How to manually modify the prompt words?
    3. I want the characters in the generated video to be closer to the original pictures. How should I set the parameters?

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน

      For your questions,
      - Anime video? Yes
      - You can use Text prompt in TextEncoder node
      - try higher step numbers, this model can run 100 -200 steps without slow down too much.

    • @hicks100
      @hicks100 2 หลายเดือนก่อน

      @@TheFutureThinkerCan Lora work on LTX video?

  • @francsharma7276
    @francsharma7276 2 หลายเดือนก่อน

    8gb VRM 3070ti not enough for image to video

  • @vasilybodnar168
    @vasilybodnar168 2 หลายเดือนก่อน

    I'm interesting only I2V and in this case LTX do nothing. It generates allmost static shots - no camera movement, no object movement - nothing. Weird.

  • @geoffphillips5293
    @geoffphillips5293 2 หลายเดือนก่อน

    It's almost insanely fast, in my ten minute test this morning, before heading out for work. Hoping the quality would improve with more steps. (edit later) Hmm. same problem as with others, face distorting, hands the same. Not so good for people walking as Cog, by a huge margin. It could do over 200 frames, using close to my 24 gigs max. If it could just be a bit more stable with people's features, it would be great. Text to video producing very fake looking people. Still, always fun to try out something new!

  • @Andro-Meta
    @Andro-Meta 2 หลายเดือนก่อน

    I've pushed it up to twelve seconds with decent results on an RTX 3080.

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน

      Nice! 👍

    • @ammarzammam2255
      @ammarzammam2255 2 หลายเดือนก่อน

      And how much time did it take to generate those 12 seconds on your GPU

    • @Andro-Meta
      @Andro-Meta 2 หลายเดือนก่อน

      @@ammarzammam2255rough estimate (I was at a friend's, showing him how to build workflows and testing LTX Video with him on his computer) is around 3-4 minutes at 20 steps. It started doing a transition into a whole new scene, so we added "transition, new scene" into the negative prompt and surprisingly that worked.
      I'll test it on my 4090 later this week, but it was surprisingly zippy, even in making a twelve second video on a 3080.

  • @FusionDeveloper
    @FusionDeveloper 2 หลายเดือนก่อน

    Update Manager. Git pull
    Update ComfyUI. Git pull
    Update All (for updating nodes). In GUI.

  • @ian2593
    @ian2593 2 หลายเดือนก่อน

    Tried adding a cigarette to the girl with blood on her face/top and it couldn't handle it. It would be interesting to nail down what its strengths and weaknesses are. A good step forwards though.

  • @YoungBloodyBlud
    @YoungBloodyBlud 2 หลายเดือนก่อน

    well now i know where this is gonna be used it starts with N and ends with W iykyk

  • @FusionDeveloper
    @FusionDeveloper 2 หลายเดือนก่อน

    Works on a 1080 ti 11gb vram.

    • @chawkibouziane
      @chawkibouziane หลายเดือนก่อน

      What time to generate the vidéo ?

  • @tinfoilhatmaninspace4944
    @tinfoilhatmaninspace4944 2 หลายเดือนก่อน

    Until it can do better than just 6 seconds I'm not bothering with the video aspect of ai, Images have perfected themselves with Flux so until their is sound and longer videos I'm not wasting my time with this.

  • @mareck6946
    @mareck6946 2 หลายเดือนก่อน

    You dont get better output with a lower end gpus. its the same output just takes longer. WHAT matters is vram tho - where nGreedia wants to rip you off. More vram less swapping or none at all - and more temporal framebuffers for temporal animation etc.

  • @insurancecasino5790
    @insurancecasino5790 2 หลายเดือนก่อน

    I think most are better off just waiting a year if they don't want to use a GPU rental service. These guys are making it way more complicated than it should be by now. We have image generators now that can generate consistent images very quickly with low VRAM. I know they are going to have to go back to basics to generate videos if they want to achieve high quality video for low VRAM machines. It's just a fact. There are limits to this. I'm sticking with paid generators for now. I will still enjoy watching your vids, but overall this stuff is a headache for most. IMO.

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน +3

      Yes, for production me too.
      For local model we just have to keep track of it, and see what's going on and if there's any potential that we can make things happen. But this model is really lightweight, and using good images it can render better quality than CogVideoX.

    • @insurancecasino5790
      @insurancecasino5790 2 หลายเดือนก่อน

      @@TheFutureThinker Well, that's good. I'm just waiting for the 1 click options like image generators have. I will try this when I have time. I just remember the early SD days. It was a nightmare for folks that were not developer savvy. Then we got pinokio that made it so simple to where I was easily downloading them and using them to this day. The early Faceswap broke my brain until I found Roop unleashed. That's like super simple. Just a few clicks. So, I'm hoping the vid models will get like that too.

    • @AB-wf8ek
      @AB-wf8ek 2 หลายเดือนก่อน +3

      Everyone gets burned out by trying to keep up with AI development. Even a lot of the developers can't keep up.
      I was telling a friend a while back, it's like trying to build a house on shifting sand. It doesn't make sense to overinvest in the current process, because it's all changing so fast.
      Whenever you reach your limit, it's ok to walk away and take a break.
      I actually completely skipped out on SDXL. I've been happy playing with AnimateDiff and SD1.5, and I'm just now getting into Flux.

    • @FusionDeveloper
      @FusionDeveloper 2 หลายเดือนก่อน

      ​@@AB-wf8ek try SDXL turbo Ultraspice with 7 to 11 steps, Euler, SGM Uniform, CFG 1 to 3. It's one of my favorite models to use. The quality is insane and it's fast.
      Flux is great, but slow.

    • @insurancecasino5790
      @insurancecasino5790 2 หลายเดือนก่อน +3

      @@AB-wf8ek My hard drive can't keep up either, so have two externals now if I want to try all the models.

  • @ja-no6fx
    @ja-no6fx 2 หลายเดือนก่อน

    every single time i try to follow one of this guys' guides, it doesnt work.

  • @SecretsandFactschannel
    @SecretsandFactschannel หลายเดือนก่อน

    bro does have a n.dify model, aw hell nah

  • @aminesoulaymani1126
    @aminesoulaymani1126 2 หลายเดือนก่อน

    maybe i'm stupid, I tested all day long the image2video functionality, I would be ashamed to release that and called it "video generation"

    • @kleber1983
      @kleber1983 2 หลายเดือนก่อน

      What do you mean? I´m doing it right now and finding the results amazing...!

  • @lenxie4501
    @lenxie4501 2 หลายเดือนก่อน

    why all my video look like shit...must be prompt problem

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน

      As detail as possible, and describe style as much as you can think of. And put prompt in a good structure.

  • @theradomguy5581
    @theradomguy5581 2 หลายเดือนก่อน

    I ran this on a 1060 6Gb, and it was slow. Wouldn't recommend unless you have probably at least a 3070 or better

  • @LagostinaCookie
    @LagostinaCookie หลายเดือนก่อน +1

    another Patreon beggar

  • @SavageKillaBees
    @SavageKillaBees 2 หลายเดือนก่อน

    This isnt ground breaking at all, it produces content to "here's my first A.I video". The content it produces looks terrible. Human animation still looks bad

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน +7

      @@SavageKillaBees yes , for arty without knowledge and understanding how AI model works and the difference of it, and concern only the generated video, it is not a good one.
      For people who know the movment of decentralize AI, they are going to see another point of view.

    • @kalakala4803
      @kalakala4803 2 หลายเดือนก่อน

      ​@@TheFutureThinkerthat's true

    • @SavageKillaBees
      @SavageKillaBees 2 หลายเดือนก่อน

      ​@TheFutureThinker do you think local models will ever come close to commercial models like KlingAI? None of the local models come close. If I want to create commercially viable image to video, I have to use websites. What do you think?

    • @TheFutureThinker
      @TheFutureThinker  2 หลายเดือนก่อน +2

      @@SavageKillaBees yes, that's right like Kling, Runway my fave. one. Because their AI model size are very large, and privately trained with dataset that it might not be from the open source dataset online. So their video preformance are way better.
      I have also suggest people if nowadays, using AI video for video ads, movies, etc. just use those commerical one. Don't waste time generate in local yet, it's not at that level for local run model.
      Like this one, only 2B parameters dataset, how can it compete with a model over 40B AI video model, with a lot more motions for referencing?
      But the local AI model, is keep improving. It will be like how a PC evolves, from a room size to a box size, or pocket size.

    • @SavageKillaBees
      @SavageKillaBees 2 หลายเดือนก่อน

      ​​@@TheFutureThinker I am just extremely impressed with what KlingAI and Runway can do. It just feels and looks superior to local models. I wonder if by next year, we get some local models we can run with top end local hardware. I have a 4080 right now but plan on getting the 5090 next year. You are right, 5 billion parameter models won't really cut it. We need much larger , more robust models for higher quality generation. I want to create commercial grade content but I can't do it locally. Only images.
      I've seen what LTX video can do but it will be sidelined for now because it's great to demonstrate the concepts, but if you're results oriented, it's just not there yet.

  • @upscalednostalgiaofficial
    @upscalednostalgiaofficial 2 หลายเดือนก่อน

    does not work on my 8g vram "torch.OutOfMemoryError: Allocation on device" :(