Stable Cascade in ComfyUI - Fully Supported!

แชร์
ฝัง
  • เผยแพร่เมื่อ 22 ก.ค. 2024
  • Update: img2img support added, check here:
    • Stable Cascade img2img...
    Using ComfyUI, we can now take a look at Stability's latest Image Generation model, Stable Cascade. A three stage model, it can produce high resolution images with good alignment very quickly compared with competing baselines.
    Article: civitai.com/articles/4161
    ComfyUI Workflow: civitai.com/models/310409
    You can download all the models inside ComfyUI Manager, be sure to update all your custom nodes in addition to ComfyUI. Consider disabling unused custom nodes if you have trouble.
    discord: / discord
    www.fivebelowfive.uk
    - Workflow Packs:
    Hyper SUPIR civitai.com/models/444002
    Merge Models civitai.com/models/432863
    cosXL Convertor civitai.com/models/420384
    Looped Motion civitai.com/models/410919
    Trio Triple Latents civitai.com/models/381021
    Ananke Hi-Red civitai.com/models/352117/ana...
    - SDXL Lora's
    civitai.com/models/384333/hel...
    civitai.com/models/401458/not...
    civitai.com/models/405640/pau...
    civitai.com/models/339881/ass...
    civitai.com/models/320332/ass...
    civitai.com/models/308724/zen...
    civitai.com/models/125217/spa...
    civitai.com/models/117624/asc...
    civitai.com/models/117546/wom...
    civitai.com/models/106864/sdx...
    - Introducing series (music/video)
    Noisee.ai • Introducing Noisee.ai ...
    Udio.com • Introducing Udio.com [...
    suno.com • Introducing Suno V3 Music
    haiper.ai • Introducing Video Gene...
    - Checkpoint Merging
    • Create The Best Model ...
    - cosXL / cosXL-edit conversion
    • Convert any SDXL model...
    • Unlock CosXL with any ...
    - 3D Generation
    • DJZ 3D Collection
    - New Diffusion Models (April '24)
    Stable Cascade:
    • Stable Cascade Comfy C...
    • Stable Cascade in Comf...
    SDXS-512:
    • SDXS - New Image Gener...
    cosXL & cosXL-edit:
    • CosXL & CosXL-Edit - N...
    - Stable Cascade series:
    • Stable Cascade Workflo...
    - Image Model Training
    datasets • Datasets in detail - M...
    colab • Updated Lora Training ...
    local • Updated Lora Training ...
    civitai • Stable Cascade LORA tr...
    civitai • SDXL Lora Training wit...
    - Music with Audacity
    • Make Music with Audaci...
    • Make Music with Audaci...
    - DJZ custom nodes (aspectsize node)
    • AspectSize (djz-nodes)...
    stable diffusion cascade
    stable diffusion lora training
    comfyui nodes explained
    comfyui video generation
    comfyui tutorial 2024
    best comfyui workflows
    comfyui image to image
    comfyui checkpoints
    civitai stable diffusion tutorial

ความคิดเห็น • 47

  • @glenyoung1809
    @glenyoung1809 5 หลายเดือนก่อน +10

    One thing to help users distinguish which format of model to use.
    Bf16(bfloat16) can only effectively be used by NVidia 30 and 40 series cards(Ampere and Ada) which have the native hardware support to read that format.
    The "full sized" models are in fp32 format which any modern GPU since 2010 can use but of course they take up more VRAM.
    Even though the bf16 models are "only" 16bits per parameter and the full sized fp32 are 32bits per parameter both are of almost the same quality in terms of the range of the numbers they can represent.
    The only difference is that the bf16 doesn't have as much accuracy(fewer decimal places) as fp32.
    Don't confuse bf16 with fp16 which is another number format mentioned with AI datasets and GPU capabilities and which has much less "range" in the values of numbers they represent.
    In so far as image quality you can think if bf16 as having almost the same dynamic range as fp32 while fp16 is nowhere near the same range as bf16.
    The "lite" vs full version should be self explanatory, lite has many less parameters than the full version therefore less encoded information in it's neural net(smaller capacity).
    Nvidia 30 and 40 series card users can mix and match between bf16/fp32 and lite/full variations.
    But if you don't have 30 or 40 series you will only be stuck with the lite/full versions of the fp32 models.
    FYI, the SC nodes as they're setup do NOT remain the models in memory after they are used, therefore your maximum VRAM usage is not accurate, the largest full model I've loaded takes up around 8-9GB VRAM so anyone with a 12GB card should be able to comfortably run the full fp32 models.

    • @MichauxJHyatt
      @MichauxJHyatt 5 หลายเดือนก่อน +1

      I appreciate your insight!

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน +1

      thanks for the detailed write up! I was using a 3060 12gb and upgraded to 4090 24gb, so i guess i never noticed, i'll be sure to mention this in future :)

    • @fabiotgarcia2
      @fabiotgarcia2 5 หลายเดือนก่อน

      So... for Mac M2 we need to use fp32 format, right? where can we find it for download?

    • @glenyoung1809
      @glenyoung1809 5 หลายเดือนก่อน

      @@fabiotgarcia2search for huggingface stable cascade, the first page should have the safetensors files want the ones marked stage a,b,c.

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน

      huggingface.co/stabilityai/stable-cascade @@fabiotgarcia2 should all be here as safetensors format

  • @electronicmusicartcollective
    @electronicmusicartcollective 5 หลายเดือนก่อน

    THanx men! Good explaining and a handy workflow

  • @magenta6
    @magenta6 5 หลายเดือนก่อน

    Excellent and concise!

  • @styrke9272
    @styrke9272 5 หลายเดือนก่อน

    cool and pretty concise!!

  • @Loutchianooo
    @Loutchianooo 5 หลายเดือนก่อน +1

    Thanks man, really helpful! A bit confusing all them folders, SDXL was checkpoints, now it's unet...

  • @gimperita3035
    @gimperita3035 5 หลายเดือนก่อน +1

    I'm running full stage C and B on a 4080. No errors so far generating 2K ish, even 2560x1440.

  • @juanchogarzonmiranda
    @juanchogarzonmiranda 5 หลายเดือนก่อน

    Thanks!!

  • @RalFingerLP
    @RalFingerLP 5 หลายเดือนก่อน

    Nice Video drift!

  • @equilibrium964
    @equilibrium964 5 หลายเดือนก่อน

    Good job, dude. Thank you very much! Do you know if it is possible to use stable cascade with SDUpscale and which model I should use in this case?

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน +1

      upscaling is done on the images, so you can go ahead and use any upscaler or upscaling nodes. It does produce very high resolution images, so it will be interesting to see how large they can go.

  • @ebzmuffshuh7513
    @ebzmuffshuh7513 5 หลายเดือนก่อน

    keep it up

  • @AI_Creatives_Toolbox
    @AI_Creatives_Toolbox 5 หลายเดือนก่อน +1

    Thanks for the excellent video! An unrelated question - How do you get the lower bar with the renders? Thanks!

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน

      in the latest comfy you can click a button iin the bottom left corner, adjust less images with bigger size will get you what i have set in mine :)

    • @LuckyWabbitDesign
      @LuckyWabbitDesign 4 หลายเดือนก่อน

      @@FiveBelowFiveUK not seeing any 'button in bottom left corner'. Pretty certain I've got the latest cumfyUI. Could you describe further, or post a screenshot? thanks

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  4 หลายเดือนก่อน +1

      i can post a screenshot on the channel feed, so look there ;) @@LuckyWabbitDesign

  • @kofteburger
    @kofteburger 5 หลายเดือนก่อน

    Tried this With Radeon 6700 10 GB on Ubuntu using the lite models. It worked once I replaced the VAE decoder with the tiled one. However the second KSampler (stage B?) is painfuly slow on the default resolution. Over 20s/Itteration slow. 1024x1024 is quite good.

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน +1

      interesting. See if the newer method released later yesterday works. I have people on

  • @magimyster
    @magimyster 5 หลายเดือนก่อน +1

    Is it possible to make it work with just the cpu?

    • @sickvr7680
      @sickvr7680 5 หลายเดือนก่อน +2

      yes

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  4 หลายเดือนก่อน +1

      yes, but you will need to install the python marked with CPU for your system, not the cuda version

  • @rsunghun
    @rsunghun 5 หลายเดือนก่อน

    Does it work with controlnet?

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน

      there is a canny model, however it's not yet implemented, i'll cover this as soon as it arrives :D

  • @dasistdiewahrheit9585
    @dasistdiewahrheit9585 5 หลายเดือนก่อน

    Thanks for the info. One point: What you wrote on that website is pretty hard to read. Please put some newlines or something.

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน

      if you meant Civit, i wrote it up a bit clearer. hope that helps, can add more if needed

    • @dasistdiewahrheit9585
      @dasistdiewahrheit9585 5 หลายเดือนก่อน

      @@FiveBelowFiveUK That was exactly my point. Thank you, very helpful information.

  • @djivanoff13
    @djivanoff13 5 หลายเดือนก่อน +3

    next video img2img plz (Stable Cascade)

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน +1

      effnetencoder not yet supported in comfy, watching the commits for an update, this is needed to encode loaded images :D we will have it up as soon as it lands!

  • @appolonius4108
    @appolonius4108 4 หลายเดือนก่อน

    the workflow is not the one you show in the video

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  4 หลายเดือนก่อน

      If you look at the Argus Page, there are more than 10 versions already. Each was created for a purpose.
      The early Workflows do not use the Checkpoints released on day three following Cascade launch. We were using them on day one with these.
      all Argus Versions up to V18 did not use the newer and current checkpoint method. You can learn more about that in the next video ;) "stable cascade comfy checkpoint update"
      I have a version of Argus, that is all in one, "Argus Cascade Studio". Once this is complete, I will release it.

  • @LouisGedo
    @LouisGedo 5 หลายเดือนก่อน

    👋

  • @jonmichaelgalindo
    @jonmichaelgalindo 5 หลายเดือนก่อน +1

    You don't have to match B and C! Use C bf16 and B lite bf16 for best results.
    "Fully supported" do img2img or inpainting. O wait, you can't. Works in SAI repo with python script only.

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน

      uhm. hold up there a minute haha.
      Img2img is supported, you are watching old videos ;)
      We got Vision and img2img in SDXL since day one.
      This video was the day one comfyui support launch.
      The big giveaway is this was released before Stability even let checkpoints out.
      Since then I released workflows that allow ANY diffusion model with all CNET etc to bridge into cascade. Should check for the more recent videos ;) before you comment.
      AI space moves fast, it's not waiting around.
      also, Ampere cards only for BF16, so 30xx & 40xx, so i try not to make assumptions when i give out advice

    • @jonmichaelgalindo
      @jonmichaelgalindo 5 หลายเดือนก่อน

      @@FiveBelowFiveUK This video is for Stable Cascade. It's a completely different model architecture from SDXL. No img2img support for Stable Cascade in Comfy as of morning Feb 24 2024.
      (In SDXL, to do img2img you only need to run the RGB pixels through a VAE to get a latent. In Stable Cascade, you first run the RGB through a VAE to get one latent, then you compress that to a second latent using a diffusion model (called "stage B"). It's this compressed latent that Stable Cascade's main pipeline diffuses ("stage C").)

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  4 หลายเดือนก่อน

      hi there, You seem to be running on maybe the day before info.
      The **effnet_encoder** is a VAE which is built into the stage C encoder. You run your image into the VAE encoder to perform img2img. There is no difference in principle between this and the VAE encoder used with SDXL. In early workflows (last week) we were doing this before the Checkpoints were released (see Argus-v18 img2img workflow)
      Please check recent videos, as if you watch a more than a day late, it's likely outdated already. @@jonmichaelgalindo
      We have had img2img support in comfyui since Feb19th. I released workflows to help people on that day that comfyui added the code to support the feature.
      As a matter of fact i have demonstrated using SD1.5, SD2.1, SDXL all with txt2img, img2img and Vision Stack in the case of Cacade, with and without Lora Loading on both sides.
      Specific nodes were introduced to aid with Cascade img2img here: (19th feb) github.com/comfyanonymous/ComfyUI/commit/a31152496990913211c6deb3267144bd3095c1ee
      Like i said, I'm here to bridge the gap between the new tools and artists.
      That means putting things in terms everyone will understand.

    • @jonmichaelgalindo
      @jonmichaelgalindo 4 หลายเดือนก่อน

      @@FiveBelowFiveUK Thanks so much, I'll try again to find the info. I honestly haven't found a single video or workflow anywhere enabling img2img. I tried using the VAE as an encoder on the 17th. I'm amazed posting that github link didn't get your comment deleted! I've found a workflow that adds images as conditioning?

  • @ushakovkirill
    @ushakovkirill 5 หลายเดือนก่อน

    nice avatar

  • @AgustinCaniglia1992
    @AgustinCaniglia1992 5 หลายเดือนก่อน

    Too many models. My pc frozes every time it loads one. Not practical

    • @FiveBelowFiveUK
      @FiveBelowFiveUK  5 หลายเดือนก่อน +1

      well lucky for you, they released two proper checkpoints to make it all easier.
      new video and workflows came out yesterday