SDXL ComfyUI Stability Workflow - What I use internally at Stability for my AI Art

แชร์
ฝัง
  • เผยแพร่เมื่อ 27 ก.ค. 2023
  • Since we have released stable diffusion SDXL to the world, I might as well show you how to get the most from the models as this is the same workflow I use on a daily basis at stability.ai. In this video I show you some of the basics on how to get the model from the models to generate your best AI artwork from our models. You will need some of the custom nodes over at civit, but you can choose the package that works best for you, as they are all pretty similar.
    We will start with a basic workflow and then complicate it with a refinement pass, but then we will add in another special twist I am sure you will enjoy. #stablediffusion #sdxl #comfyui
    Grab some of the custom nodes from civit.ai: civitai.com/tag/comfyui
    Grab the SDXL model from here (OFFICIAL): (bonus LoRA also here)
    huggingface.co/stabilityai/st...
    The refiner is also available here (OFFICIAL):
    huggingface.co/stabilityai/st...
    Additional VAE (only needed if you plan to not use the built-in version)
    huggingface.co/stabilityai/sd...
  • ภาพยนตร์และแอนิเมชัน

ความคิดเห็น • 299

  • @TedWillingham
    @TedWillingham 11 หลายเดือนก่อน +34

    I would love if you could go over some of those settings in advanced detail - like "oh, I fiddle with more conditioning steps when I want to X", etc. There are so many superstitious people out there giving bunk advice that your level-headed breakdown would be super valuable!

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +16

      Great idea! I will have to ponder where to start! :-)

  • @iiiCorrosiveiii
    @iiiCorrosiveiii 11 หลายเดือนก่อน

    thank you so much, i've become really proficient with A1111 and moving to comfyui was a big switch, so your help with how the workflows work in comfyui has made it just as easy as using A111 for me.

  • @lakislambrianides7619
    @lakislambrianides7619 11 หลายเดือนก่อน +1

    This is a great video congrats. Very informative very thorough and you left no doubts. Can't wait for the next step!

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      More to come!

  • @dxnxz53
    @dxnxz53 หลายเดือนก่อน +1

    it blew my mind that you can load an entire workflow from the image! thanks for the great content.

  • @henryabela2672
    @henryabela2672 11 หลายเดือนก่อน

    Thank You so much! videos like these are a blessing and help people to get into it more professionally.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      You're so welcome!

  • @me.shackvfx5911
    @me.shackvfx5911 11 หลายเดือนก่อน +32

    I've grown to understand and enjoy comfy UI more that the one i was using before thanks to your videos.I really appreciate you and the effort you put into making these tutorials. One of these days you can show us how to train sdxl 1 or it's lora with our faces . Thanks :)

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +8

      Great to hear! Training will be coming soon! Cheers!

  • @rsunghun
    @rsunghun 11 หลายเดือนก่อน +2

    I was waiting for it. These are very difficult for ordinary people to figure out how to use it. Thank you for the video!

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      Glad it was helpful!

  • @PieterLaroy
    @PieterLaroy 11 หลายเดือนก่อน

    My first steps into ComfyUI, and it's the kind of thing I really like 🙂

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      Glad to hear it!

  • @benkilowhiskey
    @benkilowhiskey หลายเดือนก่อน

    Great watch, thank you!

  • @ziixu
    @ziixu 10 หลายเดือนก่อน

    Great tutorial!
    This is my first time using ComfyUI and this video helped me a lot, tyvm!

    • @sedetweiler
      @sedetweiler  10 หลายเดือนก่อน

      You are most welcome!

  • @lionhearto6238
    @lionhearto6238 11 หลายเดือนก่อน

    all the info i was looking for. great video. thank you

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      Glad to hear it!

  • @wschieber
    @wschieber 11 หลายเดือนก่อน

    Thanks, Scott. I was really looking for something like this to get started with SDXL in ComfyUI.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      Glad it was helpful!

  •  6 หลายเดือนก่อน

    Thanks Scott, as a beginner your videos are great. Very well explained and "easy to learn than a bunsh of others.

    • @sedetweiler
      @sedetweiler  6 หลายเดือนก่อน

      Glad to help

  • @imperfectmammal2566
    @imperfectmammal2566 11 หลายเดือนก่อน +1

    Thank you so much! Even though I couldn’t understand much, it helped me get started with comfy.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      You’re welcome 😊 Just keep working with it and it will start to click into place.

  • @ImAlecPonce
    @ImAlecPonce 11 หลายเดือนก่อน

    Thanks!!! these boxes are actually starting to make sense

  • @JRis44
    @JRis44 11 หลายเดือนก่อน +1

    This is getting fun! Cant wait to work on img to img tomorrow after work!

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      Have fun!

  • @Feelix420
    @Feelix420 11 หลายเดือนก่อน

    I'm mind blown! never thought of using comfyui ever but seems like I'm sold over this video. very nice sir and thank you for sharing your knowledge

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      Glad you liked it! It will also teach you a lot more about how things work, which I always feel is a good idea.

    • @Feelix420
      @Feelix420 11 หลายเดือนก่อน

      @@sedetweiler absolutely! downloading it now hehe

  • @shallowandpedantic2320
    @shallowandpedantic2320 11 หลายเดือนก่อน +1

    Thanks. If you're looking for recommendations, a video focused on comparing upscalers and incorporating upscaling into this kind of workflow might help people. Seems like a nice next step. Appreciate what you've shared so far.

  • @russellmm
    @russellmm 11 หลายเดือนก่อน

    quickly becoming my goto channel. keep up the great work

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      Thank you!

  • @angryDAnerd
    @angryDAnerd 11 หลายเดือนก่อน +4

    Excellent tutorial, thanks! I got SDXL up and running with the refiner. If you have the time I'd like to see you make a video explaining how Stable Diffusion works and explain exactly what the program is doing as it sends the data through the nodes in Comfy so I can have a greater conceptual understanding of what is happening. Believe me I could watch hours of technical stuff lol.

  • @eliasdelreal3863
    @eliasdelreal3863 11 หลายเดือนก่อน

    You are da man! Thank You so much for this tutorial!!!

  • @conquestofninjacats
    @conquestofninjacats 11 หลายเดือนก่อน +1

    Love your videos :D just started using ComfyUI with SDXL. Having a lot of fun so far!

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      Glad you enjoy it!

  • @zerosequels
    @zerosequels 9 หลายเดือนก่อน

    Thanks Scott, your tutorials are great

    • @sedetweiler
      @sedetweiler  9 หลายเดือนก่อน

      Glad you like them!

  • @potusuk
    @potusuk 7 หลายเดือนก่อน

    yay finally got it working :) learning fast thank you

  • @14MTH3M00N
    @14MTH3M00N 6 หลายเดือนก่อน +2

    Love your disgust for the negative prompts lists haha. relatable stuff

    • @sedetweiler
      @sedetweiler  6 หลายเดือนก่อน +3

      (((((((((extra arms!))))))))) :-)

  • @nicolasmarques7460
    @nicolasmarques7460 8 หลายเดือนก่อน

    Amazing tutorial, thanks for sharing!

    • @sedetweiler
      @sedetweiler  8 หลายเดือนก่อน

      Glad it was helpful!

  • @gbresaleking
    @gbresaleking 11 หลายเดือนก่อน

    Thanks for this ! I was using comfy and using refiner at the beginning and it was coming up with really wacky pictures, now I can use it correctly thx

  • @archielundy3131
    @archielundy3131 5 หลายเดือนก่อน +1

    A million thanks for these. As finicky and frustrating as the program is for beginners, your calm expertise is just what's needed.

    • @sedetweiler
      @sedetweiler  5 หลายเดือนก่อน +2

      Thank you!

  • @Pfaeff
    @Pfaeff 11 หลายเดือนก่อน +25

    Why are there width and height values in the CLIPTextEncoderSDXL and what is the difference between width and target_width and why is one of them 4096?

    • @courtneyb6154
      @courtneyb6154 10 หลายเดือนก่อน +4

      Great questions and hopefully Scott can take the time to explain. Building out the workload is a great first step, but not knowing what everything does so that you can fine tune it is lame.

  • @clonosaurios
    @clonosaurios 8 หลายเดือนก่อน +1

    Thank you for your video! I learnt that comfyui is awesome :)

    • @sedetweiler
      @sedetweiler  8 หลายเดือนก่อน +1

      You're so welcome!

  • @ColbstaD
    @ColbstaD 7 หลายเดือนก่อน

    Thank you for this! I've created my own custom workflow based on this one with lots of inputs --> primitives to change stuff quickly.

    • @sedetweiler
      @sedetweiler  7 หลายเดือนก่อน +1

      Fantastic!

  • @tomaseriksson5430
    @tomaseriksson5430 9 หลายเดือนก่อน

    thanks for this! looks so crazy to a beginner but I followed and it's a great place to start generating images

    • @sedetweiler
      @sedetweiler  9 หลายเดือนก่อน

      It really is! You have officially leveled up, as this specific type of workflow shows you how things work, not how to use a specific UI. Cheers!

  • @CMak3r
    @CMak3r 11 หลายเดือนก่อน +7

    Prompt switching can be realized with additional KSampler that will render first steps with completely different prompt. For example you may want to create triangle composition, or a symmetrical image, and it can be done at early steps of a generation. Good for abstract art. And also I like that in ComfyUI it's seed can be fixed while base model and refiner will be generating on different seeds

    • @zacharykrevitt7560
      @zacharykrevitt7560 8 หลายเดือนก่อน

      good idea! just tried this out and it worked in an ineresteing way. Essentially prompting an init image

  • @Yggdrasil777
    @Yggdrasil777 11 หลายเดือนก่อน +3

    I have been in love with ComfyUI since I found it (coming from Unreal Blueprints, very familiar system). I am currently working out some torch issues with my current system, but I generate whenever I can. It is great to see you building out the workflow and explaining the nodes that you use and why. Very informative and THANKS for the tip with the shift-click to copy nodes AND connections. NICE!

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      Great to hear! I am really happy with the nodes, but I hope they really update to things like docking, etc. Cheers!

    • @digitalbear3831
      @digitalbear3831 11 หลายเดือนก่อน +1

      Same here since I come from Houdini, just love the node spagetti

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      Yusss! I also used Houdini as well as Substance Designer and I am hoping to get into nested nodes here as well. Cheers!

  • @TomMaiaroto
    @TomMaiaroto 11 หลายเดือนก่อน +2

    I'm new to ComfyUI all and really love your videos. Thanks! Maybe this is obvious to folks, but one thing I recently learned was the ability to condition after one KSampler ran so you can continue to refine your final image. It ended up being an alternative (or another tool in the toolbelt) to inpainting. I wasn't just refining, I was adding to or dramatically changing the final image - all without losing the "base" starting point that was all "locked down" in that the seed was fixed, the cfg and steps didn't change, etc. So it was a very non-destructive compositional workflow. If I wanted to add an object to the image, I could do that through a second prompt that was applied to a second KSampler.
    I could also introduce new LoRAs later on in those steps. I'm going to continue to experiment with this strategy and go through this more than once. So instead of a long prompt followed by a smaller corrective one, do more of a build up of prompts. Start simple and continue to add on to it so that elements within the image can be independently adjusted, removed, or re-arranged. Again, a more compositional approach during image generation to hopefully reduce the amount of work in post (or a series of very similar images that can be worked together in post processing). This could get a bit messy too, but maybe not if they are arranged left to right in a linear fashion building up the scene.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      That's great! It is a lot of fun adding into the pipeline. It's what we do internally as well when testing models and playing with new ideas. Cheers!

  • @DarnSylon
    @DarnSylon 10 หลายเดือนก่อน

    I kept to the two sampler idea but introduced separate prompts for the first sampler. It creates an interesting variation. Especially if you put different colors in each positive prompt. For example, make a white bridge and a black animal. I was hoping to create a background and a foreground. But the twist turned out pretty cool anyway.

  • @MonzonMedia
    @MonzonMedia 11 หลายเดือนก่อน

    Appreciate this Scott, you helped me fill in the blanks! I was wondering how the primitive nodes were used, it was driving me nuts! Hahaha! I was able to add an extra step to add an upscale process and it works very well! Looking forward to more.

    • @MonzonMedia
      @MonzonMedia 11 หลายเดือนก่อน +1

      By the way, is there a way to create an image gallery somehow? Sort of like how invoke Ai is set up?

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      I just use Bridge, since I already have an Adobe sub and it's better than most gallery apps. Comfy really isn't good for that type of thing at this point.

  • @renderartist
    @renderartist 11 หลายเดือนก่อน

    Thank you! 👍🏼

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      You're welcome!

  • @larryross9380
    @larryross9380 2 หลายเดือนก่อน

    Perhaps things have changed since this was published nine months ago, because this workflow just gave me dark, abstract images. But I learned a lot about how to build out a workflow! Thanks!5

  • @hleet
    @hleet 11 หลายเดือนก่อน +1

    WOW ! that's a super tutorial of ComfyUI there ! Thanks. I never know that there was this new addition of clipnode for SDXL !
    The only drawback that I find in ComfyUI is the way it manage the workflows. I mean when you want to change your original workflow, you need to save a local file, and if you want to do something else (like inpainting) you have to redo ALL your workflow and save it to a file to recall your workflow and switch by loading one workflow or another depending on what you want to do. Definitly not fond of this way of managing workflows. They could have done some kind of "favorite" workflow. Like 5 or more "workflow ready" that you could custom afterwards and save your "favorite custom workflow" and switch whenever you like. it would skyrocket the use and adoption of comfyui !

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      I just drop the json you get from using "save" into the interface and it loads. But,*do agree that would be nice.

    • @hleet
      @hleet 11 หลายเดือนก่อน

      @@sedetweiler ooh ! Nice another tip ! Drag and drop the json just works too ! I might be able to explore more versatile stuff with comfyui now :)

  • @lukeovermind
    @lukeovermind 11 หลายเดือนก่อน +1

    fantastic! I am looking at some advance workflows, however with no real explanations how they work. I want to use it but I dont know what some of the nodes and flows do! However I found alot of value from your vids and at this stage I am happy to just play and learn comfy and put of creating art projects/ideas with SDXL for the time being.
    That 3rd Sampler is neat! I tried to see if you can use latent upscale method in your previous video with SDXL base and refiner, didnt work but that is the beauty of comfy! You get to try stuff

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      I also think it is a pretty great way to learn how all of this works together. It really is limitless!

  • @paullangdon1852
    @paullangdon1852 11 หลายเดือนก่อน

    Great video, very usefull. I am struggling to do a workflow from load image and the put through an ultimate upscale node.

  • @Aaabii
    @Aaabii 11 หลายเดือนก่อน

    Thank you very much. ı prefer comfyUI over A1111 and you are my go to channel for my purposes.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      Happy to hear that!

  • @badmojjo
    @badmojjo 11 หลายเดือนก่อน +1

    I am definitely going to search for a good upscale workflow on your channel.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      It's coming this weekend!

  • @case0h
    @case0h 11 หลายเดือนก่อน

    thank you!

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      You're welcome!

  • @user-kn6vx2xz4e
    @user-kn6vx2xz4e 9 หลายเดือนก่อน

    very good

  • @novantha1
    @novantha1 11 หลายเดือนก่อน +3

    Huh. I wonder what would happen if you had dedicated models for a variety of tasks (hands, eyes, hair, reflections, contrast, and so on) and fed a few steps from each of them in a daisy chain until you got to the first "true" sampler...
    Truly the possibilities are endless; thanks for the food for thought and the hard work!

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      That's a great idea, and we do have those as loras. It's fun to combine them to help get what you want.

    • @tripleheadedmonkey6613
      @tripleheadedmonkey6613 11 หลายเดือนก่อน +1

      That is an interesting idea. The multitude of experts approach is proving to be the more effective of what we have developed recently.
      Not too mention that you could also combine this with prompt blending syntax to ensure that each part of the processing is focusing entirely on one subject in the prompt while still maintaining an overall mixed composition.
      If for simplification purposes you set up 5 samplers, each with an equal number of steps, 4 for the limbs and 1 for the head/torso. Then you set up a prompt blending which focuses 20% of the processing on each limb etc. it may even have better results.

    • @tripleheadedmonkey6613
      @tripleheadedmonkey6613 11 หลายเดือนก่อน

      And yeah using LORA chains would mean that we could have a separate model output for each limb, while maintaining the same initial model. Allowing for less resources used at the same time compared to multiple dedicated models.

    • @tripleheadedmonkey6613
      @tripleheadedmonkey6613 11 หลายเดือนก่อน

      I think I'm going to play around with this now actually xD Minus the dedicated limb lora of course.

  • @MikevomMars
    @MikevomMars 11 หลายเดือนก่อน +1

    Using SDXL with a 2060 SUPER 8GB + ComfyUI and it works great 👍Turned out that Comfy is 15x(!!!) faster than A1111 for the same tasks using SDXL! Also I never got any errors with ComfyUI while A1111 always gives me "NaN tensor" errors when working with SDXL. For SDXL, Comfy is a MUST! So I`m looking forward for more tutorials on ComfyUI.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      Great to hear!

    • @technoprincess95
      @technoprincess95 11 หลายเดือนก่อน

      I am on the opposite; I am only be able to use Colab. When using A111, it's very fast and smooth, allowing me to work with videos and other tasks, including upscaling to 8k,comfuyi just pictures. However, with Comfuyi on Colab, I encounter issues such as disconnects and running out of ram after just 5 images, im using workflow of olivio sakiras, using base ckpt sd1.0 + refiner sd 1.0 . overheard refiner consumes lots of Ram

    • @MikevomMars
      @MikevomMars 11 หลายเดือนก่อน

      @@technoprincess95 I`d double check your workflow used then. Also did you try to run ComfyUI locally? You don`t need a decent GPU to do so. Even an old 8GB NVIDIA will do it.

  • @GamingDaveUK
    @GamingDaveUK 11 หลายเดือนก่อน

    Very useful video and great to see videos that are not automatic 1111 lol

  • @ysy69
    @ysy69 8 หลายเดือนก่อน

    ComfyUI is truly about fine tuning the way one approaches the creation of an image using AI

    • @sedetweiler
      @sedetweiler  8 หลายเดือนก่อน +1

      I agree!

  • @Adreitz7
    @Adreitz7 11 หลายเดือนก่อน +5

    Thanks for this look at the setup that Stability uses internally. I'm not so familiar with Comfy, but I've been using and enjoying SDXL through Invoke, which has a similar Nodes capability. I have a few questions and comments:
    1. What are the Original and Target W/H actually doing for the CLIP conditioning nodes and what is the logic to setting those values? I played around with it, testing various combinations, and the only thing I could confidently say is that setting Original W/H smaller than 1024 causes the image to become blurry. I couldn't see any specific benefit to any other value, as I tried 1024, 4096, and 40960 for Original and between 64 and 40960 for Target -- setting different values made the image different, but not obviously better or worse. I settled on just setting them the same as the output image dimensions.
    2. Why are there two prompt inputs for the base text encoder node when you provide the same input to both? Invoke calls one input the prompt and the other the style. What effects are caused by, e.g. separating your prompts into a prompt and a style and sending them independently to the two inputs, switching the inputs (so prompt goes to the "style" input and vice versa), setting them both the same, or leaving one or the other blank? I've found that if I prompt the base model for a roller coaster in the first input, I get a roller coaster. But if I prompt "roller coaster" for the first input and "photograph" for the second, I get anything BUT a roller coaster -- ruined buildings, abstract paintings, etc.
    3. Connected with #2, Invoke's refiner conditioning node only includes a "style" input, but I've found that only giving it a style prompt can cause the refiner to do weird things (like making architecture look like it's made of tent fabric).
    4. You've indicated that initializing the noise with the refiner is an interesting idea, which it is, but have you seen any consequences other than just making the images different? Does it provide any actual benefit?
    5. I've experimented with higher resolution SDXL generations. I'm on a Mac and there are some apparent generation bugs with Invoke on MPS (about 1856 square and above it becomes debilitating). But I've noticed that my scenes at higher resolution (photographic sci-fi style architecture) tend to become wide angle and taken from a high vantage point, almost as if the resolution setting is correlated with the position and zoom of the virtual camera. Has Stability done any experiments at higher resolutions than 1024x1024?
    6. Is there a benefit or danger to sending the same noise seed to both the base and refiner?

    • @bobbyboe
      @bobbyboe 7 หลายเดือนก่อน +1

      Good questions... I also would like to know the answer. Did you understand the concept of why there is a field of dimensions in a node that is supposed to provide only text?

  • @San4itos
    @San4itos 8 หลายเดือนก่อน

    You set Base KSampler to return the leftover noise but there is no leftover noise because it does all of its steps. Then Refiner adds its own noise and process it further. You may see it in
    Base preview. I guess if you turn off the leftover noise from Base, result is gonna be the same. What you need to do to pass the leftover noise to the Refiner is to use for example 20 steps but end on step 12. Then disable add noise function on Refiner KSampler.

  • @skylightikab443
    @skylightikab443 8 หลายเดือนก่อน +1

    Thanks for no nativ english speaker this was a good tutorial. It was very helpful! :)

    • @sedetweiler
      @sedetweiler  8 หลายเดือนก่อน

      Glad you enjoyed it!

  • @A.polon.i.a
    @A.polon.i.a 3 หลายเดือนก่อน

    Great video Scott, I wonder could you explain how to change the image size? What do I have to alter to produce an image of 832 x 1216 for example? Or point me to a future video that explains it, as I'm only on ep.2 Thanks💖

  • @parsley8188
    @parsley8188 11 หลายเดือนก่อน

    Nice

  • @matthewharrison3813
    @matthewharrison3813 11 หลายเดือนก่อน +1

    Thanks for the great video. Could you please talk more on the clip encoder width and height and target width and height? What do they do and is there any documentation? Why are you using a different value for the target than the base?

    • @4richis
      @4richis 7 หลายเดือนก่อน

      I would love to see and answer to this as well

  • @reekster30
    @reekster30 10 หลายเดือนก่อน

    wow - great tutorial dude. I've only recently got into comfy and wondered why all the controlNETs were failing last week :D All new ones install thanks to your videos and loving all the sdxl videos... fun times ahead (but I really need a pc gaming rig for speed) haha
    Out of interest - what kind of set up for a pc would you recommend for quicker generation/processing? massive 128gb RAM and like a RTX4090? :D
    thanks for your videos - amazing

  • @courtneyb6154
    @courtneyb6154 10 หลายเดือนก่อน

    Excellent video Scott. If you could do some of us a favor and go into detail about what everything is and how it works within the cliptextencode nodes then that would be of tremendous value. I have scoured the net and am only able to find limited info about the options and nothing i have found has explained how or why they work. Building out the workflow is a great first step but not knowing how to fine tune is lame 😂Thanks!!!!

    • @sedetweiler
      @sedetweiler  10 หลายเดือนก่อน

      Sure thing!

  • @demoran
    @demoran 11 หลายเดือนก่อน

    So I was following the guidance here, and found that UniPC and the 2M variants will barf on you when the refiner steps are higher than the base steps. I tried with the 12/20 pair you've demoed here, and got an image with nasty vertical streaks in it. It was fine at 20/20, but barfed again at 20/50.

  • @DarnSylon
    @DarnSylon 10 หลายเดือนก่อน +2

    When you added the third or 'pre-sampler', why did you not pass the noise information as you had done with the first of the two samplers? I messed with that setting on the first two and didn't notice much of a change. Thank you for the videos and instructions. They are extremely helpful. And you suggest not to add things like extra fingers to the negative prompt. What is your method of not getting extra fingers or limbs, etc?

  • @tripleheadedmonkey6613
    @tripleheadedmonkey6613 11 หลายเดือนก่อน +1

    One question I had. Is there any reason why you recommend using the VAE from the refiner, when there is only 1 version of the VAE (barring custom fixes for FP16) publicly available?
    If I choose to merge the fixed FP16 base VAE with the refiner, am I getting the same experience as you are (besides fp16-fp32 differences) ?

  • @florentraffray1073
    @florentraffray1073 3 หลายเดือนก่อน +1

    Thanks for these tutorials, great to have an in depth dive into the UI.
    I'm a little confused about the start/end steps and steps in the KSampler.
    In your second sampler in the chain of them, if you start at step 3 and do 12 steps, wouldn't that leave you at step 15 for your starting point in the next one?

    • @sedetweiler
      @sedetweiler  3 หลายเดือนก่อน

      There are some advantages to skipping steps in some cases. It all has to do with the residual noise.

  • @davewills6121
    @davewills6121 10 หลายเดือนก่อน

    Excellent tutorial, learned a whole lot in a short time. Why is it that the Refined images are indeed sharper, they seem to loose some of the more acute details. EXAMPLE: undersea shot without Refiner shows a murky underwater world with subtle light refraction and a sense of DOF, refiner seems to strip that away, leaving a sterile shot with little atmosphere?. Kudos

    • @sedetweiler
      @sedetweiler  10 หลายเดือนก่อน

      There is some balancing going on for sure!

    • @davewills6121
      @davewills6121 10 หลายเดือนก่อน

      @@sedetweiler Im just an eager noob getting his toes wet. PS: Can i trouble you for one question: If i pick one image from say my ''history'', can i build from that, run further batches based off of that one image, so that i can fine tune my results?

  • @Smashachu
    @Smashachu 7 หลายเดือนก่อน +1

    Hmm i'm messing around with rendering the first 2-3 steps as something that i know SDXL is trained very well in so for example a brown horse racing for a positive prompt on the first 3 frames, then using a negative prompt for the Brown, with the new color being purple with a (purple horse:1.3). It's been working very well especially for harder to generate things, it's like it's erasing the colors and redrawing it now that there's a rough shape. I'd love to see how it will workout in combination with controlnet to maintain consistency in textures and shapes.

    • @sedetweiler
      @sedetweiler  7 หลายเดือนก่อน

      That method can also help with LoRA images that are not as strong as you prefer. It's a great workflow. 🥂

  • @bigbo1764
    @bigbo1764 10 หลายเดือนก่อน +1

    I’m curious, how would I implement a lora in this setup? I tried inserting 2 lora nodes after the checkpoint nodes and connecting them like I would in SD 1.5, but it seems to not be registering the existence of my Lora and just skipping over it. My checkpoints are connected to the Lora nodes only, except for the VAE, which is used for the decoding, what am I doing wrong and how exactly do I fix this?

  • @deafponi
    @deafponi 5 หลายเดือนก่อน

    Hi there Scott, thank you for the excellent tut. I must admit though, my robots did not look anything close to how refined yours came out. I wonder if I missed anything somewhere...

  • @Aksuloid
    @Aksuloid 10 หลายเดือนก่อน

    Would it be possible for you to make a video explaining the ideas behind negative prompting? You've mentioned it twice in the ComfyUI playlist now, and I am quite interested to know why and what one would actually put in there.

    • @sedetweiler
      @sedetweiler  10 หลายเดือนก่อน +1

      It was really only an issue with SD 1.5, and not so much with SDXL. Consider for a moment that the model has no clue how many fingers a person has, so using a negative prompt of ((((((extra fingers))))) is just silly. People making it higher priority with all the "(" is like yelling English to someone that doesn't speak the language, cause that helps. The model was not trained on what "extra fingers" are, and that is just one example.

    • @Aksuloid
      @Aksuloid 10 หลายเดือนก่อน

      @@sedetweiler That makes alot of sense, thank you. From what I've gathered then, its going to be more useful for removing unwanted objects/themes/concepts from an image (for example), or steering the model away from unwanted styles, instead of fixing small anatomical mistakes, etc.

  • @MisterKerstov
    @MisterKerstov 11 หลายเดือนก่อน +1

    Thank for this really concise and helpful tutorial. Just one thought, you did not enable the "return with leftover noise" for the "initial conditioning" node. Wouldn't it make sense to do so?

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      It actually returns so much that things go sideways. Give it a try. I have not found that to work well.

  • @ysy69
    @ysy69 11 หลายเดือนก่อน +1

    Hi Scott, really appreciate your giving us the most recent update on SDXL. Do you know how to fine tuned a model using SDXL 1.0 and Dreambooth? Is this something you can create a tutorial video for us?

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +2

      That is coming soon. It is going to be easier to train, results wise, but still getting methodology together.

    • @ysy69
      @ysy69 11 หลายเดือนก่อน

      @@sedetweiler 🙏🙏looking forward to… do you know if the new dataset should be set at minimum at 1024 by 1024?

  • @karenreddy
    @karenreddy 11 หลายเดือนก่อน

    Even the simpler setup is convoluted. I've worked with shader graphs, so it's alright, but I can see how this has a bad learning curve for many. I just don't see the big gain in using this setup for this utility quite yet.
    Stability should allow it to be "baked" into a simple GUI, so you can create a front end with different graphs, then not mess with it much, unless you want to add more pieces to the front end. Saving this front end would allow it to be shared with beginners and make it easy to get into, the complexity would be hidden until they're ready to explore.
    What is the advantage to conditioning with the refiner first?

  • @adam4adam890
    @adam4adam890 11 หลายเดือนก่อน

    For some reason I always get an "overcooked" or flat images from refiner, in some cases it takes good result from the base and just removes the depth from image and turn it into drawings. Regardless very helpful information, TY

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      It is a balancing act. You might want to minimize the steps in base to around 12 and also be sure to send over the remaining noise.

  • @spiralofhope
    @spiralofhope 8 หลายเดือนก่อน +1

    I was able to follow the tutorial well. I'm a bit confused at the three separate seeds. I can adjust the first (the conditioner/initializer) and get changed results, do I care about the others? In a previous video you said it wouldn't matter much for that context. Is that also true here?

  • @vanstonecypher
    @vanstonecypher 11 หลายเดือนก่อน

    Thank you so much for the tutorial. It really helped with some basic knowledge that was not obvious as a new user of ComfyUI (double-clicking to get a list of nodes for example.) A couple of question though. Why are the width and height of the CLIPTextEncodeSDXL nodes are set to 4096? What does this mean since the output is still 1024?

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      It is the resolution CLIP was conditioned at prior to scaling. I tend to use it and prefer the result.

  • @kenjix7316
    @kenjix7316 10 หลายเดือนก่อน

    why exactly dont we change the return leftover noise for the first sampler (the 2nd refiner one added at the end)?

  • @TomSweeney-ov8qs
    @TomSweeney-ov8qs 10 หลายเดือนก่อน

    Do you have any videos (or recommendations for other videos) that go in depth on debunking the negative prompt urban legends you mention?

    • @sedetweiler
      @sedetweiler  10 หลายเดือนก่อน

      No, but I should make one. It's just terrible what people pass on as the perfect negative. Do they think the model was trained on "bad anatomy" and "extra fingers?"

  • @TissaUnderscore
    @TissaUnderscore 3 หลายเดือนก่อน

    Should i use a refiner for a custom model? for example if i use juggernaut xl?

  • @GuitarWithMe100
    @GuitarWithMe100 11 หลายเดือนก่อน +1

    Im still confused on what the ClipTextEncodeSDXL does? and how does the value 4096 affect it?

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน

      That was the initial conditioning prior to scaling, so we just prefer that for the refiner.

  • @ianwilliams7740
    @ianwilliams7740 9 หลายเดือนก่อน

    on that third sampler you added you kept the return with leftover noise to disable.. does that mean you use up all the noise in those 3 early steps? what's the thought in not setting that to enable??

  • @JohnSundayBigChin
    @JohnSundayBigChin 9 หลายเดือนก่อน +1

    Hi Scott, im rewatching the whole series again, you have done a good job. I have a question in this particular episode with the Sampler...why do you have the possibility of using the denoise within the KSampler but not with the advanced KSampler? Do they work differently?

    • @sedetweiler
      @sedetweiler  9 หลายเดือนก่อน +1

      it was to simplify things. when you start at a later step with the advanced sampler, you are "skipping" some of the pieces you do not want to denoise, so it is the same thing but harder to explain.

    • @JohnSundayBigChin
      @JohnSundayBigChin 9 หลายเดือนก่อน +1

      When you make Img2Img in one of the videos I saw that you used the common Ksampler because you needed the denoiser. Now everything is much clearer to me, thank you very much for answering.

  • @technoprincess95
    @technoprincess95 11 หลายเดือนก่อน +1

    Would you mind sharing this workflow through a gdrive ❤

  • @97DarkSkull
    @97DarkSkull 11 หลายเดือนก่อน

    How would you precess sequences of images (I'm going or similar) with Comfyui?

  • @ysy69
    @ysy69 8 หลายเดือนก่อน

    Very helpful. Have you experimented with learned with using multiple Ksamplers? Are you still keeping at least 3 at different steps today as part of your workflow?

    • @sedetweiler
      @sedetweiler  8 หลายเดือนก่อน +1

      I use 2 most of the time.

  • @ethanhorizon
    @ethanhorizon หลายเดือนก่อน

    Thanks for the tutorial! Is the "noise seed" in Ksampler Advanced same as "seed" in Ksampler? You set noise seed as 4, what's the meaning of the number? What if I left it as zero?

  • @97DarkSkull
    @97DarkSkull 11 หลายเดือนก่อน

    How would you process sequences of images (I'm going or similar) with Confyui?

  • @kabochsaygo6976
    @kabochsaygo6976 8 หลายเดือนก่อน

    Is it possible for you to do a tutorial showing the ComfyUI ->Models folder structure, and what goes into each of them? I manually installed the manager with no issues. But other things such as diffusers, embeddings, clip_vision, etc. are unknown to me. And a lot of things on huggingface can't be found within the manager. Thanks. PS: Just getting started with SDXL and using Comfy. So going through your videos one at a time.

  • @kick851
    @kick851 หลายเดือนก่อน

    for the cliptextencodesdxl
    if my latent image is 768x1280 do i still use 4096 for width and height and what about the target width and height

  • @gameplayfirst6548
    @gameplayfirst6548 11 หลายเดือนก่อน +2

    How is there any noise left during handover to the refiner, if you don't use the "end_at_step" parameter? Don't you get images without any noise from the base sampler if you don't limit the end in any way?
    Your base preview image confirms that you don't have any noise left after the base, which doesn't match the workflow described in the SD-XL documentation.
    And why do you overlap steps? For example you do 12 steps in base, but start at step 12 in refiner, instead of starting at step 13.

  • @dreaminspirer
    @dreaminspirer 11 หลายเดือนก่อน +1

    thanks so much for the video.
    I'm having BASE Steps and TOTAL Steps Primitives. So I'm trying to use a Primitive node to feed the PRERUN steps to 1st Refiner (let's call it PRERUN KSampler) but i bumped into a problem.
    - Feeding "steps" into PRERUN Ksampler is fine but I can not feed this "steps" INT to "start at step" for the BASE KSampler . they're both INT, but perhaps ComfyUI considers "steps" and "start/end at step" are different types. 😒
    - The other way around is feeding "end at step" for PRERUN and feeding this value to "start at step" for BASE and feed all KSampler with same "steps" value. But for some reason, the PRERUN Ksampler needs to be fed with exact amount of steps otherwise the result is nothing but NOISE. 😒
    please help , thanks again.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +2

      I have also noted that, and I think it is a bug. That should work just fine. I got around it by using a math node, since that was the end goal anyway.

    • @dreaminspirer
      @dreaminspirer 11 หลายเดือนก่อน

      @@sedetweiler thats exactly what i found. Derfuu VAR nodes and MATH nodes did the trick without any problem.
      Having said that, i found PRERUN step should not be more than 3 or it's all crap :)
      Thanks again and pls keep sharing with us the quirky tricks to play with Comfyui

  • @Zizos
    @Zizos 10 หลายเดือนก่อน

    I just did download the official base and refiner but it seems I've got the VAE version from somewhere else in the past.
    What's the difference? I get that the VAE is built-in to the model. Does this mean you get to delete the VAE Decode node or some other node?
    Can you just keep the VAE version and follow your workflow with no difference in results or at least no negative results in quality?
    As in the last step you showed you can 1st generate a blank latent and then into the base and refiner... Seems like you can do all sorts of tricks like that to experiment with the resulting image. I wonder if it makes sense. If I get it right, it seems that the latent creates a base noise ignoring the models so that you can just get something a bit out of the box (model). Is that right?
    Thank you for the tutorial. I have lot's of stuff to learn.

  • @scottmahony4742
    @scottmahony4742 5 หลายเดือนก่อน

    models, the refiner, etc. Where can I find definitions for all these variables?

  • @alexlindgren1
    @alexlindgren1 8 หลายเดือนก่อน

    I'm aware that SD don't take account of spatial relationships, but I want to be able to replace for example a sofa in an existing image with an image of another sofa, but not sure on how to take on that challenge with SD, do you have any suggestions where to start? I don't want to manually mask each image, but I want the AI to recognize what part of the image is a sofa and mask it for me, I should just provide the image of the sofa and the "base image" of the livingroom.

  • @RamonGuthrie
    @RamonGuthrie 11 หลายเดือนก่อน

    Is there a way to change the folder path where the images are saved to easily? 📂

  • @appolonius4108
    @appolonius4108 11 หลายเดือนก่อน

    Another great video. I learned a lot doing this, but could you build off of this one and show how to add a lora to this setup? thanks :D

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      Sure thing!

    • @appolonius4108
      @appolonius4108 11 หลายเดือนก่อน

      @@sedetweiler oh! I forgot, maybe also, if you could, do an upscaler node tutorial? These would be very helpful.. I think that would be all basic stuff. These tutorials are well done and easy to follow. They really help me understand all this stuff lol.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      @@appolonius4108 I did schedule a video for tomorrow that will cover basic img2img. That is the start of the basic steps. I will get an upscale video done super soon.

    • @appolonius4108
      @appolonius4108 11 หลายเดือนก่อน

      @@sedetweiler that is awesome news, I'll be on the look out for those! Thanks, you rock!

  • @imperfectmammal2566
    @imperfectmammal2566 11 หลายเดือนก่อน +1

    Can you tell me how to use the Loras offset that came with sdxl in comfy

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +2

      Yes, I will post a video on that and it is SUPER easy to do! Cheers!

  • @lioncrud9096
    @lioncrud9096 11 หลายเดือนก่อน +1

    any tips on adding an upscaler?

    • @sedetweiler
      @sedetweiler  10 หลายเดือนก่อน +1

      Upscaler video is out today! Woot!

    • @lioncrud9096
      @lioncrud9096 10 หลายเดือนก่อน

      thanks!!@@sedetweiler

  • @lechatsportif124
    @lechatsportif124 6 หลายเดือนก่อน

    Couple of things, isn't it recommended for the refiner to actually be started at 80% of total steps? Also, is conditioning via the refiner really a thing or did you just kind of mess around with it? You didn't select pass on noise, so I'm not sure what that means.
    Thank you for the tutorials, they are great!

  • @martdiamond6109
    @martdiamond6109 11 หลายเดือนก่อน

    Do you also use an upscaler with SDXL? All the Comfyui examples I've seen never include it, so I'm just wondering how that would look in this workflow?

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      Ya, you can use any upscaler and use them repeatedly. It's way more flexible than AUTO1111. I will do a video on this super soon. Cheers!

  • @RobertWildling
    @RobertWildling 6 หลายเดือนก่อน

    Hmmm... at around 14:15, when you add the first refiner with the 3 steps, shouldn't the last refiner's "start_at_step" be changed to 15?

  • @michaelroper87
    @michaelroper87 5 หลายเดือนก่อน

    What does it mean if it nevers leave the load checkpoint stage?

  • @Darkwing8707
    @Darkwing8707 11 หลายเดือนก่อน +3

    Why did you choose 4096 for the height and width in the conditioners?

    • @digitalbear3831
      @digitalbear3831 11 หลายเดือนก่อน +1

      I'd like to know that one too

    • @AdamDesrosiers
      @AdamDesrosiers 11 หลายเดือนก่อน +2

      also would like to know what these conditioners numbers do. And somehow, I've been happier with outputs when I set those number to 2048. But why? I don't know what those are doing.

    • @sedetweiler
      @sedetweiler  11 หลายเดือนก่อน +1

      The refiner was initially conditioned at that size prior to scaling, so we tend to use that size.

    • @jonnyfat
      @jonnyfat 11 หลายเดือนก่อน

      @@sedetweiler Thanks for this tutorial - great reference. Great to have tutorials on this by someone who knows what they're talking about :-) I picked up on the size thing too - so it's 4096 for the base and 1024 for the refiner? Thanks!

    • @petec737
      @petec737 4 หลายเดือนก่อน

      @@sedetweiler "we tend to use that size" isn't really an answer. The only reason you'd have those numbers different is if you want to CROP a portion of the image..so in your case it's like wanting to crop out a 4096x4096 OUT OF a 1024x1024 image; which obviously is not how math works :)