Google's prompt-to-prompt AI for Stable Diffusion - tutorial!

แชร์
ฝัง
  • เผยแพร่เมื่อ 27 ก.ย. 2024

ความคิดเห็น • 106

  • @darkstatehk
    @darkstatehk ปีที่แล้ว +32

    Second time your channel has taught me something amazing.

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +2

      Awesome, and many thanks! 😀

    • @alpineswiftai
      @alpineswiftai ปีที่แล้ว +2

      Im on my 6th time.

  • @ducktaperules
    @ducktaperules ปีที่แล้ว +24

    This prompt manipulation looks way better than the previous img2img that exists. But i know that if i spend time installing manually, then it will be added to 1111 within a few days.

    • @chris-hayes
      @chris-hayes ปีที่แล้ว +7

      1111 is wild

    • @LOLO3000
      @LOLO3000 ปีที่แล้ว +2

      1111 is giving us Christmas twice a week

  • @darkstatehk
    @darkstatehk ปีที่แล้ว +15

    You sir are a national knowledge sharing treasure.

  • @ASteelNeuron
    @ASteelNeuron ปีที่แล้ว +6

    Hey! Thanks for the great breakdown :) I have a quick question: Is it possible to run prompt-to-prompt on images that weren't originally SD generated via prompt? (i.e instead of generating an original image of potates, just start from a real picture of potatoes and add the prompt to it).

  • @kavellion
    @kavellion ปีที่แล้ว +1

    nice to finally see your actual face. thanks for the tutorial!

  • @patricksweetman3285
    @patricksweetman3285 ปีที่แล้ว

    Making it all look just too easy, as usual. Thanks.

  • @mathef
    @mathef ปีที่แล้ว +2

    Great video, thank you :) Is it possible to run this on Collab, please?

  • @banzai316
    @banzai316 ปีที่แล้ว

    Thank you, thanks Google as well!

  • @tiagotiagot
    @tiagotiagot ปีที่แล้ว +1

    At this rate, we're gonna have holodeck-style natural language image editing before the end of the year...

  • @yourrightimsooosorry884
    @yourrightimsooosorry884 ปีที่แล้ว +1

    Just downloaded it on my Android phone and it's amazing!!!🖖😁

    • @EonSound
      @EonSound ปีที่แล้ว +2

      @Mr Mavvy malware

    • @chris-hayes
      @chris-hayes ปีที่แล้ว +1

      Maybe a reference to the dude that created an app to run Stable Diffusion on mobile. Called "Maple Diffusion" or something like that.

    • @chris-hayes
      @chris-hayes ปีที่แล้ว

      @Mr Mavvy a demo was shared on Twitter, though I wouldn't recommend downloading it unless you're okay with waiting 10 minutes per a single image 😂

  • @PeeP_Gainz_DEV
    @PeeP_Gainz_DEV ปีที่แล้ว +2

    Based on my prediction, stabilization of the MetaVerse will definitely depend on our prompt to prompt interactions. It’s becoming clear that each computer connection will have its own visual interface, but the rules must remain constant. Nice video

  • @adam0514
    @adam0514 ปีที่แล้ว +1

    Thank you very much for this knowledge, I will wait for the Colab link please.

  • @ysy69
    @ysy69 ปีที่แล้ว +4

    Thanks a great deal for your videos! I saw the other one about bringing the AI generated characters to video! When you bring your different avatar images to life on every video, do you use the sane technique or able to use another solution to record your videos using different avatar images in real time ?

  • @culpritdesign
    @culpritdesign ปีที่แล้ว +1

    Fantastic, thank you so much

  • @swannschilling474
    @swannschilling474 ปีที่แล้ว

    Thank you so much for this one!! 🤩

  • @aa-xn5hc
    @aa-xn5hc ปีที่แล้ว

    Outstanding channel!

  • @jagratpatel99
    @jagratpatel99 ปีที่แล้ว +3

    Can you suggest me a best method to run stable diffusion local in 4GB NVIDIA card

    • @starblazer64
      @starblazer64 ปีที่แล้ว +1

      That's a good question! Whenever I have free time, I'll have to figure it out too!

    • @GreenAppelPie
      @GreenAppelPie ปีที่แล้ว +3

      Automatic1111 runs it reasonably well on a 2070 with 4GB.

    • @mrdragonrider
      @mrdragonrider ปีที่แล้ว +2

      Your best bet is running cmdr2's Stable Diffusion UI. I run it on GTX 1050 with 3 GB VRam without a hitch. In fact, I am watching this video while Stable Diffusion is rendering a batch of images in the background.

    • @jagratpatel99
      @jagratpatel99 ปีที่แล้ว

      @@mrdragonrider thanks man

    • @hepzibahhez9965
      @hepzibahhez9965 ปีที่แล้ว

      @@mrdragonrider what resolution can you go upto?

  • @shiftyj1495
    @shiftyj1495 ปีที่แล้ว

    I was just trying to install it on a cloud gpu provider using a jupyter. When I reached step "4.) conda activate diffusers2" I got the following error:
    CommandNotFoundError: Your shell has not been properly configured to use 'conda activate'.
    To initialize your shell, run
    $ conda init
    Currently supported shells are:
    - bash
    - fish
    - tcsh
    - xonsh
    - zsh
    - powershell
    See 'conda init --help' for more information and options.
    IMPORTANT: You may need to close and restart your shell after running 'conda init'.
    Any help would be much appreciated!

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      I wouldn’t use conda in colab personally as it’s already a separate environment

  • @gamedev1905
    @gamedev1905 ปีที่แล้ว

    I want this!

  • @arockiyadiaz1464
    @arockiyadiaz1464 ปีที่แล้ว

    Please give us the jupyter notebook to work with

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      Link is in the description ;)

  • @b0b6O6
    @b0b6O6 ปีที่แล้ว

    How are you doing the talking head? Is it DFL? Please, I need to know.

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +2

      I’m using the thin plate spline motion model for image animation. Link is in the video description 😉

    • @b0b6O6
      @b0b6O6 ปีที่แล้ว

      @@NerdyRodent Thank you!

  • @flonixcorn
    @flonixcorn ปีที่แล้ว

    Great video, do u know if this is already implemented in automatic1111 webui

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      It’s been 1 day so I’d be amazed if it hasn’t 😉

  • @KingZero69
    @KingZero69 ปีที่แล้ว

    how do you make your face cam... what program do you use to remap your video onto that anime looking character... stable diffusion? what else?

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +1

      Link in the description 😉

    • @KingZero69
      @KingZero69 ปีที่แล้ว +1

      @@NerdyRodent that’s very awesome… thank you for all you do 😎💪

  • @Fangornmmc
    @Fangornmmc ปีที่แล้ว

    Question for more advanced users. What if I want to train SD with multiple different images of my own. For example: suppose I want to train SD on blue eyes and on green eyes. Currently it's easy enough to train on either and get a .ckpt, but if I want to include both trainings in my model, what then?
    1. Using checkpoint merger - gives bad results, destroys oldest training the most.
    2. Train on blue eyes, use output .ckpt as input .ckpt for training on green eyes - seems to work but older trainings do get lost over time.
    2b. Same method as above but use different class images for each training. Even though both trainings are eyes (and therefore all the class images are images of eyes). The idea being that loss is reduced as new trainings are added,
    I am just not sure what the best way is to expand my model with multiple new trainings. Hope someone here has any pointers.

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      I just used different tokens and daisy chained, but there is a repo that uses image filenames to allow multiple things at once :)

  • @templeofleila
    @templeofleila ปีที่แล้ว

    I hate to ask a question off topic but what program are you using for your avatar? And not obs.. I mean just the mocap Avatar

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +1

      Link in the description 😉

  • @guumaster_
    @guumaster_ ปีที่แล้ว

    Is this already implemented in AUTOMATIC1111 web-ui? or weighted prompts are different than this?

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      The code has already been out for a day so it should be implemented already!

    • @fabgeb667
      @fabgeb667 ปีที่แล้ว

      @@NerdyRodent hmmm so how do i install itt for local windows?

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      @@fabgeb667 Run the 7 commands in the description 😉

  • @havemoney
    @havemoney ปีที่แล้ว

    When can we expect Stable Diffusion for Radeon?

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +1

      Already works on AMD cards 😄 Easiest way is to use the rocM pytorch.

    • @havemoney
      @havemoney ปีที่แล้ว

      ​@@NerdyRodent thanks

  • @fernandoguerravisualart2405
    @fernandoguerravisualart2405 ปีที่แล้ว

    Is strange that I have a problem running this on colab?

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      Should run ok on a T4, but if you get OOM try the low vram setting

    • @fernandoguerravisualart2405
      @fernandoguerravisualart2405 ปีที่แล้ว

      @@NerdyRodent no my problem is more about a ptp_utils module not found, even if I clone the repo

  • @Punpunsara
    @Punpunsara ปีที่แล้ว

    can it img2img?

  • @RHYTE
    @RHYTE ปีที่แล้ว

    now there only needs to be a gui for this lmao

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      Other than the jupyter web ui? 😆

    • @RHYTE
      @RHYTE ปีที่แล้ว

      ​@@NerdyRodent My smooth brain self is barely able to use automatic's web GUI so ... probably yeah

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      @@RHYTE Lols XD

  • @shadowolf3998
    @shadowolf3998 ปีที่แล้ว

    How did you make a V-Avatar like that from the AI Image?

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      Link is in the description 😉

  • @starblazer64
    @starblazer64 ปีที่แล้ว

    One of these days, ill fifure out how to do a face swap and voice change

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +2

      I’d check out my videos for voice cloning and avatar animations 😉

  • @MitrichDX
    @MitrichDX ปีที่แล้ว

    GUI for this???

    • @4.0.4
      @4.0.4 ปีที่แล้ว

      Probably Automatic's webui will have it.

  • @drawmaster77
    @drawmaster77 ปีที่แล้ว

    rodent with 3 legs and lion with 12 fingers.

  • @IceMetalPunk
    @IceMetalPunk ปีที่แล้ว +3

    Just waiting for it to be optimized to work with 8GB of VRAM and added to Automatic1111's repo... I give it a week 😁

  • @timeTegus
    @timeTegus ปีที่แล้ว +2

    I love that pytourch works on every platform and os. And u don't have to port the application.

  • @contrarian8870
    @contrarian8870 ปีที่แล้ว +1

    Why does the visualization split "rodent" into "ro" and "dent"? These two syllables make no sense by themselves.

  • @AginoEvolutionHD
    @AginoEvolutionHD ปีที่แล้ว

    I'm new here, but let me ask this question but how to make you face like this wend you toking I know the is face to face but I need to know who you did which a webcam or phone

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      Link is in the description 😉

    • @AginoEvolutionHD
      @AginoEvolutionHD ปีที่แล้ว

      @@NerdyRodent thx i will look up later

  • @bloomp7999
    @bloomp7999 ปีที่แล้ว

    this next gen coding when there's the parameter "croutons"

  • @catrocks
    @catrocks ปีที่แล้ว

    Thanks Nerdy ♥

  • @dacentafielda12
    @dacentafielda12 ปีที่แล้ว +1

    I haven't bothered with Automatic's web ui yet. But i'll get into it when this is added. I don't have the time to bother with troubleshooting installation with this method cause I always run into an issue.

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +1

      Yeah, I prefer just using the normal conda method for A1111 as well!

  • @the_hero_shep
    @the_hero_shep ปีที่แล้ว +1

    Any chance you could put this into a colab?? I'd be super thankful

  • @4.0.4
    @4.0.4 ปีที่แล้ว +1

    Ohh, can't wait till I press git pull on Automatic's webui and it magically just runs like a charm.

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +3

      We'll probably have to wait at least 10 minutes for that though! ;)

  • @Because_Reasons
    @Because_Reasons ปีที่แล้ว +1

    is it possible to just use a local CKPT instead of diffusors token?

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +1

      Yup, you can convert the ckpt to diffusers format then use that directory instead!

    • @Because_Reasons
      @Because_Reasons ปีที่แล้ว

      @@NerdyRodent Been trying all morning. Managed to convert the model to diffusors, but now getting all types of errors trying to invoke it using.
      device = torch.device("cuda" if torch.cuda.is_available() else "mps" if torch.backends.mps.is_available() else "cpu")
      pipe = StableDiffusionPipeline.from_pretrained("C:/Users/chlyw/Desktop/Automatic1111/prompt-to-prompt-main/diffusors/model_index.json")
      pipe = pipe.to(device)
      **
      Repo id must be in the form 'repo_name' or 'namespace/repo_name':

  • @RhetoricalTraveling
    @RhetoricalTraveling ปีที่แล้ว

    "hold on to your papers" 🤣

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      I love Two Minute Papers as well - awesome channel!

  • @ariefnurandono9917
    @ariefnurandono9917 ปีที่แล้ว

    how do you make your avatar?

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว

      Link is in the description 😉

  • @NeonTheCoder
    @NeonTheCoder ปีที่แล้ว

    Can't wait for this to be in auto's gui

  • @DajaMythBusters
    @DajaMythBusters ปีที่แล้ว

    Dude, your videos are the best, thanks for teaching us all this cool stuff.

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +1

      Glad you’re having fun! 😀

  • @scienceandmatter8739
    @scienceandmatter8739 ปีที่แล้ว

    Can you make a tutorial for streaming like you . What Programm, how to choose what you Look like etc.

    • @hajiamano22
      @hajiamano22 ปีที่แล้ว +3

      He's using the Thin-Plate Spline Motion Model for Image Animation. I sent a link in the first comment, but I think TH-cam got rid of it in the spam filter. You can type in the name of the model with "hugging face" after it to see a hosted version.

    • @tiagotiagot
      @tiagotiagot ปีที่แล้ว +2

      He has done a tutorial already. No streaming though, it's done after recording; not fast enough for real-time just yet.

  • @michaelli7000
    @michaelli7000 ปีที่แล้ว

    this is really amzing, may i ask how to make the talking avatar realtime at the right side corner? thanks a lot

    • @NerdyRodent
      @NerdyRodent  ปีที่แล้ว +1

      Link is in the description 😉

    • @michaelli7000
      @michaelli7000 ปีที่แล้ว

      @@NerdyRodent ok thanks a lot : )