Bernard Maltais
Bernard Maltais
  • 27
  • 388 376
New Stable Diffusion amazing LoHA training method
I present a new training method to obtain very clean and sharp models with high likeness and flexibility. The method does not require caption files and produce better results than all previous methods used.
Here is the resulting model if you want to try it: civitai.com/models/79210/kelly-jenner
มุมมอง: 10 022

วีดีโอ

How to use tensorboard to look at Stable Diffusion training logs
มุมมอง 4.2Kปีที่แล้ว
TensorBoard is a tool for visualizing and understanding machine learning models and experiments1. It can track and display metrics such as loss and accuracy, show the model graph, and project embeddings to a lower dimensional space2. Stable Diffusion is a method for training generative models using diffusion processes. In this video, you will learn how to use TensorBoard to look at Stable Diffu...
Modeling The White Queen as a Stable Diffusion LoRA
มุมมอง 6Kปีที่แล้ว
This tutorial will go through my new process for producing LoRAs. It should give you tips and tricks on how to create your own. Civit.ai model: civitai.com/models/10347/the-white-queen
Quickly fix bad faces using inpaint
มุมมอง 24Kปีที่แล้ว
This is a short tutorial on how to fix bad faces in stable diffusion using the inpaint feature.
Creating a LoRA weight using kohya_ss GUI, part 2: Training the model and creating outputs
มุมมอง 63Kปีที่แล้ว
This is the part two of creating a LoRA weight. It will explain how you go about creating your model and generating realistic outputs from prompts. kohya_ss GUI: github.com/bmaltais/kohya_ss kohya_ss LoRA extension for webui: github.com/kohya-ss/sd-webui-additional-networks Part 1: th-cam.com/video/N4_-fB62Hwk/w-d-xo.html
Creating a LoRA weight using kohya_ss GUI, part 1: Dataset creation and preparation
มุมมอง 82Kปีที่แล้ว
This is the part one of creating a LoRA weight. It will explain how you can go about using images available via google search to create your model. kohya_ss GUI: github.com/bmaltais/kohya_ss kohya_ss LoRA extension for webui: github.com/kohya-ss/sd-webui-additional-networks Part two: th-cam.com/video/k5imq01uvUY/w-d-xo.html
Quickstart: Training a stable diffusion model using kohya_ss GUI
มุมมอง 43Kปีที่แล้ว
Quickstart tutorial on how to train a Stable Diffusion model using kohya_ss GUI.
FSD Beta 10.11.1 Canada - Une promade de samedi matin - Complet
มุมมอง 7172 ปีที่แล้ว
Une promenade d'environ une heure par un beau samedi matin d'avril à Gatineau, Quebec, Canada. En français pour mes amis Quebecois et français de partout dans le monde.
FSD Beta 10.11.1 - Canada, Quebec - Test new camera extension and audio recorder
มุมมอง 2262 ปีที่แล้ว
Short test to see how the camera extension and new audio recorder perform. You are the judges. Test court pour évaluer comment l'extension pour la caméra et le nouvel enregistreur audio performe. Vous êtes les juges.
FSD Beta 10.11.1 20220329 end of day drive
มุมมอง 1502 ปีที่แล้ว
Conduite avec FSD Beta 10.11.1 sur routes avec de très mauvaise lignes suite à notre hiver rigoureux. Drive with FSD Beta 10.11.1 on roads will really bad markings due to the rough winter we just had.
Model 3 - Beta FSD 10.11.1 - 2022.4.5.20 - 28 March 2022 - Long drive with roundabouts
มุมมอง 2302 ปีที่แล้ว
Highlights: 0:00 Intro 10:37 Uncomfortably close to roundabout exits 13:22 Kids playing hockey in the street 13:37 Almost running over a bicycle on the side of the road 14:08 Passing a rolling car exiting from his driveway in a neighbourhood street... not cool... 15:30 Turning the wrong direction toward another car 24:03 Not positioning in the right lane to turn right after the highway exit 31:...
FSD Beta 10.11.1 Canada, from good to bad (weather)
มุมมอง 3142 ปีที่แล้ว
A drive with FSD Beta leaving with nice clear weather and ending in a snow-covered road challenging FSD Beta.
Model 3 - Beta FSD 10.11.1 - 2022.4.5.20 - 27 March 2022 - Route du Pirate, Gatineau, Quebec
มุมมอง 2372 ปีที่แล้ว
Etablissons le standard d'une route typique répétable afin d'évaluer la progression de FSD Beta à Gatineau, Quebec.
Model 3 - Beta FSD 10.11.1 - 2022.4.5.20 - 27 March 2022 - Premiere Sortie Gatineau Quebec
มุมมอง 1.9K2 ปีที่แล้ว
1iere video de FSD Beta Canada 10.11.1 prise le jour du lancement de celle-ci.
Cybertruck testing side mirrors and turning
มุมมอง 38K2 ปีที่แล้ว
www.google.com/maps/contrib/104756815649499553761/place/ChIJDQFm_N1qkYARblRN1PJYIpU/@37.9765144,-121.0400129,9z/data=!4m6!1m5!8m4!1e2!2s104756815649499553761!3m1!1e1
Tesla Cybertruck testing turning radius and visible side mirrors
มุมมอง 100K2 ปีที่แล้ว
Tesla Cybertruck testing turning radius and visible side mirrors
Tesla Gigafactory Texas Waterfall
มุมมอง 3663 ปีที่แล้ว
Tesla Gigafactory Texas Waterfall
Stuff you see driving a Tesla vol.1
มุมมอง 3783 ปีที่แล้ว
Stuff you see driving a Tesla vol.1
Comment ajuster son frunk sur sa Tesla Model 3
มุมมอง 3503 ปีที่แล้ว
Comment ajuster son frunk sur sa Tesla Model 3
YouTube on my Tesla with authentication
มุมมอง 1963 ปีที่แล้ว
TH-cam on my Tesla with authentication
Chauffage de la Tesla model 3 2021
มุมมอง 1943 ปีที่แล้ว
Chauffage de la Tesla model 3 2021
Tesla Model 3 - Recharge de la batterie
มุมมอง 2.4K3 ปีที่แล้ว
Tesla Model 3 - Recharge de la batterie
HOWTO Deploy Azure infrastructure using the GC Cloud Accelerators Azure Development VM
มุมมอง 1245 ปีที่แล้ว
HOWTO Deploy Azure infrastructure using the GC Cloud Accelerators Azure Development VM
HOWTO GC Cloud Accelerators Azure Dev VM
มุมมอง 1215 ปีที่แล้ว
HOWTO GC Cloud Accelerators Azure Dev VM
How to add custom ringtone on android
มุมมอง 836 ปีที่แล้ว
How to add custom ringtone on android
Halloween with Philips hue lightning strobe light
มุมมอง 10K7 ปีที่แล้ว
Halloween with Philips hue lightning strobe light
Confused Canadian geese
มุมมอง 2749 ปีที่แล้ว
Confused Canadian geese

ความคิดเห็น

  • @tim1398
    @tim1398 17 วันที่ผ่านมา

    How can we log more items like time_step etc?

  • @MAASTEER007
    @MAASTEER007 29 วันที่ผ่านมา

    Thanks a lot

  • @user-wk4np7fh2v
    @user-wk4np7fh2v หลายเดือนก่อน

    cant even open the file browser to pick a model "TCL was probably not installed correctly" holy shit trying to do anything with this makes me want to commit suicide

  • @vivavideo-videofilmer
    @vivavideo-videofilmer หลายเดือนก่อน

    horrible tutorial

  • @bokabanana6398
    @bokabanana6398 2 หลายเดือนก่อน

    hello does kohya_ss upload my training images to a server or something? i have alot output traffic if i try to train with it

  • @tharindumadushan1569
    @tharindumadushan1569 2 หลายเดือนก่อน

    Can you give us New Tutorial they is change everything now :(

  • @HanSolocambo
    @HanSolocambo 3 หลายเดือนก่อน

    2:45 So g-r-e-a-t, thanks. I can finally check my tensorboards even after a Kohya/Onetrainer training is finished. Didn't how to access those logs folders in tensorboard anymore. Cheers for that. Very helpful video (b^-^)b

  • @jameswest9552
    @jameswest9552 3 หลายเดือนก่อน

    Juste efficace, merci 😉

  • @ismaildugu1998
    @ismaildugu1998 4 หลายเดือนก่อน

    help please it says File "D:\AI\KohyaTraining\kohya_ss\sd-scripts\train_db.py", line 118, in train text_encoder, vae, unet, load_stable_diffusion_format = train_util.load_target_model(args, weight_dtype, accelerator) File "D:\AI\KohyaTraining\kohya_ss\sd-scripts\library\train_util.py", line 4113, in load_target_model text_encoder, vae, unet, load_stable_diffusion_format = _load_target_model( File "D:\AI\KohyaTraining\kohya_ss\sd-scripts\library\train_util.py", line 4067, in _load_target_model text_encoder, vae, unet = model_util.load_models_from_stable_diffusion_checkpoint( File "D:\AI\KohyaTraining\kohya_ss\sd-scripts\library\model_util.py", line 1008, in load_models_from_stable_diffusion_checkpoint info = unet.load_state_dict(converted_unet_checkpoint) File "D:\AI\KohyaTraining\kohya_ss\venv\lib\site-packages\torch n\modules\module.py", line 2152, in load_state_dict raise RuntimeError('Error(s) in loading state_dict for {}: \t{}'.format( RuntimeError: Error(s) in loading state_dict for UNet2DConditionModel: size mismatch for down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for mid_block.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for mid_block.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). Traceback (most recent call last): File "C:\Users\isodu\AppData\Local\Programs\Python\Python310\lib unpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "C:\Users\isodu\AppData\Local\Programs\Python\Python310\lib unpy.py", line 86, in _run_code exec(code, run_globals) File "D:\AI\KohyaTraining\kohya_ss\venv\Scripts\accelerate.exe\__main__.py", line 7, in <module> File "D:\AI\KohyaTraining\kohya_ss\venv\lib\site-packages\accelerate\commands\accelerate_cli.py", line 47, in main args.func(args) File "D:\AI\KohyaTraining\kohya_ss\venv\lib\site-packages\accelerate\commands\launch.py", line 1017, in launch_command simple_launcher(args) File "D:\AI\KohyaTraining\kohya_ss\venv\lib\site-packages\accelerate\commands\launch.py", line 637, in simple_launcher raise subprocess.CalledProcessError(returncode=process.returncode, cmd=cmd) subprocess.CalledProcessError: Command '['D:\\AI\\KohyaTraining\\kohya_ss\\venv\\Scripts\\python.exe', 'D:\\AI\\KohyaTraining\\kohya_ss/sd-scripts/train_db.py', '--bucket_no_upscale', '--bucket_reso_steps=64', '--cache_latents', '--enable_bucket', '--min_bucket_reso=256', '--max_bucket_reso=2048', '--learning_rate=1e-05', '--learning_rate_te=1e-05', '--logging_dir=D:/AI/kohya destination\\log', '--lr_scheduler=cosine', '--lr_scheduler_num_cycles=1', '--lr_warmup_steps=200', '--max_data_loader_n_workers=0', '--resolution=512,512', '--max_train_steps=2000', '--mixed_precision=fp16', '--optimizer_type=AdamW8bit', '--output_dir=D:/AI/kohya destination\\model', '--output_name=last', '--pretrained_model_name_or_path=D:/AI/sd.webui/webui/models/Stable-diffusion/dreamshaper_8.safetensors', '--save_every_n_epochs=1', '--save_model_as=safetensors', '--save_precision=fp16', '--train_batch_size=1', '--train_data_dir=D:/AI/kohya destination\\img', '--v2', '--xformers']' returned non-zero exit status 1.

  • @tristanwheeler2300
    @tristanwheeler2300 4 หลายเดือนก่อน

    oldie but goldie

  • @bulgingbattery2050
    @bulgingbattery2050 4 หลายเดือนก่อน

    I have a folder with 200+ images in it, and Kohya_SS says that there are not images in that folder.

    • @justinschumann8215
      @justinschumann8215 4 หลายเดือนก่อน

      You need to put in a folder containing the folder with the images, aka the parent folder of the image folder.

  • @user-kh2zk2rn7i
    @user-kh2zk2rn7i 4 หลายเดือนก่อน

    Hello Bernard, I followed your excellent presentation step by step, but I cannot obtain models under the "safetensors" extension! In the "Models" folder there is only one .json file... Do you have a solution to offer me? Thanking you.

  • @quercus3290
    @quercus3290 4 หลายเดือนก่อน

    awesome, a tutorial that actualy tells you what you need to know. Well done.

  • @oritfx
    @oritfx 5 หลายเดือนก่อน

    Thank you for the tutorial Bernard. I was wondering if it was possible to generate a smaller file (I am thinking LoRA size with dreambooth results). Do you know if that can be accomplished?

  • @jackkoveleskie2504
    @jackkoveleskie2504 6 หลายเดือนก่อน

    I just had the exact same phenomenon over my house! Hundreds of confused geese. It was weird! So I googled "confused geese" and saw your video. Gettysburg, PA. 12 Jan 2024. 1615

  • @flodyto
    @flodyto 6 หลายเดือนก่อน

    Hola gracias por el tutorial. Hay alguna alternativa gratuita a Gigapixel que pueda hacer lo mismo???

  • @Genteka
    @Genteka 7 หลายเดือนก่อน

    This videos obsolete now, all can be done on tensor art.

  • @RodrigoLM92
    @RodrigoLM92 7 หลายเดือนก่อน

    hey! I recently started following your channel and watching your videos about Kohya, LoRa and other stuff... but I'm having difficulty with some things... I would like to know if you can help me create a LoRa and in the meantime clarify some doubts that I'm having... I'm willing to pay for your time. btw congratulations on your videos... they are very enlightening and dynamic haha

  • @damasoroma
    @damasoroma 7 หลายเดือนก่อน

    Hi Bernard, Thanks a million for your efforts and your tutorial, I watched so many tricks thanks to your video! I was wondering I could I create a specific part of a body I would like to focus on man's chest (muscles and hair) and I was wondering if I need to takejust chest training images or the full body pictures of a man. And what about regularization pics? Should I take just chest or face or the full body? I'm a bit confused. Thanks a million!

  • @MrCracktuber
    @MrCracktuber 7 หลายเดือนก่อน

    Merci Berny

  • @tonychong8181
    @tonychong8181 7 หลายเดือนก่อน

    Thank you so much for all your work! Learned a lot from this video. Subscribed!

  • @RobertMichael-jh6fu
    @RobertMichael-jh6fu 8 หลายเดือนก่อน

    I have a 4090, if i don't care about speed, should i do float precision for better accuracy? whats a good basic config for highest accuracy?

  • @gnetwork
    @gnetwork 8 หลายเดือนก่อน

    First, thank you for your Guide. But as other mentioned, it has been trained on a known person what makes it a bit hard to see if the values are correctly. Could you make a video with an unknown person?

  • @laperchaude
    @laperchaude 8 หลายเดือนก่อน

    Merci pour le tutorial et l'outil!!

  • @DraikRoan
    @DraikRoan 8 หลายเดือนก่อน

    MFW when you check OP's textual embeddings 😏 allfoursnaked corneosidedoggy sexyback bimbostyle busty_15

  • @MonkeyDIvan
    @MonkeyDIvan 8 หลายเดือนก่อน

    Amazing stuff! Thank you!

  • @s3bl31
    @s3bl31 8 หลายเดือนก่อน

    Dont work for me i dont know what the problem is in the preview i see a good face but in the last step i turns back to the bad face and the output is just a even worse oversharpend face

    • @marksanders3662
      @marksanders3662 3 หลายเดือนก่อน

      I have the same problem. Have you solved it?

    • @s3bl31
      @s3bl31 3 หลายเดือนก่อน

      @@marksanders3662 Are you using a amd card? if so i think i fixed with --no-half in the command line. But idk since its that long ago and i switched to Nvidia.

  • @whyjordie
    @whyjordie 9 หลายเดือนก่อน

    this works!! thank you! i was struggling so hard lol

  • @bhupendraking9370
    @bhupendraking9370 9 หลายเดือนก่อน

    How to train SDXL model with large number of images? like example 30k images

  • @botlifegamer7026
    @botlifegamer7026 9 หลายเดือนก่อน

    I know this might be outdated not but for some reason When I train using BF16 I can only use a batch size of 1 or I get out of memory. even at 512,512. for a test. How are you able to do 8 I have a 3090 by the way. followed your settings im using BF16 and 23 images.

  • @lemur1870
    @lemur1870 9 หลายเดือนก่อน

    I have trained the lora, and I have used it for stable diffusion, but I keep on getting black images, doesnt mater what i type in

  • @adair546
    @adair546 9 หลายเดือนก่อน

    Hi! @Bernard Maltais, do you have a relation whit HandyPrint?

  • @rondovk
    @rondovk 9 หลายเดือนก่อน

    grateful for video but damn that audio makes me wanna die

  • @LouisCubingChannel
    @LouisCubingChannel 10 หลายเดือนก่อน

    kohya_ss 和 webui 可以共用虚拟环境吗

  • @christopherquenneville
    @christopherquenneville 10 หลายเดือนก่อน

    Quebec Power

  • @Soulsphere001
    @Soulsphere001 10 หลายเดือนก่อน

    One problem that I'm running into is that it takes over three hours to train using 7500 steps, but looking at my CPU and GPU performance shows that I'm only using around 10% of my CPU and usually only around 3% or lower on my GPU.

  • @didierlebouc2788
    @didierlebouc2788 10 หลายเดือนก่อน

    Excellent presentation

  • @hotmarcnet
    @hotmarcnet 10 หลายเดือนก่อน

    🎯 Key Takeaways for quick navigation: 00:02 🌐 El video explica cómo recopilar imágenes de internet y prepararlas para crear un modelo LoRA utilizando Kohya SS GUI. 01:13 🖼️ Es importante seleccionar imágenes de alta resolución para el modelo LoRA, no importa la proporción. 03:19 📂 Se organiza una carpeta llamada "Jennifer Aniston" para recopilar las imágenes y otra subcarpeta llamada "image" para almacenarlas. 09:27 📸 Utiliza Gigapixel para mejorar la resolución de las imágenes y luego las guarda en formato JPEG. 24:10 🖼️ Emplea el Auto1-1-1 Web UI para mejorar la calidad de algunas imágenes, ajustando parámetros como el denoising strength. 28:03 📈 Refina las imágenes con Auto1-1-1 Web UI, manteniendo la apariencia original de los rostros para no comprometer el modelo final. 29:42 🖼️ La preparación de un conjunto de datos para crear un modelo LoRA comienza con la limpieza y mejora de la resolución de las imágenes. 32:18 📷 Después de la preparación de las imágenes, es importante agregar descripciones a cada una de ellas para enseñar al modelo sobre el tema relevante y no importantes detalles. 34:33 📝 Las descripciones se pueden validar abriendo cada archivo de texto junto con la imagen correspondiente para asegurarse de que sean precisas. 44:10 🚀 Una vez validadas las descripciones, se está listo para configurar el modelo LoRA y proceder con la fase dos del entrenamiento. Made with HARPA AI

  • @armandoaddel
    @armandoaddel 10 หลายเดือนก่อน

    Having issues with the Group images script, It only does 10 images at the time and only once. Windows 11 machine. Anyone else having that issue.

  • @Sabrinahuskydog
    @Sabrinahuskydog 10 หลายเดือนก่อน

    Your audio is fucking awful with an echo in it. We can't understand you. I just had to mute it and use CC instead.

    • @RobertJene
      @RobertJene 5 หลายเดือนก่อน

      ANNNNNDDDD he's eating

  • @ThaRaiZe
    @ThaRaiZe 10 หลายเดือนก่อน

    These settings dont work. I keep getting CUDA out of memory issues even when using a 24 GB card. LoRa's train no problem. There's something off in these settings

  • @I77AGIC
    @I77AGIC 10 หลายเดือนก่อน

    45 minutes??

  • @metasamsara
    @metasamsara 10 หลายเดือนก่อน

    thank you, really clear and concise tutorial

  • @GooseAlarm
    @GooseAlarm 11 หลายเดือนก่อน

    Does anyone know why I continue to get - CalledProcessError: Command '['C:\\Users\\Name\\Desktop\\Vlad Diffusion\\kohya_ss\\venv\\Scripts\\python.exe', returned non-zero exit status 2. I can't find anything that fixes this issue and after watching so many videos on Kohya it feels like I'm the only one with this issue.

  • @NetTubeUser
    @NetTubeUser 11 หลายเดือนก่อน

    10:00 -- Question: Have you used the Internet since 1991 or just recently, like two years ago? Your image is smaller because, for some strange reason and logic, you think that dragging an image from Google Search into a folder will restore it to its original size which is completely wrong! You have taken the image from the temporary cache of your Internet browser! To get the right size of the image that you want to save, you need to access that image FROM the website, directly. But because you don't want to visit the whole site, you only need to use the direct link to the image. To do that: On the Google Search page, click on the image to open and see the Google preview of that image FROM the website, then right-click and select "Open image in new tab". And NOW and only NOW you have the right size image and can drag it into your folder (or right-click and save) ... not the smaller version of the image from the temporary cache!

  • @owenpaulmeyer
    @owenpaulmeyer 11 หลายเดือนก่อน

    🎯 Key Takeaways for quick navigation: The video is about creating a LoRA weight using kohya_ss GUI for model training. The process involves collecting images from the internet and enhancing them for model use. The speaker demonstrates searching for high-resolution images of Jennifer Aniston from the 1990s. The images are collected in a specific folder structure for processing. The speaker uses Gigapixel AI to upscale and clarify the images. Some images are processed using the Koya SS GUI tool to improve their quality further. The speaker discusses considerations for aspect ratios and cropping images. The images are being prepared for use in a model training process. Some images are discussed as potentially usable while others may require more improvement. The video highlights the workflow of collecting, enhancing, and preparing images for training. 32:18 The next stage involves adding text captions to the images to train the LoRA model. 32:49 Captions are added by pasting the image directory and using utility tools for text captioning. 33:20 The process involves configuring parameters like batch size, beams, and minimum caption length. 34:18 Captions start appearing next to each image file as the process proceeds. 35:01 Descriptions are reviewed and edited for accuracy and relevance. 35:26 Descriptions focus on clothing and essential details while avoiding unnecessary specifics. 40:10 Providing clothing details helps the model learn without enforcing specific attributes. 44:31 The process of captioning and validating images is complete, preparing for the next phase of training the LoRA model. Made with HARPA AI

  • @ysy69
    @ysy69 11 หลายเดือนก่อน

    Appreciate your tutorials. From all of your use and testing of SD, would you arrive at your best preference being Lora training and not Dreambooth ?

  • @letitiadominica2588
    @letitiadominica2588 11 หลายเดือนก่อน

    learned from the video, but tried 2 on the same settings and it resulted in a melted faces like the one you are showing in this video @ 80%. Is hires fix a requirement to make it work?

  • @nikoleifalkon
    @nikoleifalkon ปีที่แล้ว

    Thanks, now I've made 3 successful LoHA files and made a tutorial guide in spanish, if you want to look and made some corrections I would be so thankful, again, thank you very much

  • @VooDooEf
    @VooDooEf ปีที่แล้ว

    32:28 she looks more like a trans woman now :D